var/home/core/zuul-output/0000755000175000017500000000000015112320255014521 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015112330765015474 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log0000644000000000000000003140615715112330757017710 0ustar rootrootNov 28 13:20:42 crc systemd[1]: Starting Kubernetes Kubelet... Nov 28 13:20:42 crc restorecon[4578]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:42 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 28 13:20:43 crc restorecon[4578]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 28 13:20:43 crc restorecon[4578]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Nov 28 13:20:43 crc kubenswrapper[4631]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 28 13:20:43 crc kubenswrapper[4631]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Nov 28 13:20:43 crc kubenswrapper[4631]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 28 13:20:43 crc kubenswrapper[4631]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 28 13:20:43 crc kubenswrapper[4631]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Nov 28 13:20:43 crc kubenswrapper[4631]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.333223 4631 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.336382 4631 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.336404 4631 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.336409 4631 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.336414 4631 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.336422 4631 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.336427 4631 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.336432 4631 feature_gate.go:330] unrecognized feature gate: SignatureStores Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.336437 4631 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.336441 4631 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.336445 4631 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.336449 4631 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.336453 4631 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.336456 4631 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.336461 4631 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.336465 4631 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.336469 4631 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.336474 4631 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.336479 4631 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.336483 4631 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.336487 4631 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.336491 4631 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.336495 4631 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.336498 4631 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.336503 4631 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.336507 4631 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.336511 4631 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.336516 4631 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.336521 4631 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.336527 4631 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.336532 4631 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.336537 4631 feature_gate.go:330] unrecognized feature gate: Example Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.336544 4631 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.336550 4631 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.336556 4631 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.336561 4631 feature_gate.go:330] unrecognized feature gate: GatewayAPI Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.336567 4631 feature_gate.go:330] unrecognized feature gate: PinnedImages Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.336571 4631 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.336577 4631 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.336583 4631 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.336588 4631 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.336592 4631 feature_gate.go:330] unrecognized feature gate: PlatformOperators Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.336595 4631 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.336599 4631 feature_gate.go:330] unrecognized feature gate: OVNObservability Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.336603 4631 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.336606 4631 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.336611 4631 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.336615 4631 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.336621 4631 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.336625 4631 feature_gate.go:330] unrecognized feature gate: NewOLM Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.336628 4631 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.336632 4631 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.336635 4631 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.336641 4631 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.336646 4631 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.336650 4631 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.336654 4631 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.336658 4631 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.336662 4631 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.336665 4631 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.336669 4631 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.336673 4631 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.336677 4631 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.336680 4631 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.336684 4631 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.336687 4631 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.336691 4631 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.336694 4631 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.336699 4631 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.336702 4631 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.336706 4631 feature_gate.go:330] unrecognized feature gate: InsightsConfig Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.336709 4631 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.336965 4631 flags.go:64] FLAG: --address="0.0.0.0" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.336978 4631 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.336985 4631 flags.go:64] FLAG: --anonymous-auth="true" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.336992 4631 flags.go:64] FLAG: --application-metrics-count-limit="100" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.337000 4631 flags.go:64] FLAG: --authentication-token-webhook="false" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.337004 4631 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.337010 4631 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.337016 4631 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.337021 4631 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.337025 4631 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.337030 4631 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.337035 4631 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.337039 4631 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.337043 4631 flags.go:64] FLAG: --cgroup-root="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.337047 4631 flags.go:64] FLAG: --cgroups-per-qos="true" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.337051 4631 flags.go:64] FLAG: --client-ca-file="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.337055 4631 flags.go:64] FLAG: --cloud-config="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.337059 4631 flags.go:64] FLAG: --cloud-provider="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.337063 4631 flags.go:64] FLAG: --cluster-dns="[]" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.337069 4631 flags.go:64] FLAG: --cluster-domain="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.337074 4631 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.337079 4631 flags.go:64] FLAG: --config-dir="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.337084 4631 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.337088 4631 flags.go:64] FLAG: --container-log-max-files="5" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.337095 4631 flags.go:64] FLAG: --container-log-max-size="10Mi" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.337100 4631 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.337104 4631 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.337109 4631 flags.go:64] FLAG: --containerd-namespace="k8s.io" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.337113 4631 flags.go:64] FLAG: --contention-profiling="false" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.337117 4631 flags.go:64] FLAG: --cpu-cfs-quota="true" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.337122 4631 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.337126 4631 flags.go:64] FLAG: --cpu-manager-policy="none" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.337130 4631 flags.go:64] FLAG: --cpu-manager-policy-options="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.337135 4631 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.337139 4631 flags.go:64] FLAG: --enable-controller-attach-detach="true" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.337144 4631 flags.go:64] FLAG: --enable-debugging-handlers="true" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.337148 4631 flags.go:64] FLAG: --enable-load-reader="false" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.337152 4631 flags.go:64] FLAG: --enable-server="true" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.337156 4631 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.337161 4631 flags.go:64] FLAG: --event-burst="100" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.337166 4631 flags.go:64] FLAG: --event-qps="50" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.337171 4631 flags.go:64] FLAG: --event-storage-age-limit="default=0" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.337175 4631 flags.go:64] FLAG: --event-storage-event-limit="default=0" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.337179 4631 flags.go:64] FLAG: --eviction-hard="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.337185 4631 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.337189 4631 flags.go:64] FLAG: --eviction-minimum-reclaim="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.337193 4631 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.337197 4631 flags.go:64] FLAG: --eviction-soft="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.337201 4631 flags.go:64] FLAG: --eviction-soft-grace-period="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.337205 4631 flags.go:64] FLAG: --exit-on-lock-contention="false" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.337210 4631 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.337214 4631 flags.go:64] FLAG: --experimental-mounter-path="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.337218 4631 flags.go:64] FLAG: --fail-cgroupv1="false" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.337222 4631 flags.go:64] FLAG: --fail-swap-on="true" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.337226 4631 flags.go:64] FLAG: --feature-gates="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.337233 4631 flags.go:64] FLAG: --file-check-frequency="20s" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.337237 4631 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.337242 4631 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.337247 4631 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.337251 4631 flags.go:64] FLAG: --healthz-port="10248" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.337255 4631 flags.go:64] FLAG: --help="false" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.337262 4631 flags.go:64] FLAG: --hostname-override="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.337266 4631 flags.go:64] FLAG: --housekeeping-interval="10s" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.337270 4631 flags.go:64] FLAG: --http-check-frequency="20s" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.337275 4631 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.337280 4631 flags.go:64] FLAG: --image-credential-provider-config="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.337303 4631 flags.go:64] FLAG: --image-gc-high-threshold="85" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.337308 4631 flags.go:64] FLAG: --image-gc-low-threshold="80" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.337312 4631 flags.go:64] FLAG: --image-service-endpoint="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.337316 4631 flags.go:64] FLAG: --kernel-memcg-notification="false" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.337320 4631 flags.go:64] FLAG: --kube-api-burst="100" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.337324 4631 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.337329 4631 flags.go:64] FLAG: --kube-api-qps="50" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.337333 4631 flags.go:64] FLAG: --kube-reserved="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.337337 4631 flags.go:64] FLAG: --kube-reserved-cgroup="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.337341 4631 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.337345 4631 flags.go:64] FLAG: --kubelet-cgroups="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.337350 4631 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.337354 4631 flags.go:64] FLAG: --lock-file="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.337358 4631 flags.go:64] FLAG: --log-cadvisor-usage="false" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.337362 4631 flags.go:64] FLAG: --log-flush-frequency="5s" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.337366 4631 flags.go:64] FLAG: --log-json-info-buffer-size="0" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.337373 4631 flags.go:64] FLAG: --log-json-split-stream="false" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.337377 4631 flags.go:64] FLAG: --log-text-info-buffer-size="0" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.337382 4631 flags.go:64] FLAG: --log-text-split-stream="false" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.337386 4631 flags.go:64] FLAG: --logging-format="text" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.337390 4631 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.337395 4631 flags.go:64] FLAG: --make-iptables-util-chains="true" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.337399 4631 flags.go:64] FLAG: --manifest-url="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.337404 4631 flags.go:64] FLAG: --manifest-url-header="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.337410 4631 flags.go:64] FLAG: --max-housekeeping-interval="15s" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.337415 4631 flags.go:64] FLAG: --max-open-files="1000000" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.337426 4631 flags.go:64] FLAG: --max-pods="110" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.337431 4631 flags.go:64] FLAG: --maximum-dead-containers="-1" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.337435 4631 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.337439 4631 flags.go:64] FLAG: --memory-manager-policy="None" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.337444 4631 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.337448 4631 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.337453 4631 flags.go:64] FLAG: --node-ip="192.168.126.11" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.337457 4631 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.337467 4631 flags.go:64] FLAG: --node-status-max-images="50" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.337471 4631 flags.go:64] FLAG: --node-status-update-frequency="10s" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.337475 4631 flags.go:64] FLAG: --oom-score-adj="-999" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.337480 4631 flags.go:64] FLAG: --pod-cidr="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.337484 4631 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.337492 4631 flags.go:64] FLAG: --pod-manifest-path="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.337496 4631 flags.go:64] FLAG: --pod-max-pids="-1" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.337500 4631 flags.go:64] FLAG: --pods-per-core="0" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.337504 4631 flags.go:64] FLAG: --port="10250" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.337508 4631 flags.go:64] FLAG: --protect-kernel-defaults="false" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.337512 4631 flags.go:64] FLAG: --provider-id="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.337517 4631 flags.go:64] FLAG: --qos-reserved="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.337521 4631 flags.go:64] FLAG: --read-only-port="10255" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.337525 4631 flags.go:64] FLAG: --register-node="true" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.337529 4631 flags.go:64] FLAG: --register-schedulable="true" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.337533 4631 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.337540 4631 flags.go:64] FLAG: --registry-burst="10" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.337544 4631 flags.go:64] FLAG: --registry-qps="5" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.337548 4631 flags.go:64] FLAG: --reserved-cpus="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.337552 4631 flags.go:64] FLAG: --reserved-memory="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.337557 4631 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.337562 4631 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.337566 4631 flags.go:64] FLAG: --rotate-certificates="false" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.337570 4631 flags.go:64] FLAG: --rotate-server-certificates="false" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.337577 4631 flags.go:64] FLAG: --runonce="false" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.337581 4631 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.337586 4631 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.337591 4631 flags.go:64] FLAG: --seccomp-default="false" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.337596 4631 flags.go:64] FLAG: --serialize-image-pulls="true" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.337600 4631 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.337605 4631 flags.go:64] FLAG: --storage-driver-db="cadvisor" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.337609 4631 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.337613 4631 flags.go:64] FLAG: --storage-driver-password="root" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.337618 4631 flags.go:64] FLAG: --storage-driver-secure="false" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.337622 4631 flags.go:64] FLAG: --storage-driver-table="stats" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.337627 4631 flags.go:64] FLAG: --storage-driver-user="root" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.337631 4631 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.337636 4631 flags.go:64] FLAG: --sync-frequency="1m0s" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.337640 4631 flags.go:64] FLAG: --system-cgroups="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.337644 4631 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.337652 4631 flags.go:64] FLAG: --system-reserved-cgroup="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.337655 4631 flags.go:64] FLAG: --tls-cert-file="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.337659 4631 flags.go:64] FLAG: --tls-cipher-suites="[]" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.337664 4631 flags.go:64] FLAG: --tls-min-version="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.337669 4631 flags.go:64] FLAG: --tls-private-key-file="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.337673 4631 flags.go:64] FLAG: --topology-manager-policy="none" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.337677 4631 flags.go:64] FLAG: --topology-manager-policy-options="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.337681 4631 flags.go:64] FLAG: --topology-manager-scope="container" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.337685 4631 flags.go:64] FLAG: --v="2" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.337691 4631 flags.go:64] FLAG: --version="false" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.337697 4631 flags.go:64] FLAG: --vmodule="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.337702 4631 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.337707 4631 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.337820 4631 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.337825 4631 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.337829 4631 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.337836 4631 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.337840 4631 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.337845 4631 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.337849 4631 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.337853 4631 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.337857 4631 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.337862 4631 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.337867 4631 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.337871 4631 feature_gate.go:330] unrecognized feature gate: PlatformOperators Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.337875 4631 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.337879 4631 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.337883 4631 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.337887 4631 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.337890 4631 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.337894 4631 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.337897 4631 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.337901 4631 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.337904 4631 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.337907 4631 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.337911 4631 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.337914 4631 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.337918 4631 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.337921 4631 feature_gate.go:330] unrecognized feature gate: GatewayAPI Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.337925 4631 feature_gate.go:330] unrecognized feature gate: SignatureStores Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.337928 4631 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.337932 4631 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.337935 4631 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.337939 4631 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.337942 4631 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.337946 4631 feature_gate.go:330] unrecognized feature gate: OVNObservability Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.337949 4631 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.337953 4631 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.337958 4631 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.337962 4631 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.337966 4631 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.337970 4631 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.337974 4631 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.337978 4631 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.337981 4631 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.337986 4631 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.337991 4631 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.337995 4631 feature_gate.go:330] unrecognized feature gate: Example Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.337999 4631 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.338003 4631 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.338007 4631 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.338011 4631 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.338014 4631 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.338018 4631 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.338021 4631 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.338025 4631 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.338028 4631 feature_gate.go:330] unrecognized feature gate: NewOLM Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.338032 4631 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.338035 4631 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.338039 4631 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.338042 4631 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.338046 4631 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.338050 4631 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.338069 4631 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.338073 4631 feature_gate.go:330] unrecognized feature gate: PinnedImages Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.338077 4631 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.338080 4631 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.338084 4631 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.338087 4631 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.338091 4631 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.338096 4631 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.338101 4631 feature_gate.go:330] unrecognized feature gate: InsightsConfig Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.338113 4631 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.338117 4631 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.338305 4631 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.349549 4631 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.349595 4631 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.349941 4631 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.349960 4631 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.349968 4631 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.349974 4631 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.349979 4631 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.349984 4631 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.349989 4631 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.349994 4631 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.349998 4631 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.350003 4631 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.350015 4631 feature_gate.go:330] unrecognized feature gate: GatewayAPI Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.350019 4631 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.350024 4631 feature_gate.go:330] unrecognized feature gate: Example Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.350030 4631 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.350037 4631 feature_gate.go:330] unrecognized feature gate: InsightsConfig Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.350064 4631 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.350069 4631 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.350073 4631 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.350077 4631 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.350082 4631 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.350087 4631 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.350091 4631 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.350100 4631 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.350104 4631 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.350110 4631 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.350115 4631 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.350119 4631 feature_gate.go:330] unrecognized feature gate: SignatureStores Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.350125 4631 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.350131 4631 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.350135 4631 feature_gate.go:330] unrecognized feature gate: NewOLM Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.350142 4631 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.350146 4631 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.350151 4631 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.350156 4631 feature_gate.go:330] unrecognized feature gate: PlatformOperators Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.350160 4631 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.350168 4631 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.350173 4631 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.350179 4631 feature_gate.go:330] unrecognized feature gate: OVNObservability Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.350184 4631 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.350188 4631 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.350192 4631 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.350197 4631 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.350202 4631 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.350208 4631 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.350212 4631 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.350216 4631 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.350221 4631 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.350229 4631 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.350234 4631 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.350238 4631 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.350243 4631 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.350247 4631 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.350251 4631 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.350256 4631 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.350260 4631 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.350265 4631 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.350272 4631 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.350277 4631 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.350301 4631 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.350309 4631 feature_gate.go:330] unrecognized feature gate: PinnedImages Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.350313 4631 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.350318 4631 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.350324 4631 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.350330 4631 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.350335 4631 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.350341 4631 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.350346 4631 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.350353 4631 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.350357 4631 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.350362 4631 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.350368 4631 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.350376 4631 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.350709 4631 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.350722 4631 feature_gate.go:330] unrecognized feature gate: PinnedImages Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.350727 4631 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.350732 4631 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.350737 4631 feature_gate.go:330] unrecognized feature gate: OVNObservability Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.350741 4631 feature_gate.go:330] unrecognized feature gate: NewOLM Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.350745 4631 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.350751 4631 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.350756 4631 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.350766 4631 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.350772 4631 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.350778 4631 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.350783 4631 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.350789 4631 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.350793 4631 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.350797 4631 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.350803 4631 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.350808 4631 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.350813 4631 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.350817 4631 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.350821 4631 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.350832 4631 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.350837 4631 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.350842 4631 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.350847 4631 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.350851 4631 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.350856 4631 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.350860 4631 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.350865 4631 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.350870 4631 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.350874 4631 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.350879 4631 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.350883 4631 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.350892 4631 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.350896 4631 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.350901 4631 feature_gate.go:330] unrecognized feature gate: PlatformOperators Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.350906 4631 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.350912 4631 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.350917 4631 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.350922 4631 feature_gate.go:330] unrecognized feature gate: SignatureStores Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.350926 4631 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.350931 4631 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.350936 4631 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.350940 4631 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.350945 4631 feature_gate.go:330] unrecognized feature gate: InsightsConfig Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.350949 4631 feature_gate.go:330] unrecognized feature gate: GatewayAPI Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.350958 4631 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.350965 4631 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.350971 4631 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.350976 4631 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.350982 4631 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.350987 4631 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.350992 4631 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.350997 4631 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.351002 4631 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.351006 4631 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.351010 4631 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.351015 4631 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.351019 4631 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.351027 4631 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.351031 4631 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.351036 4631 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.351041 4631 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.351046 4631 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.351051 4631 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.351055 4631 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.351060 4631 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.351065 4631 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.351070 4631 feature_gate.go:330] unrecognized feature gate: Example Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.351075 4631 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.351080 4631 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.351092 4631 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.352343 4631 server.go:940] "Client rotation is on, will bootstrap in background" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.356890 4631 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.357046 4631 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.357891 4631 server.go:997] "Starting client certificate rotation" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.357930 4631 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.358252 4631 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2025-11-19 09:49:30.784776359 +0000 UTC Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.358457 4631 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.364471 4631 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Nov 28 13:20:43 crc kubenswrapper[4631]: E1128 13:20:43.367014 4631 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 38.102.83.194:6443: connect: connection refused" logger="UnhandledError" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.367151 4631 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.377269 4631 log.go:25] "Validated CRI v1 runtime API" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.395458 4631 log.go:25] "Validated CRI v1 image API" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.397538 4631 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.400337 4631 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2025-11-28-13-15-16-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.400365 4631 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:42 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:41 fsType:tmpfs blockSize:0}] Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.412170 4631 manager.go:217] Machine: {Timestamp:2025-11-28 13:20:43.410785621 +0000 UTC m=+0.218089005 CPUVendorID:AuthenticAMD NumCores:8 NumPhysicalCores:1 NumSockets:8 CpuFrequency:2799998 MemoryCapacity:25199472640 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:0670c19c-6aaa-48af-8d2a-ee47e63640da BootID:4d70419e-7663-4d3f-a15c-3c3ee0ecadaa Filesystems:[{Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:12599738368 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:42 Capacity:2519945216 Type:vfs Inodes:615221 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:41 Capacity:1073741824 Type:vfs Inodes:3076107 HasInodes:true} {Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:12599734272 Type:vfs Inodes:3076107 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:5039894528 Type:vfs Inodes:819200 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:429496729600 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:5d:00:81 Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:5d:00:81 Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:fd:19:92 Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:18:3e:d6 Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:a0:a6:27 Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:17:5a:a5 Speed:-1 Mtu:1496} {Name:eth10 MacAddress:b2:57:70:35:25:42 Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:de:11:fb:f7:a0:46 Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:25199472640 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.412398 4631 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.412520 4631 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.413018 4631 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.413196 4631 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.413226 4631 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.413482 4631 topology_manager.go:138] "Creating topology manager with none policy" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.413495 4631 container_manager_linux.go:303] "Creating device plugin manager" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.413714 4631 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.413756 4631 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.414062 4631 state_mem.go:36] "Initialized new in-memory state store" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.414153 4631 server.go:1245] "Using root directory" path="/var/lib/kubelet" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.414817 4631 kubelet.go:418] "Attempting to sync node with API server" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.414841 4631 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.414872 4631 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.414891 4631 kubelet.go:324] "Adding apiserver pod source" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.414927 4631 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.418049 4631 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.194:6443: connect: connection refused Nov 28 13:20:43 crc kubenswrapper[4631]: E1128 13:20:43.418154 4631 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.194:6443: connect: connection refused" logger="UnhandledError" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.418401 4631 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.418779 4631 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.419547 4631 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.194:6443: connect: connection refused Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.419760 4631 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Nov 28 13:20:43 crc kubenswrapper[4631]: E1128 13:20:43.419721 4631 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.194:6443: connect: connection refused" logger="UnhandledError" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.421103 4631 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.421166 4631 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.421181 4631 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.421194 4631 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.421216 4631 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.421228 4631 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.421241 4631 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.421262 4631 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.421278 4631 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.421317 4631 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.421338 4631 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.421350 4631 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.421579 4631 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.422443 4631 server.go:1280] "Started kubelet" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.422845 4631 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.423061 4631 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.423117 4631 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.194:6443: connect: connection refused Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.423683 4631 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Nov 28 13:20:43 crc systemd[1]: Started Kubernetes Kubelet. Nov 28 13:20:43 crc kubenswrapper[4631]: E1128 13:20:43.425263 4631 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.194:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.187c2e489c9c2040 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-11-28 13:20:43.422392384 +0000 UTC m=+0.229695758,LastTimestamp:2025-11-28 13:20:43.422392384 +0000 UTC m=+0.229695758,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.426221 4631 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.426277 4631 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.426376 4631 volume_manager.go:287] "The desired_state_of_world populator starts" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.426388 4631 volume_manager.go:289] "Starting Kubelet Volume Manager" Nov 28 13:20:43 crc kubenswrapper[4631]: E1128 13:20:43.426452 4631 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.426784 4631 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-02 06:01:19.442956063 +0000 UTC Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.426874 4631 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 832h40m36.016086426s for next certificate rotation Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.427724 4631 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.194:6443: connect: connection refused Nov 28 13:20:43 crc kubenswrapper[4631]: E1128 13:20:43.427819 4631 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.194:6443: connect: connection refused" logger="UnhandledError" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.428096 4631 factory.go:55] Registering systemd factory Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.428123 4631 factory.go:221] Registration of the systemd container factory successfully Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.428228 4631 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Nov 28 13:20:43 crc kubenswrapper[4631]: E1128 13:20:43.428866 4631 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.194:6443: connect: connection refused" interval="200ms" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.428905 4631 server.go:460] "Adding debug handlers to kubelet server" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.429014 4631 factory.go:153] Registering CRI-O factory Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.429035 4631 factory.go:221] Registration of the crio container factory successfully Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.429112 4631 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.429141 4631 factory.go:103] Registering Raw factory Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.429165 4631 manager.go:1196] Started watching for new ooms in manager Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.430195 4631 manager.go:319] Starting recovery of all containers Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.452004 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.452087 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.452105 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.452121 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.452136 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.452148 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.452164 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.452180 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.452200 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.452214 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.452230 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.452245 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.452259 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.452277 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.452314 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.452327 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.452343 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.452355 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.452370 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.452384 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.452397 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.452412 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.452426 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.452437 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.452449 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.452463 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.452478 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.452492 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.452504 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.452514 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.453248 4631 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.453331 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.453345 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.453360 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.453376 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.453391 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.453405 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.453418 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.453431 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.453446 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.453459 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.453474 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.453485 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.453496 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.453510 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.453531 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.453557 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.453573 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.453589 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.453608 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.453620 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.453639 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.453652 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.453674 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.453687 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.453703 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.453717 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.453729 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.453747 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.453763 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.453781 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.453794 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.453809 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.453825 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.453841 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.453857 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.453872 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.453892 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.453912 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.453930 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.453981 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.454000 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.454019 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.454039 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.454056 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.454072 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.454087 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.454101 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.454116 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.454133 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.454148 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.454164 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.454179 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.454200 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.454217 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.454244 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.454262 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.454280 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.454316 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.454333 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.454351 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.454367 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.454386 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.454416 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.454430 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.454444 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.454455 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.454468 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.454483 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.454501 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.454517 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.454531 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.454544 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.454557 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.454571 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.454594 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.454609 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.454626 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.454641 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.454656 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.454673 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.454688 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.454703 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.454718 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.454732 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.454746 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.454763 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.454777 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.454793 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.454809 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.454859 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.454874 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.454890 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.454908 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.454921 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.454935 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.454951 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.454967 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.454984 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.454999 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.455015 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.455031 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.455044 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.455060 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.455073 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.455085 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.455101 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.455115 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.455129 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.455142 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.455156 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.455173 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.455186 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.455199 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.455214 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.455231 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.455248 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.455265 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.455282 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.455341 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.455355 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.455413 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.455428 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.455447 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.455464 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.455480 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.455496 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.455511 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.455526 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.455542 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.455558 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.455571 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.455587 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.455604 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.455618 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.455635 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.455652 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.455669 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.455685 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.455700 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.455732 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.455746 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.455761 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.455776 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.455791 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.455806 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.455824 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.455840 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.455854 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.455869 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.455884 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.455897 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.455912 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.455926 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.455940 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.455955 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.455969 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.455982 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.455995 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.456007 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.456020 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.456034 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.456048 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.456064 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.456079 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.456098 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.456115 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.456130 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.456146 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.456163 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.456181 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.456197 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.456216 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.456232 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.456245 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.456260 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.456272 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.456308 4631 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.456321 4631 reconstruct.go:97] "Volume reconstruction finished" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.456329 4631 reconciler.go:26] "Reconciler: start to sync state" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.474961 4631 manager.go:324] Recovery completed Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.487186 4631 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.488794 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.489183 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.489212 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.491191 4631 cpu_manager.go:225] "Starting CPU manager" policy="none" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.491212 4631 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.491238 4631 state_mem.go:36] "Initialized new in-memory state store" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.502992 4631 policy_none.go:49] "None policy: Start" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.504198 4631 memory_manager.go:170] "Starting memorymanager" policy="None" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.504228 4631 state_mem.go:35] "Initializing new in-memory state store" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.506324 4631 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.511666 4631 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.511742 4631 status_manager.go:217] "Starting to sync pod status with apiserver" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.511791 4631 kubelet.go:2335] "Starting kubelet main sync loop" Nov 28 13:20:43 crc kubenswrapper[4631]: E1128 13:20:43.511844 4631 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.515763 4631 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.194:6443: connect: connection refused Nov 28 13:20:43 crc kubenswrapper[4631]: E1128 13:20:43.515852 4631 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.194:6443: connect: connection refused" logger="UnhandledError" Nov 28 13:20:43 crc kubenswrapper[4631]: E1128 13:20:43.526771 4631 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.553596 4631 manager.go:334] "Starting Device Plugin manager" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.553665 4631 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.553686 4631 server.go:79] "Starting device plugin registration server" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.554336 4631 eviction_manager.go:189] "Eviction manager: starting control loop" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.554362 4631 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.554854 4631 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.554948 4631 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.554959 4631 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Nov 28 13:20:43 crc kubenswrapper[4631]: E1128 13:20:43.566269 4631 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.612865 4631 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc","openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc"] Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.613021 4631 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.614540 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.614618 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.614632 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.614882 4631 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.615228 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.615370 4631 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.616522 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.616559 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.616573 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.616738 4631 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.616765 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.616801 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.616818 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.616928 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.617531 4631 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.618367 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.618437 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.618456 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.618788 4631 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.619147 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.619245 4631 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.621485 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.621517 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.621519 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.621539 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.621545 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.621560 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.621611 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.621565 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.621685 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.621937 4631 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.622150 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.622217 4631 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.623555 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.623581 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.623593 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.623768 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.623803 4631 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.624177 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.624208 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.624218 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.625017 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.625078 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.625093 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 13:20:43 crc kubenswrapper[4631]: E1128 13:20:43.629802 4631 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.194:6443: connect: connection refused" interval="400ms" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.654789 4631 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.656163 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.656259 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.656273 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.656365 4631 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 28 13:20:43 crc kubenswrapper[4631]: E1128 13:20:43.657425 4631 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.194:6443: connect: connection refused" node="crc" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.657613 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.657654 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.657688 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.657714 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.657772 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.657798 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.657858 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.657974 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.658038 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.658066 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.658096 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.658122 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.658146 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.658172 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.658196 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.759215 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.759267 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.759312 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.759335 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.759359 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.759385 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.759406 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.759430 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.759444 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.759486 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.759451 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.759526 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.759529 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.759403 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.759583 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.759601 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.759605 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.759618 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.759627 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.759639 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.759670 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.759657 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.759692 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.759699 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.759768 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.759766 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.759875 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.759793 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.759807 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.759785 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.858056 4631 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.859948 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.860009 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.860022 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.860069 4631 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 28 13:20:43 crc kubenswrapper[4631]: E1128 13:20:43.860719 4631 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.194:6443: connect: connection refused" node="crc" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.958533 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 28 13:20:43 crc kubenswrapper[4631]: I1128 13:20:43.981754 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 28 13:20:43 crc kubenswrapper[4631]: W1128 13:20:43.985871 4631 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-83e4ad9604116fcf6952e555358fdad3b9ba1ab02692941aae55d53a905dd3a2 WatchSource:0}: Error finding container 83e4ad9604116fcf6952e555358fdad3b9ba1ab02692941aae55d53a905dd3a2: Status 404 returned error can't find the container with id 83e4ad9604116fcf6952e555358fdad3b9ba1ab02692941aae55d53a905dd3a2 Nov 28 13:20:44 crc kubenswrapper[4631]: I1128 13:20:44.002943 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 28 13:20:44 crc kubenswrapper[4631]: W1128 13:20:44.002943 4631 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-cd071f7fa93e6389b7002fd65ba80f0af38e68b599a12370a5461663b2b45f59 WatchSource:0}: Error finding container cd071f7fa93e6389b7002fd65ba80f0af38e68b599a12370a5461663b2b45f59: Status 404 returned error can't find the container with id cd071f7fa93e6389b7002fd65ba80f0af38e68b599a12370a5461663b2b45f59 Nov 28 13:20:44 crc kubenswrapper[4631]: I1128 13:20:44.022768 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Nov 28 13:20:44 crc kubenswrapper[4631]: I1128 13:20:44.030349 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 13:20:44 crc kubenswrapper[4631]: W1128 13:20:44.031259 4631 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-49f5ad81178622477a02e63c3b001a6ab2276a7ef36e7520774dbdf4d700abca WatchSource:0}: Error finding container 49f5ad81178622477a02e63c3b001a6ab2276a7ef36e7520774dbdf4d700abca: Status 404 returned error can't find the container with id 49f5ad81178622477a02e63c3b001a6ab2276a7ef36e7520774dbdf4d700abca Nov 28 13:20:44 crc kubenswrapper[4631]: E1128 13:20:44.031419 4631 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.194:6443: connect: connection refused" interval="800ms" Nov 28 13:20:44 crc kubenswrapper[4631]: W1128 13:20:44.041392 4631 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-b455fee146449c4995c1755cfe13facb5de4de5212195d2a9cab8da3e6ff2c35 WatchSource:0}: Error finding container b455fee146449c4995c1755cfe13facb5de4de5212195d2a9cab8da3e6ff2c35: Status 404 returned error can't find the container with id b455fee146449c4995c1755cfe13facb5de4de5212195d2a9cab8da3e6ff2c35 Nov 28 13:20:44 crc kubenswrapper[4631]: W1128 13:20:44.053254 4631 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-d69184066326704104555d75a1d85a12396c0bfb79f58e65222906fd5074701b WatchSource:0}: Error finding container d69184066326704104555d75a1d85a12396c0bfb79f58e65222906fd5074701b: Status 404 returned error can't find the container with id d69184066326704104555d75a1d85a12396c0bfb79f58e65222906fd5074701b Nov 28 13:20:44 crc kubenswrapper[4631]: I1128 13:20:44.261001 4631 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 13:20:44 crc kubenswrapper[4631]: W1128 13:20:44.261849 4631 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.194:6443: connect: connection refused Nov 28 13:20:44 crc kubenswrapper[4631]: E1128 13:20:44.261981 4631 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.194:6443: connect: connection refused" logger="UnhandledError" Nov 28 13:20:44 crc kubenswrapper[4631]: I1128 13:20:44.262087 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 13:20:44 crc kubenswrapper[4631]: I1128 13:20:44.262127 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 13:20:44 crc kubenswrapper[4631]: I1128 13:20:44.262137 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 13:20:44 crc kubenswrapper[4631]: I1128 13:20:44.262160 4631 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 28 13:20:44 crc kubenswrapper[4631]: E1128 13:20:44.262594 4631 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.194:6443: connect: connection refused" node="crc" Nov 28 13:20:44 crc kubenswrapper[4631]: W1128 13:20:44.305904 4631 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.194:6443: connect: connection refused Nov 28 13:20:44 crc kubenswrapper[4631]: E1128 13:20:44.305985 4631 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.194:6443: connect: connection refused" logger="UnhandledError" Nov 28 13:20:44 crc kubenswrapper[4631]: I1128 13:20:44.424959 4631 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.194:6443: connect: connection refused Nov 28 13:20:44 crc kubenswrapper[4631]: I1128 13:20:44.524714 4631 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="71899c0fc4ad31c5ef073eeed3391b2959bdb528b944677e0828d2945332e400" exitCode=0 Nov 28 13:20:44 crc kubenswrapper[4631]: I1128 13:20:44.524804 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"71899c0fc4ad31c5ef073eeed3391b2959bdb528b944677e0828d2945332e400"} Nov 28 13:20:44 crc kubenswrapper[4631]: I1128 13:20:44.524922 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"49f5ad81178622477a02e63c3b001a6ab2276a7ef36e7520774dbdf4d700abca"} Nov 28 13:20:44 crc kubenswrapper[4631]: I1128 13:20:44.525015 4631 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 13:20:44 crc kubenswrapper[4631]: I1128 13:20:44.526174 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 13:20:44 crc kubenswrapper[4631]: I1128 13:20:44.526208 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 13:20:44 crc kubenswrapper[4631]: I1128 13:20:44.526218 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 13:20:44 crc kubenswrapper[4631]: I1128 13:20:44.528779 4631 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="20ca4fb5cafafc9c2007201ef10b323bd5255b3a9a3e6d9a9e84370f7a8aa4ed" exitCode=0 Nov 28 13:20:44 crc kubenswrapper[4631]: I1128 13:20:44.528861 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"20ca4fb5cafafc9c2007201ef10b323bd5255b3a9a3e6d9a9e84370f7a8aa4ed"} Nov 28 13:20:44 crc kubenswrapper[4631]: I1128 13:20:44.528915 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"cd071f7fa93e6389b7002fd65ba80f0af38e68b599a12370a5461663b2b45f59"} Nov 28 13:20:44 crc kubenswrapper[4631]: I1128 13:20:44.529010 4631 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 13:20:44 crc kubenswrapper[4631]: I1128 13:20:44.530438 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 13:20:44 crc kubenswrapper[4631]: I1128 13:20:44.530493 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 13:20:44 crc kubenswrapper[4631]: I1128 13:20:44.530514 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 13:20:44 crc kubenswrapper[4631]: I1128 13:20:44.532346 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"35d912d041da9a7ed307ebea6e999cf087dacf84a3cefbfd0f3a9c891c979408"} Nov 28 13:20:44 crc kubenswrapper[4631]: I1128 13:20:44.532403 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"83e4ad9604116fcf6952e555358fdad3b9ba1ab02692941aae55d53a905dd3a2"} Nov 28 13:20:44 crc kubenswrapper[4631]: I1128 13:20:44.534067 4631 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="f0b61d56ed070810021d253875a7bb5b9cb94a3446c56c4a92fb5418022dd643" exitCode=0 Nov 28 13:20:44 crc kubenswrapper[4631]: I1128 13:20:44.534127 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"f0b61d56ed070810021d253875a7bb5b9cb94a3446c56c4a92fb5418022dd643"} Nov 28 13:20:44 crc kubenswrapper[4631]: I1128 13:20:44.534161 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"d69184066326704104555d75a1d85a12396c0bfb79f58e65222906fd5074701b"} Nov 28 13:20:44 crc kubenswrapper[4631]: I1128 13:20:44.534349 4631 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 13:20:44 crc kubenswrapper[4631]: I1128 13:20:44.535639 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 13:20:44 crc kubenswrapper[4631]: I1128 13:20:44.535667 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 13:20:44 crc kubenswrapper[4631]: I1128 13:20:44.535677 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 13:20:44 crc kubenswrapper[4631]: I1128 13:20:44.536306 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"455b79e8da5d8a47fcc29e917fc26e396068b1f03d67712ecbf8cd9d5bfb9779"} Nov 28 13:20:44 crc kubenswrapper[4631]: I1128 13:20:44.536276 4631 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="455b79e8da5d8a47fcc29e917fc26e396068b1f03d67712ecbf8cd9d5bfb9779" exitCode=0 Nov 28 13:20:44 crc kubenswrapper[4631]: I1128 13:20:44.536368 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"b455fee146449c4995c1755cfe13facb5de4de5212195d2a9cab8da3e6ff2c35"} Nov 28 13:20:44 crc kubenswrapper[4631]: I1128 13:20:44.536397 4631 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 13:20:44 crc kubenswrapper[4631]: I1128 13:20:44.537147 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 13:20:44 crc kubenswrapper[4631]: I1128 13:20:44.537168 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 13:20:44 crc kubenswrapper[4631]: I1128 13:20:44.537178 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 13:20:44 crc kubenswrapper[4631]: I1128 13:20:44.537923 4631 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 13:20:44 crc kubenswrapper[4631]: I1128 13:20:44.538934 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 13:20:44 crc kubenswrapper[4631]: I1128 13:20:44.538983 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 13:20:44 crc kubenswrapper[4631]: I1128 13:20:44.539001 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 13:20:44 crc kubenswrapper[4631]: W1128 13:20:44.694340 4631 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.194:6443: connect: connection refused Nov 28 13:20:44 crc kubenswrapper[4631]: E1128 13:20:44.694444 4631 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.194:6443: connect: connection refused" logger="UnhandledError" Nov 28 13:20:44 crc kubenswrapper[4631]: W1128 13:20:44.808778 4631 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.194:6443: connect: connection refused Nov 28 13:20:44 crc kubenswrapper[4631]: E1128 13:20:44.808890 4631 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.194:6443: connect: connection refused" logger="UnhandledError" Nov 28 13:20:44 crc kubenswrapper[4631]: E1128 13:20:44.832779 4631 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.194:6443: connect: connection refused" interval="1.6s" Nov 28 13:20:45 crc kubenswrapper[4631]: I1128 13:20:45.062938 4631 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 13:20:45 crc kubenswrapper[4631]: I1128 13:20:45.063837 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 13:20:45 crc kubenswrapper[4631]: I1128 13:20:45.063862 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 13:20:45 crc kubenswrapper[4631]: I1128 13:20:45.063872 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 13:20:45 crc kubenswrapper[4631]: I1128 13:20:45.063898 4631 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 28 13:20:45 crc kubenswrapper[4631]: E1128 13:20:45.064447 4631 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.194:6443: connect: connection refused" node="crc" Nov 28 13:20:45 crc kubenswrapper[4631]: I1128 13:20:45.543342 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"8ac30b44d2fe8ddb4ede530d4b99dab4dc641a097e21d4ae69cfcbeddc38802d"} Nov 28 13:20:45 crc kubenswrapper[4631]: I1128 13:20:45.543419 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"eff71f12d67a15adda3183a742f2869b104d773904a6c5954d1ccb6a2c64dd71"} Nov 28 13:20:45 crc kubenswrapper[4631]: I1128 13:20:45.543434 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"53ea681ea681973dd5177ec7f2925ef48fec88ac004b41fca867ce671f5a76d1"} Nov 28 13:20:45 crc kubenswrapper[4631]: I1128 13:20:45.543446 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"c4f323ff537cd79159a51b592401e92fc06d54c490ff005486f27f41d42a35f2"} Nov 28 13:20:45 crc kubenswrapper[4631]: I1128 13:20:45.545532 4631 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="d3c2c475d1721535be5ead73a9e4c97b70237f121784bb0a032913737f9163e5" exitCode=0 Nov 28 13:20:45 crc kubenswrapper[4631]: I1128 13:20:45.545636 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"d3c2c475d1721535be5ead73a9e4c97b70237f121784bb0a032913737f9163e5"} Nov 28 13:20:45 crc kubenswrapper[4631]: I1128 13:20:45.545696 4631 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Nov 28 13:20:45 crc kubenswrapper[4631]: I1128 13:20:45.545701 4631 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 13:20:45 crc kubenswrapper[4631]: I1128 13:20:45.547206 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 13:20:45 crc kubenswrapper[4631]: I1128 13:20:45.547257 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 13:20:45 crc kubenswrapper[4631]: I1128 13:20:45.547272 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 13:20:45 crc kubenswrapper[4631]: I1128 13:20:45.550520 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"977b41b2843ab4377885a9b4bfa9173e8b247f528939c931f325b5868b5bc363"} Nov 28 13:20:45 crc kubenswrapper[4631]: I1128 13:20:45.550558 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"9e9de632f58c3b05196acde927ecb273de709345ee60661b24f4d97f21ef5659"} Nov 28 13:20:45 crc kubenswrapper[4631]: I1128 13:20:45.550574 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"d02db3bdc0a6daffb1fb930ed13b513eb41fdf380cdcb3cad24ea6f1fe9a6e19"} Nov 28 13:20:45 crc kubenswrapper[4631]: I1128 13:20:45.550714 4631 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 13:20:45 crc kubenswrapper[4631]: I1128 13:20:45.552126 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 13:20:45 crc kubenswrapper[4631]: I1128 13:20:45.552150 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 13:20:45 crc kubenswrapper[4631]: I1128 13:20:45.552161 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 13:20:45 crc kubenswrapper[4631]: I1128 13:20:45.552861 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"729ba9c465551eed47fb3442b11ce8ff0ddb7c40bcaeb3474a305dae44cd2834"} Nov 28 13:20:45 crc kubenswrapper[4631]: I1128 13:20:45.552951 4631 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 13:20:45 crc kubenswrapper[4631]: I1128 13:20:45.554554 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 13:20:45 crc kubenswrapper[4631]: I1128 13:20:45.554598 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 13:20:45 crc kubenswrapper[4631]: I1128 13:20:45.554617 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 13:20:45 crc kubenswrapper[4631]: I1128 13:20:45.556054 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"b7cd0089f761e1315fe18e82695a8cf9170990bb535ea7235ee39b68edd79f23"} Nov 28 13:20:45 crc kubenswrapper[4631]: I1128 13:20:45.556081 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"d396f0adc112df728f80c8e1167526cdf1458a21e4802e2b70104f1a68086ff0"} Nov 28 13:20:45 crc kubenswrapper[4631]: I1128 13:20:45.556092 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"f11c70496a148e22446931e8183b52f16bd431cc7bd17e9678f875df424d6a03"} Nov 28 13:20:45 crc kubenswrapper[4631]: I1128 13:20:45.556151 4631 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 13:20:45 crc kubenswrapper[4631]: I1128 13:20:45.556744 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 13:20:45 crc kubenswrapper[4631]: I1128 13:20:45.556767 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 13:20:45 crc kubenswrapper[4631]: I1128 13:20:45.556775 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 13:20:46 crc kubenswrapper[4631]: I1128 13:20:46.563737 4631 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="a619de981c98d25c93f008dcd2150247ca7a86b266248f9df2722781a44a8eb8" exitCode=0 Nov 28 13:20:46 crc kubenswrapper[4631]: I1128 13:20:46.563869 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"a619de981c98d25c93f008dcd2150247ca7a86b266248f9df2722781a44a8eb8"} Nov 28 13:20:46 crc kubenswrapper[4631]: I1128 13:20:46.563972 4631 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 13:20:46 crc kubenswrapper[4631]: I1128 13:20:46.565782 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 13:20:46 crc kubenswrapper[4631]: I1128 13:20:46.565845 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 13:20:46 crc kubenswrapper[4631]: I1128 13:20:46.565872 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 13:20:46 crc kubenswrapper[4631]: I1128 13:20:46.570819 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"6138e1e9121a843fd730def8c3f019dd68a6f38d963ddb00124f2c532762e012"} Nov 28 13:20:46 crc kubenswrapper[4631]: I1128 13:20:46.570862 4631 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 13:20:46 crc kubenswrapper[4631]: I1128 13:20:46.570892 4631 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 13:20:46 crc kubenswrapper[4631]: I1128 13:20:46.572749 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 13:20:46 crc kubenswrapper[4631]: I1128 13:20:46.572789 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 13:20:46 crc kubenswrapper[4631]: I1128 13:20:46.572826 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 13:20:46 crc kubenswrapper[4631]: I1128 13:20:46.572863 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 13:20:46 crc kubenswrapper[4631]: I1128 13:20:46.572830 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 13:20:46 crc kubenswrapper[4631]: I1128 13:20:46.572938 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 13:20:46 crc kubenswrapper[4631]: I1128 13:20:46.665090 4631 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 13:20:46 crc kubenswrapper[4631]: I1128 13:20:46.666818 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 13:20:46 crc kubenswrapper[4631]: I1128 13:20:46.666866 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 13:20:46 crc kubenswrapper[4631]: I1128 13:20:46.666881 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 13:20:46 crc kubenswrapper[4631]: I1128 13:20:46.666910 4631 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 28 13:20:46 crc kubenswrapper[4631]: I1128 13:20:46.915648 4631 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 13:20:47 crc kubenswrapper[4631]: I1128 13:20:47.290909 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 28 13:20:47 crc kubenswrapper[4631]: I1128 13:20:47.516555 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 28 13:20:47 crc kubenswrapper[4631]: I1128 13:20:47.516769 4631 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 13:20:47 crc kubenswrapper[4631]: I1128 13:20:47.518569 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 13:20:47 crc kubenswrapper[4631]: I1128 13:20:47.518739 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 13:20:47 crc kubenswrapper[4631]: I1128 13:20:47.518770 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 13:20:47 crc kubenswrapper[4631]: I1128 13:20:47.578961 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"4ac3bc7c2c1a70895b1e52886bee358501d6b403a84cd411ed66e35951590cd8"} Nov 28 13:20:47 crc kubenswrapper[4631]: I1128 13:20:47.579022 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"6f931f1aeb2cdc11d5b20adc19b5da6fda9763670ed4e7113c4e6d2fae512099"} Nov 28 13:20:47 crc kubenswrapper[4631]: I1128 13:20:47.579039 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"2e8c487eb87f69106af1ccc90b2a95c5882f9bc9080551594fb14b34d1cb3646"} Nov 28 13:20:47 crc kubenswrapper[4631]: I1128 13:20:47.579049 4631 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 13:20:47 crc kubenswrapper[4631]: I1128 13:20:47.579119 4631 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 13:20:47 crc kubenswrapper[4631]: I1128 13:20:47.579053 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"1b2911ca29670fcb2dc4179a8cc263e1d0bb9a22f55f8bd85056cd14d62fbc28"} Nov 28 13:20:47 crc kubenswrapper[4631]: I1128 13:20:47.579253 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 13:20:47 crc kubenswrapper[4631]: I1128 13:20:47.580306 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 13:20:47 crc kubenswrapper[4631]: I1128 13:20:47.580343 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 13:20:47 crc kubenswrapper[4631]: I1128 13:20:47.580357 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 13:20:47 crc kubenswrapper[4631]: I1128 13:20:47.580907 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 13:20:47 crc kubenswrapper[4631]: I1128 13:20:47.580942 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 13:20:47 crc kubenswrapper[4631]: I1128 13:20:47.580954 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 13:20:48 crc kubenswrapper[4631]: I1128 13:20:48.587131 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"20476713555c549d817c22a89fb8e6697d18cab8b156e25a045903b94cab9f8c"} Nov 28 13:20:48 crc kubenswrapper[4631]: I1128 13:20:48.587320 4631 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 13:20:48 crc kubenswrapper[4631]: I1128 13:20:48.587456 4631 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 13:20:48 crc kubenswrapper[4631]: I1128 13:20:48.588665 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 13:20:48 crc kubenswrapper[4631]: I1128 13:20:48.588703 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 13:20:48 crc kubenswrapper[4631]: I1128 13:20:48.588716 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 13:20:48 crc kubenswrapper[4631]: I1128 13:20:48.588701 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 13:20:48 crc kubenswrapper[4631]: I1128 13:20:48.588787 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 13:20:48 crc kubenswrapper[4631]: I1128 13:20:48.588802 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 13:20:48 crc kubenswrapper[4631]: I1128 13:20:48.662893 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 13:20:49 crc kubenswrapper[4631]: I1128 13:20:49.590062 4631 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 13:20:49 crc kubenswrapper[4631]: I1128 13:20:49.590100 4631 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 13:20:49 crc kubenswrapper[4631]: I1128 13:20:49.592020 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 13:20:49 crc kubenswrapper[4631]: I1128 13:20:49.592049 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 13:20:49 crc kubenswrapper[4631]: I1128 13:20:49.592080 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 13:20:49 crc kubenswrapper[4631]: I1128 13:20:49.592080 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 13:20:49 crc kubenswrapper[4631]: I1128 13:20:49.592101 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 13:20:49 crc kubenswrapper[4631]: I1128 13:20:49.592144 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 13:20:49 crc kubenswrapper[4631]: I1128 13:20:49.962216 4631 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 28 13:20:49 crc kubenswrapper[4631]: I1128 13:20:49.962708 4631 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 13:20:49 crc kubenswrapper[4631]: I1128 13:20:49.964900 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 13:20:49 crc kubenswrapper[4631]: I1128 13:20:49.964951 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 13:20:49 crc kubenswrapper[4631]: I1128 13:20:49.964962 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 13:20:50 crc kubenswrapper[4631]: I1128 13:20:50.232260 4631 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 28 13:20:50 crc kubenswrapper[4631]: I1128 13:20:50.241031 4631 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 28 13:20:50 crc kubenswrapper[4631]: I1128 13:20:50.593121 4631 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 13:20:50 crc kubenswrapper[4631]: I1128 13:20:50.594536 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 13:20:50 crc kubenswrapper[4631]: I1128 13:20:50.594572 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 13:20:50 crc kubenswrapper[4631]: I1128 13:20:50.594584 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 13:20:51 crc kubenswrapper[4631]: I1128 13:20:51.596415 4631 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 13:20:51 crc kubenswrapper[4631]: I1128 13:20:51.597973 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 13:20:51 crc kubenswrapper[4631]: I1128 13:20:51.598041 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 13:20:51 crc kubenswrapper[4631]: I1128 13:20:51.598063 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 13:20:52 crc kubenswrapper[4631]: I1128 13:20:52.195384 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Nov 28 13:20:52 crc kubenswrapper[4631]: I1128 13:20:52.195721 4631 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 13:20:52 crc kubenswrapper[4631]: I1128 13:20:52.197815 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 13:20:52 crc kubenswrapper[4631]: I1128 13:20:52.198064 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 13:20:52 crc kubenswrapper[4631]: I1128 13:20:52.200120 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 13:20:52 crc kubenswrapper[4631]: I1128 13:20:52.962376 4631 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Nov 28 13:20:52 crc kubenswrapper[4631]: I1128 13:20:52.963425 4631 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 28 13:20:53 crc kubenswrapper[4631]: E1128 13:20:53.566603 4631 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Nov 28 13:20:54 crc kubenswrapper[4631]: I1128 13:20:54.513460 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 28 13:20:54 crc kubenswrapper[4631]: I1128 13:20:54.513713 4631 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 13:20:54 crc kubenswrapper[4631]: I1128 13:20:54.514679 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 13:20:54 crc kubenswrapper[4631]: I1128 13:20:54.514704 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 13:20:54 crc kubenswrapper[4631]: I1128 13:20:54.514715 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 13:20:54 crc kubenswrapper[4631]: I1128 13:20:54.518378 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 28 13:20:54 crc kubenswrapper[4631]: I1128 13:20:54.608486 4631 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 13:20:54 crc kubenswrapper[4631]: I1128 13:20:54.609403 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 13:20:54 crc kubenswrapper[4631]: I1128 13:20:54.609444 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 13:20:54 crc kubenswrapper[4631]: I1128 13:20:54.609455 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 13:20:55 crc kubenswrapper[4631]: I1128 13:20:55.425568 4631 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": net/http: TLS handshake timeout Nov 28 13:20:55 crc kubenswrapper[4631]: E1128 13:20:55.548053 4631 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": net/http: TLS handshake timeout" logger="UnhandledError" Nov 28 13:20:56 crc kubenswrapper[4631]: E1128 13:20:56.433652 4631 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" interval="3.2s" Nov 28 13:20:56 crc kubenswrapper[4631]: I1128 13:20:56.520893 4631 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Nov 28 13:20:56 crc kubenswrapper[4631]: I1128 13:20:56.520945 4631 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Nov 28 13:20:56 crc kubenswrapper[4631]: I1128 13:20:56.529039 4631 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Nov 28 13:20:56 crc kubenswrapper[4631]: I1128 13:20:56.529106 4631 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Nov 28 13:20:56 crc kubenswrapper[4631]: I1128 13:20:56.922968 4631 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Nov 28 13:20:56 crc kubenswrapper[4631]: [+]log ok Nov 28 13:20:56 crc kubenswrapper[4631]: [+]etcd ok Nov 28 13:20:56 crc kubenswrapper[4631]: [+]poststarthook/start-apiserver-admission-initializer ok Nov 28 13:20:56 crc kubenswrapper[4631]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Nov 28 13:20:56 crc kubenswrapper[4631]: [+]poststarthook/openshift.io-api-request-count-filter ok Nov 28 13:20:56 crc kubenswrapper[4631]: [+]poststarthook/openshift.io-startkubeinformers ok Nov 28 13:20:56 crc kubenswrapper[4631]: [+]poststarthook/openshift.io-openshift-apiserver-reachable ok Nov 28 13:20:56 crc kubenswrapper[4631]: [+]poststarthook/openshift.io-oauth-apiserver-reachable ok Nov 28 13:20:56 crc kubenswrapper[4631]: [+]poststarthook/generic-apiserver-start-informers ok Nov 28 13:20:56 crc kubenswrapper[4631]: [+]poststarthook/priority-and-fairness-config-consumer ok Nov 28 13:20:56 crc kubenswrapper[4631]: [+]poststarthook/priority-and-fairness-filter ok Nov 28 13:20:56 crc kubenswrapper[4631]: [+]poststarthook/storage-object-count-tracker-hook ok Nov 28 13:20:56 crc kubenswrapper[4631]: [+]poststarthook/start-apiextensions-informers ok Nov 28 13:20:56 crc kubenswrapper[4631]: [+]poststarthook/start-apiextensions-controllers ok Nov 28 13:20:56 crc kubenswrapper[4631]: [+]poststarthook/crd-informer-synced ok Nov 28 13:20:56 crc kubenswrapper[4631]: [+]poststarthook/start-system-namespaces-controller ok Nov 28 13:20:56 crc kubenswrapper[4631]: [+]poststarthook/start-cluster-authentication-info-controller ok Nov 28 13:20:56 crc kubenswrapper[4631]: [+]poststarthook/start-kube-apiserver-identity-lease-controller ok Nov 28 13:20:56 crc kubenswrapper[4631]: [+]poststarthook/start-kube-apiserver-identity-lease-garbage-collector ok Nov 28 13:20:56 crc kubenswrapper[4631]: [+]poststarthook/start-legacy-token-tracking-controller ok Nov 28 13:20:56 crc kubenswrapper[4631]: [+]poststarthook/start-service-ip-repair-controllers ok Nov 28 13:20:56 crc kubenswrapper[4631]: [-]poststarthook/rbac/bootstrap-roles failed: reason withheld Nov 28 13:20:56 crc kubenswrapper[4631]: [-]poststarthook/scheduling/bootstrap-system-priority-classes failed: reason withheld Nov 28 13:20:56 crc kubenswrapper[4631]: [+]poststarthook/priority-and-fairness-config-producer ok Nov 28 13:20:56 crc kubenswrapper[4631]: [+]poststarthook/bootstrap-controller ok Nov 28 13:20:56 crc kubenswrapper[4631]: [+]poststarthook/aggregator-reload-proxy-client-cert ok Nov 28 13:20:56 crc kubenswrapper[4631]: [+]poststarthook/start-kube-aggregator-informers ok Nov 28 13:20:56 crc kubenswrapper[4631]: [+]poststarthook/apiservice-status-local-available-controller ok Nov 28 13:20:56 crc kubenswrapper[4631]: [+]poststarthook/apiservice-status-remote-available-controller ok Nov 28 13:20:56 crc kubenswrapper[4631]: [+]poststarthook/apiservice-registration-controller ok Nov 28 13:20:56 crc kubenswrapper[4631]: [+]poststarthook/apiservice-wait-for-first-sync ok Nov 28 13:20:56 crc kubenswrapper[4631]: [+]poststarthook/apiservice-discovery-controller ok Nov 28 13:20:56 crc kubenswrapper[4631]: [+]poststarthook/kube-apiserver-autoregistration ok Nov 28 13:20:56 crc kubenswrapper[4631]: [+]autoregister-completion ok Nov 28 13:20:56 crc kubenswrapper[4631]: [+]poststarthook/apiservice-openapi-controller ok Nov 28 13:20:56 crc kubenswrapper[4631]: [+]poststarthook/apiservice-openapiv3-controller ok Nov 28 13:20:56 crc kubenswrapper[4631]: livez check failed Nov 28 13:20:56 crc kubenswrapper[4631]: I1128 13:20:56.923054 4631 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 28 13:20:57 crc kubenswrapper[4631]: I1128 13:20:57.508493 4631 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Nov 28 13:20:57 crc kubenswrapper[4631]: I1128 13:20:57.508687 4631 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 13:20:57 crc kubenswrapper[4631]: I1128 13:20:57.509784 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 13:20:57 crc kubenswrapper[4631]: I1128 13:20:57.509834 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 13:20:57 crc kubenswrapper[4631]: I1128 13:20:57.509852 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 13:20:57 crc kubenswrapper[4631]: I1128 13:20:57.577769 4631 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Nov 28 13:20:57 crc kubenswrapper[4631]: I1128 13:20:57.614159 4631 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 13:20:57 crc kubenswrapper[4631]: I1128 13:20:57.615154 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 13:20:57 crc kubenswrapper[4631]: I1128 13:20:57.615182 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 13:20:57 crc kubenswrapper[4631]: I1128 13:20:57.615193 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 13:20:57 crc kubenswrapper[4631]: I1128 13:20:57.626409 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Nov 28 13:20:58 crc kubenswrapper[4631]: I1128 13:20:58.616249 4631 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 13:20:58 crc kubenswrapper[4631]: I1128 13:20:58.617431 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 13:20:58 crc kubenswrapper[4631]: I1128 13:20:58.617484 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 13:20:58 crc kubenswrapper[4631]: I1128 13:20:58.617496 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 13:20:59 crc kubenswrapper[4631]: I1128 13:20:59.945532 4631 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Nov 28 13:20:59 crc kubenswrapper[4631]: I1128 13:20:59.965146 4631 reflector.go:368] Caches populated for *v1.CertificateSigningRequest from k8s.io/client-go/tools/watch/informerwatcher.go:146 Nov 28 13:21:01 crc kubenswrapper[4631]: I1128 13:21:01.503153 4631 trace.go:236] Trace[790485157]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (28-Nov-2025 13:20:46.697) (total time: 14805ms): Nov 28 13:21:01 crc kubenswrapper[4631]: Trace[790485157]: ---"Objects listed" error: 14805ms (13:21:01.503) Nov 28 13:21:01 crc kubenswrapper[4631]: Trace[790485157]: [14.805460759s] [14.805460759s] END Nov 28 13:21:01 crc kubenswrapper[4631]: I1128 13:21:01.503587 4631 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Nov 28 13:21:01 crc kubenswrapper[4631]: I1128 13:21:01.503681 4631 trace.go:236] Trace[880148378]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (28-Nov-2025 13:20:47.359) (total time: 14144ms): Nov 28 13:21:01 crc kubenswrapper[4631]: Trace[880148378]: ---"Objects listed" error: 14144ms (13:21:01.503) Nov 28 13:21:01 crc kubenswrapper[4631]: Trace[880148378]: [14.144526257s] [14.144526257s] END Nov 28 13:21:01 crc kubenswrapper[4631]: I1128 13:21:01.503746 4631 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Nov 28 13:21:01 crc kubenswrapper[4631]: I1128 13:21:01.503171 4631 trace.go:236] Trace[614983929]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (28-Nov-2025 13:20:47.032) (total time: 14470ms): Nov 28 13:21:01 crc kubenswrapper[4631]: Trace[614983929]: ---"Objects listed" error: 14470ms (13:21:01.503) Nov 28 13:21:01 crc kubenswrapper[4631]: Trace[614983929]: [14.470929499s] [14.470929499s] END Nov 28 13:21:01 crc kubenswrapper[4631]: I1128 13:21:01.503815 4631 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Nov 28 13:21:01 crc kubenswrapper[4631]: E1128 13:21:01.506311 4631 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes \"crc\" is forbidden: autoscaling.openshift.io/ManagedNode infra config cache not synchronized" node="crc" Nov 28 13:21:01 crc kubenswrapper[4631]: I1128 13:21:01.507665 4631 trace.go:236] Trace[2098170355]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (28-Nov-2025 13:20:47.446) (total time: 14061ms): Nov 28 13:21:01 crc kubenswrapper[4631]: Trace[2098170355]: ---"Objects listed" error: 14061ms (13:21:01.507) Nov 28 13:21:01 crc kubenswrapper[4631]: Trace[2098170355]: [14.061194761s] [14.061194761s] END Nov 28 13:21:01 crc kubenswrapper[4631]: I1128 13:21:01.507696 4631 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Nov 28 13:21:01 crc kubenswrapper[4631]: I1128 13:21:01.509125 4631 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Nov 28 13:21:01 crc kubenswrapper[4631]: I1128 13:21:01.687435 4631 csr.go:261] certificate signing request csr-v8gvv is approved, waiting to be issued Nov 28 13:21:01 crc kubenswrapper[4631]: I1128 13:21:01.718186 4631 csr.go:257] certificate signing request csr-v8gvv is issued Nov 28 13:21:01 crc kubenswrapper[4631]: I1128 13:21:01.922597 4631 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 13:21:01 crc kubenswrapper[4631]: I1128 13:21:01.928788 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.211596 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.426547 4631 apiserver.go:52] "Watching apiserver" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.429452 4631 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.429731 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/iptables-alerter-4ln5h","openshift-network-operator/network-operator-58b4c7f79c-55gtf","openshift-kube-apiserver/kube-apiserver-crc","openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf"] Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.430021 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.430210 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 13:21:02 crc kubenswrapper[4631]: E1128 13:21:02.430269 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.430332 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 13:21:02 crc kubenswrapper[4631]: E1128 13:21:02.430354 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.430387 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 13:21:02 crc kubenswrapper[4631]: E1128 13:21:02.430405 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.430439 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.430735 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.433727 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.433770 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.433834 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.433834 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.433909 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.434025 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.434446 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.434604 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.541554 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.543501 4631 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.563160 4631 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"07b4914f-d3c8-4e6a-909c-3af046e1759c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:20:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:20:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:20:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4f323ff537cd79159a51b592401e92fc06d54c490ff005486f27f41d42a35f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T13:20:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eff71f12d67a15adda3183a742f2869b104d773904a6c5954d1ccb6a2c64dd71\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T13:20:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://53ea681ea681973dd5177ec7f2925ef48fec88ac004b41fca867ce671f5a76d1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T13:20:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6138e1e9121a843fd730def8c3f019dd68a6f38d963ddb00124f2c532762e012\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T13:20:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ac30b44d2fe8ddb4ede530d4b99dab4dc641a097e21d4ae69cfcbeddc38802d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T13:20:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0b61d56ed070810021d253875a7bb5b9cb94a3446c56c4a92fb5418022dd643\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f0b61d56ed070810021d253875a7bb5b9cb94a3446c56c4a92fb5418022dd643\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T13:20:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T13:20:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T13:20:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.575713 4631 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:02Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.584840 4631 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:02Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.592927 4631 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:02Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.600393 4631 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:02Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.608896 4631 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:02Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.621989 4631 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:02Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.622974 4631 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.633874 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 28 13:21:02 crc kubenswrapper[4631]: E1128 13:21:02.638986 4631 kubelet.go:1929] "Failed creating a mirror pod for" err="pods \"kube-apiserver-crc\" already exists" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.639371 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.641358 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.641506 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.641637 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.641743 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.641849 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.641965 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.642082 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.642190 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.642313 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.642033 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.642187 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.642193 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.642356 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.642405 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.642428 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.642485 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.642512 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.642534 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.642557 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.642579 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.642599 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.642644 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.642668 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.642687 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.642708 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.642751 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.642751 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.642773 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.642764 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.642793 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.642857 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.642876 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.642915 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.642952 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.642988 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.643026 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.643061 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.643089 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.643112 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.643146 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.643172 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.643203 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.643251 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.643341 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.643374 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.643405 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.643436 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.643466 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.643652 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.643686 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.643711 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.643737 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.643765 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.643837 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.643867 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.643894 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.643922 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.643958 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.643980 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.644004 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.644026 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.644050 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.644075 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.644100 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.644123 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.644145 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.644168 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.644196 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.644227 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.644259 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.644306 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.644330 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.644352 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.644409 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.642986 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.643015 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.643154 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.643277 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.643304 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.643400 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.643591 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.644015 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.644015 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.644064 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.644160 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.644207 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.644223 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.644329 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.644409 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.644423 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.644433 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.644543 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.644570 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.644594 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.644619 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.644643 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.644665 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.644681 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.644696 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.644711 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.644714 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.644728 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.644730 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.644745 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.644762 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.644777 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.644792 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.644844 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.644860 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.644876 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.644893 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.644891 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.644926 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.645161 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.645179 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.645197 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.645215 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.645206 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.645231 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.645275 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.645326 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.645353 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.645420 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.645445 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.645467 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.645493 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.645517 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.645550 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.645582 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.645611 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.645643 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.645653 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.645674 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.645707 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.645737 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.645768 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.645802 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.645839 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.645872 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.645896 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.645918 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.645941 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.645966 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.645987 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.646010 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.646032 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.646054 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.646076 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.646097 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.646120 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.646142 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.646164 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.646186 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.646220 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.646245 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.646268 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.646314 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.646357 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.646381 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.646404 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.646427 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.646452 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.646483 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.646516 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.646547 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.646584 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.646633 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.646667 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.646699 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.646731 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.646765 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.646808 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.646846 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.646880 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.646903 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.646928 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.646951 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.646974 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.646998 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.647021 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.647045 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.647068 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.647091 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.647115 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.647148 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.647172 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.647200 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.647235 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.647257 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.647282 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.647326 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.647350 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.647375 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.647399 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.647423 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.647446 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.647470 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.647493 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.647519 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.647553 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.647584 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.647608 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.647634 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.647659 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.647683 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.647707 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.647731 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.647755 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.647779 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.647807 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.647843 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.647869 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.647895 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.647920 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.647946 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.647968 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.647991 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.648014 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.648038 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.648061 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.648085 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.648110 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.648134 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.648157 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.648529 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.648591 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.648595 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.648784 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.648802 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.648808 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.648917 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.649019 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.649156 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.649358 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.649372 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.649500 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.649588 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.649635 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.649700 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.649742 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.649866 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.650021 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.650046 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.650184 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.650328 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.650473 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.650647 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.650803 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.650931 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.651005 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.651238 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.651625 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.651734 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.651931 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.652071 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.652235 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.652758 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.652777 4631 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:02Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.652965 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.653191 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.653254 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.653325 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.653332 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.653349 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.653621 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.653775 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.654381 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.654493 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.659212 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.659544 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.659827 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.660117 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.660316 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.660520 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.660689 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.662148 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.663756 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.663821 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.663864 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.663909 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.663942 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.663970 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.663995 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.664033 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.664056 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.664082 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.664127 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.664135 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.664548 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.665071 4631 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.673345 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.673401 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.674775 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.675111 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.677715 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.677911 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.678111 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.678267 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.678555 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.678568 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.678759 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.679444 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.679505 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.679687 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.679712 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.679873 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.680084 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.679884 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.680322 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:21:02 crc kubenswrapper[4631]: E1128 13:21:02.680545 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 13:21:03.180523112 +0000 UTC m=+19.987826446 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.680630 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.681506 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.681845 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.681987 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.682227 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.682421 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.682622 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.682807 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.683225 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.684599 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.684622 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.684963 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.685357 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.685427 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.685687 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.686528 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.687201 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.687690 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.688149 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.688490 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.688580 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.688763 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.689418 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.689881 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.690388 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.690486 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.690666 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.690755 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.691011 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.696396 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.696756 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.696799 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.696812 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.696890 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.696976 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.697049 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.697086 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.697099 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.697577 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.701500 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.701656 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.701677 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.701689 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:21:02 crc kubenswrapper[4631]: E1128 13:21:02.701841 4631 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 28 13:21:02 crc kubenswrapper[4631]: E1128 13:21:02.701931 4631 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 28 13:21:02 crc kubenswrapper[4631]: E1128 13:21:02.702016 4631 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.702100 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 13:21:02 crc kubenswrapper[4631]: E1128 13:21:02.702212 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-28 13:21:03.20219317 +0000 UTC m=+20.009496524 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.702724 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.702912 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.702940 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.703164 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.703373 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.703387 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.703592 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.703611 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.703817 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.703967 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.664133 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.704030 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.704057 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.704085 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.704114 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.704139 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.704165 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.704206 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.704274 4631 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.704309 4631 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.704323 4631 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.704338 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.704350 4631 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.704362 4631 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.704374 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.704386 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.704396 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.704408 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.704420 4631 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.704431 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.704443 4631 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.704454 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.704465 4631 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.704477 4631 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.704489 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.704501 4631 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.704512 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.704523 4631 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.704535 4631 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.704473 4631 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"07b4914f-d3c8-4e6a-909c-3af046e1759c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:20:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:20:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:20:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4f323ff537cd79159a51b592401e92fc06d54c490ff005486f27f41d42a35f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T13:20:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eff71f12d67a15adda3183a742f2869b104d773904a6c5954d1ccb6a2c64dd71\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T13:20:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://53ea681ea681973dd5177ec7f2925ef48fec88ac004b41fca867ce671f5a76d1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T13:20:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6138e1e9121a843fd730def8c3f019dd68a6f38d963ddb00124f2c532762e012\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T13:20:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ac30b44d2fe8ddb4ede530d4b99dab4dc641a097e21d4ae69cfcbeddc38802d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T13:20:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0b61d56ed070810021d253875a7bb5b9cb94a3446c56c4a92fb5418022dd643\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f0b61d56ed070810021d253875a7bb5b9cb94a3446c56c4a92fb5418022dd643\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T13:20:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T13:20:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T13:20:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.704546 4631 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.704684 4631 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.704698 4631 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.704711 4631 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.704723 4631 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.704735 4631 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.704747 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.704759 4631 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.704770 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.704781 4631 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.704792 4631 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.704804 4631 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.704815 4631 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.704825 4631 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.704837 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.704848 4631 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.704859 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.704870 4631 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.704881 4631 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.704892 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.704904 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.704915 4631 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.704927 4631 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.704931 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.704938 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.704950 4631 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.704961 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.704975 4631 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.704986 4631 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.705001 4631 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.705013 4631 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Nov 28 13:21:02 crc kubenswrapper[4631]: E1128 13:21:02.704782 4631 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 28 13:21:02 crc kubenswrapper[4631]: E1128 13:21:02.705072 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-28 13:21:03.20505469 +0000 UTC m=+20.012358024 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.705025 4631 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.705100 4631 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.705980 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.706737 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.706896 4631 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.706917 4631 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.706930 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.706942 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.706956 4631 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.706964 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:21:02 crc kubenswrapper[4631]: E1128 13:21:02.705008 4631 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.706968 4631 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.706991 4631 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.707001 4631 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 28 13:21:02 crc kubenswrapper[4631]: E1128 13:21:02.707019 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-28 13:21:03.20700462 +0000 UTC m=+20.014307964 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.707036 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.707050 4631 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.707062 4631 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.707072 4631 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.707082 4631 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.707124 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.707138 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.707150 4631 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.707161 4631 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.707173 4631 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.707185 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.707196 4631 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.707208 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.707218 4631 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.707232 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.707243 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.707258 4631 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.707270 4631 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.707282 4631 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.707311 4631 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.707322 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.707336 4631 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.707348 4631 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.707360 4631 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.707371 4631 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.707381 4631 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.707392 4631 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.707403 4631 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.707416 4631 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.707427 4631 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.707439 4631 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.707449 4631 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.707462 4631 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.707473 4631 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.707484 4631 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.707495 4631 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.707505 4631 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.707516 4631 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.707527 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.707537 4631 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.707546 4631 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.707555 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.707565 4631 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.707575 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.707585 4631 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.707595 4631 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.707607 4631 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.707632 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.707644 4631 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.707656 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.707667 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.707678 4631 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.707691 4631 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.707703 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.707714 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.707726 4631 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.707738 4631 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.707749 4631 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.707762 4631 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.707775 4631 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.707786 4631 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.707798 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.707811 4631 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.707823 4631 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.707835 4631 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.707847 4631 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.707857 4631 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.707867 4631 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.707878 4631 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.707889 4631 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.707900 4631 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.707911 4631 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.707921 4631 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.707932 4631 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.707942 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.707952 4631 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.707962 4631 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.707973 4631 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.707984 4631 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.707995 4631 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.708007 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.708018 4631 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.708028 4631 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.708039 4631 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.708051 4631 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.708061 4631 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.708073 4631 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.708084 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.708096 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.708109 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.708119 4631 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.708130 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.708142 4631 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.708153 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.708182 4631 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.708196 4631 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.708209 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.708360 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.708397 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.703823 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.712201 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.712359 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.712394 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.712468 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 13:21:02 crc kubenswrapper[4631]: E1128 13:21:02.712477 4631 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.712492 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 13:21:02 crc kubenswrapper[4631]: E1128 13:21:02.712498 4631 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 28 13:21:02 crc kubenswrapper[4631]: E1128 13:21:02.712515 4631 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.712538 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:21:02 crc kubenswrapper[4631]: E1128 13:21:02.712590 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-28 13:21:03.212570466 +0000 UTC m=+20.019873900 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.716449 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.716679 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.717309 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.719190 4631 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-11-28 13:16:01 +0000 UTC, rotation deadline is 2026-10-22 05:13:22.099975949 +0000 UTC Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.719213 4631 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 7863h52m19.380765696s for next certificate rotation Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.724682 4631 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:02Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.727640 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.727712 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.729498 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.735319 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.735856 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.743596 4631 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"07b4914f-d3c8-4e6a-909c-3af046e1759c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:20:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:20:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:20:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4f323ff537cd79159a51b592401e92fc06d54c490ff005486f27f41d42a35f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T13:20:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eff71f12d67a15adda3183a742f2869b104d773904a6c5954d1ccb6a2c64dd71\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T13:20:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://53ea681ea681973dd5177ec7f2925ef48fec88ac004b41fca867ce671f5a76d1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T13:20:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6138e1e9121a843fd730def8c3f019dd68a6f38d963ddb00124f2c532762e012\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T13:20:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ac30b44d2fe8ddb4ede530d4b99dab4dc641a097e21d4ae69cfcbeddc38802d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T13:20:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0b61d56ed070810021d253875a7bb5b9cb94a3446c56c4a92fb5418022dd643\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f0b61d56ed070810021d253875a7bb5b9cb94a3446c56c4a92fb5418022dd643\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T13:20:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T13:20:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T13:20:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.745162 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.745369 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.745796 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.745972 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.746371 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.747355 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.747503 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.747657 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.747886 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.748674 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.749472 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.749635 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.749925 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.750159 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.751115 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.752484 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.753845 4631 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:02Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.757439 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.762702 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.766895 4631 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:02Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.775836 4631 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:02Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.784157 4631 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:02Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.792775 4631 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6d7f49b-96d5-4d70-a0e3-9d9f37f59073\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:20:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:20:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:20:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f11c70496a148e22446931e8183b52f16bd431cc7bd17e9678f875df424d6a03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T13:20:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35d912d041da9a7ed307ebea6e999cf087dacf84a3cefbfd0f3a9c891c979408\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T13:20:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d396f0adc112df728f80c8e1167526cdf1458a21e4802e2b70104f1a68086ff0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T13:20:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7cd0089f761e1315fe18e82695a8cf9170990bb535ea7235ee39b68edd79f23\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T13:20:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T13:20:43Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.807096 4631 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:02Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.809271 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.809327 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.809377 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.809386 4631 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.809395 4631 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.809403 4631 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.809411 4631 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.809419 4631 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.809427 4631 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.809435 4631 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.809443 4631 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.809450 4631 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.809460 4631 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.809468 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.809476 4631 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.809485 4631 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.809492 4631 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.809501 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.809508 4631 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.809516 4631 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.809523 4631 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.809530 4631 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.809538 4631 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.809546 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.809554 4631 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.809563 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.809571 4631 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.809599 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.809607 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.809615 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.809622 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.809631 4631 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.809639 4631 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.809647 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.809655 4631 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.809663 4631 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.809698 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.809801 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.817337 4631 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:02Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 28 13:21:02 crc kubenswrapper[4631]: I1128 13:21:02.852559 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 28 13:21:03 crc kubenswrapper[4631]: I1128 13:21:03.042657 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 28 13:21:03 crc kubenswrapper[4631]: I1128 13:21:03.213061 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 13:21:03 crc kubenswrapper[4631]: I1128 13:21:03.213151 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 13:21:03 crc kubenswrapper[4631]: I1128 13:21:03.213173 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 13:21:03 crc kubenswrapper[4631]: E1128 13:21:03.213207 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 13:21:04.213178976 +0000 UTC m=+21.020482320 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 13:21:03 crc kubenswrapper[4631]: E1128 13:21:03.213231 4631 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 28 13:21:03 crc kubenswrapper[4631]: I1128 13:21:03.213256 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 13:21:03 crc kubenswrapper[4631]: I1128 13:21:03.213322 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 13:21:03 crc kubenswrapper[4631]: E1128 13:21:03.213374 4631 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 28 13:21:03 crc kubenswrapper[4631]: E1128 13:21:03.213420 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-28 13:21:04.213402531 +0000 UTC m=+21.020705975 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 28 13:21:03 crc kubenswrapper[4631]: E1128 13:21:03.213421 4631 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 28 13:21:03 crc kubenswrapper[4631]: E1128 13:21:03.213462 4631 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 28 13:21:03 crc kubenswrapper[4631]: E1128 13:21:03.213476 4631 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 28 13:21:03 crc kubenswrapper[4631]: E1128 13:21:03.213435 4631 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 28 13:21:03 crc kubenswrapper[4631]: E1128 13:21:03.213521 4631 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 28 13:21:03 crc kubenswrapper[4631]: E1128 13:21:03.213447 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-28 13:21:04.213439302 +0000 UTC m=+21.020742646 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 28 13:21:03 crc kubenswrapper[4631]: E1128 13:21:03.213536 4631 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 28 13:21:03 crc kubenswrapper[4631]: E1128 13:21:03.213554 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-28 13:21:04.213537034 +0000 UTC m=+21.020840378 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 28 13:21:03 crc kubenswrapper[4631]: E1128 13:21:03.213578 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-28 13:21:04.213571635 +0000 UTC m=+21.020874979 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 28 13:21:03 crc kubenswrapper[4631]: I1128 13:21:03.359238 4631 transport.go:147] "Certificate rotation detected, shutting down client connections to start using new credentials" Nov 28 13:21:03 crc kubenswrapper[4631]: W1128 13:21:03.359565 4631 reflector.go:484] object-"openshift-network-node-identity"/"kube-root-ca.crt": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-network-node-identity"/"kube-root-ca.crt": Unexpected watch close - watch lasted less than a second and no items received Nov 28 13:21:03 crc kubenswrapper[4631]: W1128 13:21:03.359625 4631 reflector.go:484] object-"openshift-network-node-identity"/"env-overrides": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-network-node-identity"/"env-overrides": Unexpected watch close - watch lasted less than a second and no items received Nov 28 13:21:03 crc kubenswrapper[4631]: W1128 13:21:03.359675 4631 reflector.go:484] object-"openshift-network-operator"/"iptables-alerter-script": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-network-operator"/"iptables-alerter-script": Unexpected watch close - watch lasted less than a second and no items received Nov 28 13:21:03 crc kubenswrapper[4631]: W1128 13:21:03.359708 4631 reflector.go:484] object-"openshift-network-operator"/"kube-root-ca.crt": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-network-operator"/"kube-root-ca.crt": Unexpected watch close - watch lasted less than a second and no items received Nov 28 13:21:03 crc kubenswrapper[4631]: W1128 13:21:03.360306 4631 reflector.go:484] object-"openshift-network-node-identity"/"ovnkube-identity-cm": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-network-node-identity"/"ovnkube-identity-cm": Unexpected watch close - watch lasted less than a second and no items received Nov 28 13:21:03 crc kubenswrapper[4631]: W1128 13:21:03.360349 4631 reflector.go:484] object-"openshift-network-operator"/"openshift-service-ca.crt": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-network-operator"/"openshift-service-ca.crt": Unexpected watch close - watch lasted less than a second and no items received Nov 28 13:21:03 crc kubenswrapper[4631]: W1128 13:21:03.360358 4631 reflector.go:484] object-"openshift-network-node-identity"/"network-node-identity-cert": watch of *v1.Secret ended with: very short watch: object-"openshift-network-node-identity"/"network-node-identity-cert": Unexpected watch close - watch lasted less than a second and no items received Nov 28 13:21:03 crc kubenswrapper[4631]: W1128 13:21:03.360399 4631 reflector.go:484] object-"openshift-network-operator"/"metrics-tls": watch of *v1.Secret ended with: very short watch: object-"openshift-network-operator"/"metrics-tls": Unexpected watch close - watch lasted less than a second and no items received Nov 28 13:21:03 crc kubenswrapper[4631]: W1128 13:21:03.360375 4631 reflector.go:484] object-"openshift-network-node-identity"/"openshift-service-ca.crt": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-network-node-identity"/"openshift-service-ca.crt": Unexpected watch close - watch lasted less than a second and no items received Nov 28 13:21:03 crc kubenswrapper[4631]: I1128 13:21:03.516047 4631 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Nov 28 13:21:03 crc kubenswrapper[4631]: I1128 13:21:03.516780 4631 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Nov 28 13:21:03 crc kubenswrapper[4631]: I1128 13:21:03.517479 4631 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Nov 28 13:21:03 crc kubenswrapper[4631]: I1128 13:21:03.518123 4631 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Nov 28 13:21:03 crc kubenswrapper[4631]: I1128 13:21:03.518952 4631 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Nov 28 13:21:03 crc kubenswrapper[4631]: I1128 13:21:03.519581 4631 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Nov 28 13:21:03 crc kubenswrapper[4631]: I1128 13:21:03.520364 4631 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Nov 28 13:21:03 crc kubenswrapper[4631]: I1128 13:21:03.521076 4631 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Nov 28 13:21:03 crc kubenswrapper[4631]: I1128 13:21:03.521896 4631 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Nov 28 13:21:03 crc kubenswrapper[4631]: I1128 13:21:03.522617 4631 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Nov 28 13:21:03 crc kubenswrapper[4631]: I1128 13:21:03.523263 4631 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Nov 28 13:21:03 crc kubenswrapper[4631]: I1128 13:21:03.524151 4631 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Nov 28 13:21:03 crc kubenswrapper[4631]: I1128 13:21:03.524840 4631 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Nov 28 13:21:03 crc kubenswrapper[4631]: I1128 13:21:03.525471 4631 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Nov 28 13:21:03 crc kubenswrapper[4631]: I1128 13:21:03.526100 4631 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Nov 28 13:21:03 crc kubenswrapper[4631]: I1128 13:21:03.526040 4631 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6d7f49b-96d5-4d70-a0e3-9d9f37f59073\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:20:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:20:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:20:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f11c70496a148e22446931e8183b52f16bd431cc7bd17e9678f875df424d6a03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T13:20:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35d912d041da9a7ed307ebea6e999cf087dacf84a3cefbfd0f3a9c891c979408\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T13:20:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d396f0adc112df728f80c8e1167526cdf1458a21e4802e2b70104f1a68086ff0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T13:20:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7cd0089f761e1315fe18e82695a8cf9170990bb535ea7235ee39b68edd79f23\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T13:20:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T13:20:43Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 28 13:21:03 crc kubenswrapper[4631]: I1128 13:21:03.526738 4631 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Nov 28 13:21:03 crc kubenswrapper[4631]: I1128 13:21:03.527423 4631 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Nov 28 13:21:03 crc kubenswrapper[4631]: I1128 13:21:03.536973 4631 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:02Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 28 13:21:03 crc kubenswrapper[4631]: I1128 13:21:03.550930 4631 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:02Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 28 13:21:03 crc kubenswrapper[4631]: I1128 13:21:03.572002 4631 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:02Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 28 13:21:03 crc kubenswrapper[4631]: I1128 13:21:03.592738 4631 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:02Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 28 13:21:03 crc kubenswrapper[4631]: I1128 13:21:03.602574 4631 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Nov 28 13:21:03 crc kubenswrapper[4631]: I1128 13:21:03.610022 4631 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"07b4914f-d3c8-4e6a-909c-3af046e1759c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:20:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:20:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:20:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4f323ff537cd79159a51b592401e92fc06d54c490ff005486f27f41d42a35f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T13:20:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eff71f12d67a15adda3183a742f2869b104d773904a6c5954d1ccb6a2c64dd71\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T13:20:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://53ea681ea681973dd5177ec7f2925ef48fec88ac004b41fca867ce671f5a76d1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T13:20:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6138e1e9121a843fd730def8c3f019dd68a6f38d963ddb00124f2c532762e012\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T13:20:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ac30b44d2fe8ddb4ede530d4b99dab4dc641a097e21d4ae69cfcbeddc38802d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T13:20:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0b61d56ed070810021d253875a7bb5b9cb94a3446c56c4a92fb5418022dd643\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f0b61d56ed070810021d253875a7bb5b9cb94a3446c56c4a92fb5418022dd643\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T13:20:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T13:20:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T13:20:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 28 13:21:03 crc kubenswrapper[4631]: I1128 13:21:03.625422 4631 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:02Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 28 13:21:03 crc kubenswrapper[4631]: I1128 13:21:03.642707 4631 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:02Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 28 13:21:04 crc kubenswrapper[4631]: I1128 13:21:04.223770 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 13:21:04 crc kubenswrapper[4631]: I1128 13:21:04.223898 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 13:21:04 crc kubenswrapper[4631]: I1128 13:21:04.223952 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 13:21:04 crc kubenswrapper[4631]: I1128 13:21:04.224001 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 13:21:04 crc kubenswrapper[4631]: I1128 13:21:04.224035 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 13:21:04 crc kubenswrapper[4631]: E1128 13:21:04.224151 4631 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 28 13:21:04 crc kubenswrapper[4631]: E1128 13:21:04.224230 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-28 13:21:06.224205561 +0000 UTC m=+23.031508945 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 28 13:21:04 crc kubenswrapper[4631]: E1128 13:21:04.224538 4631 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 28 13:21:04 crc kubenswrapper[4631]: E1128 13:21:04.224615 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 13:21:06.224573038 +0000 UTC m=+23.031876422 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 13:21:04 crc kubenswrapper[4631]: E1128 13:21:04.224744 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-28 13:21:06.224720611 +0000 UTC m=+23.032023955 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 28 13:21:04 crc kubenswrapper[4631]: E1128 13:21:04.224815 4631 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 28 13:21:04 crc kubenswrapper[4631]: E1128 13:21:04.224869 4631 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 28 13:21:04 crc kubenswrapper[4631]: E1128 13:21:04.224893 4631 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 28 13:21:04 crc kubenswrapper[4631]: E1128 13:21:04.224959 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-28 13:21:06.224935166 +0000 UTC m=+23.032238550 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 28 13:21:04 crc kubenswrapper[4631]: E1128 13:21:04.224821 4631 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 28 13:21:04 crc kubenswrapper[4631]: E1128 13:21:04.224999 4631 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 28 13:21:04 crc kubenswrapper[4631]: E1128 13:21:04.225015 4631 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 28 13:21:04 crc kubenswrapper[4631]: E1128 13:21:04.225069 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-28 13:21:06.225056938 +0000 UTC m=+23.032360322 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 28 13:21:04 crc kubenswrapper[4631]: I1128 13:21:04.376825 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Nov 28 13:21:04 crc kubenswrapper[4631]: I1128 13:21:04.554865 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Nov 28 13:21:04 crc kubenswrapper[4631]: I1128 13:21:04.583979 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Nov 28 13:21:04 crc kubenswrapper[4631]: I1128 13:21:04.636557 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Nov 28 13:21:04 crc kubenswrapper[4631]: I1128 13:21:04.652405 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Nov 28 13:21:04 crc kubenswrapper[4631]: I1128 13:21:04.706646 4631 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 13:21:04 crc kubenswrapper[4631]: I1128 13:21:04.708157 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 13:21:04 crc kubenswrapper[4631]: I1128 13:21:04.708218 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 13:21:04 crc kubenswrapper[4631]: I1128 13:21:04.708237 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 13:21:04 crc kubenswrapper[4631]: I1128 13:21:04.708372 4631 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 28 13:21:04 crc kubenswrapper[4631]: I1128 13:21:04.716701 4631 kubelet_node_status.go:115] "Node was previously registered" node="crc" Nov 28 13:21:04 crc kubenswrapper[4631]: I1128 13:21:04.716775 4631 kubelet_node_status.go:79] "Successfully registered node" node="crc" Nov 28 13:21:04 crc kubenswrapper[4631]: I1128 13:21:04.717795 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 13:21:04 crc kubenswrapper[4631]: I1128 13:21:04.717845 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 13:21:04 crc kubenswrapper[4631]: I1128 13:21:04.717862 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 13:21:04 crc kubenswrapper[4631]: I1128 13:21:04.717884 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 13:21:04 crc kubenswrapper[4631]: I1128 13:21:04.717904 4631 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T13:21:04Z","lastTransitionTime":"2025-11-28T13:21:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 13:21:04 crc kubenswrapper[4631]: I1128 13:21:04.737871 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Nov 28 13:21:04 crc kubenswrapper[4631]: E1128 13:21:04.740512 4631 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148060Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608860Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T13:21:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T13:21:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:04Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T13:21:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T13:21:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:04Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4d70419e-7663-4d3f-a15c-3c3ee0ecadaa\\\",\\\"systemUUID\\\":\\\"0670c19c-6aaa-48af-8d2a-ee47e63640da\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 28 13:21:04 crc kubenswrapper[4631]: I1128 13:21:04.744789 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 13:21:04 crc kubenswrapper[4631]: I1128 13:21:04.744846 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 13:21:04 crc kubenswrapper[4631]: I1128 13:21:04.744866 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 13:21:04 crc kubenswrapper[4631]: I1128 13:21:04.744890 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 13:21:04 crc kubenswrapper[4631]: I1128 13:21:04.744908 4631 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T13:21:04Z","lastTransitionTime":"2025-11-28T13:21:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 13:21:04 crc kubenswrapper[4631]: E1128 13:21:04.758226 4631 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148060Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608860Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T13:21:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T13:21:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:04Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T13:21:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T13:21:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:04Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4d70419e-7663-4d3f-a15c-3c3ee0ecadaa\\\",\\\"systemUUID\\\":\\\"0670c19c-6aaa-48af-8d2a-ee47e63640da\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 28 13:21:04 crc kubenswrapper[4631]: I1128 13:21:04.762473 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 13:21:04 crc kubenswrapper[4631]: I1128 13:21:04.762496 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 13:21:04 crc kubenswrapper[4631]: I1128 13:21:04.762505 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 13:21:04 crc kubenswrapper[4631]: I1128 13:21:04.762516 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 13:21:04 crc kubenswrapper[4631]: I1128 13:21:04.762525 4631 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T13:21:04Z","lastTransitionTime":"2025-11-28T13:21:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 13:21:04 crc kubenswrapper[4631]: E1128 13:21:04.777653 4631 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148060Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608860Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T13:21:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T13:21:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:04Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T13:21:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T13:21:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:04Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4d70419e-7663-4d3f-a15c-3c3ee0ecadaa\\\",\\\"systemUUID\\\":\\\"0670c19c-6aaa-48af-8d2a-ee47e63640da\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 28 13:21:04 crc kubenswrapper[4631]: I1128 13:21:04.779675 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 28 13:21:04 crc kubenswrapper[4631]: I1128 13:21:04.792365 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 13:21:04 crc kubenswrapper[4631]: I1128 13:21:04.792397 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 13:21:04 crc kubenswrapper[4631]: I1128 13:21:04.792409 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 13:21:04 crc kubenswrapper[4631]: I1128 13:21:04.792423 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 13:21:04 crc kubenswrapper[4631]: I1128 13:21:04.792434 4631 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T13:21:04Z","lastTransitionTime":"2025-11-28T13:21:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 13:21:04 crc kubenswrapper[4631]: I1128 13:21:04.800091 4631 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Nov 28 13:21:04 crc kubenswrapper[4631]: I1128 13:21:04.801312 4631 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Nov 28 13:21:04 crc kubenswrapper[4631]: I1128 13:21:04.801994 4631 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Nov 28 13:21:04 crc kubenswrapper[4631]: I1128 13:21:04.802838 4631 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Nov 28 13:21:04 crc kubenswrapper[4631]: I1128 13:21:04.803626 4631 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Nov 28 13:21:04 crc kubenswrapper[4631]: I1128 13:21:04.804381 4631 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Nov 28 13:21:04 crc kubenswrapper[4631]: I1128 13:21:04.807283 4631 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Nov 28 13:21:04 crc kubenswrapper[4631]: I1128 13:21:04.812356 4631 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Nov 28 13:21:04 crc kubenswrapper[4631]: E1128 13:21:04.836205 4631 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148060Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608860Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T13:21:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T13:21:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:04Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T13:21:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T13:21:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:04Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4d70419e-7663-4d3f-a15c-3c3ee0ecadaa\\\",\\\"systemUUID\\\":\\\"0670c19c-6aaa-48af-8d2a-ee47e63640da\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 28 13:21:04 crc kubenswrapper[4631]: I1128 13:21:04.836535 4631 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Nov 28 13:21:04 crc kubenswrapper[4631]: I1128 13:21:04.838191 4631 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Nov 28 13:21:04 crc kubenswrapper[4631]: I1128 13:21:04.839510 4631 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Nov 28 13:21:04 crc kubenswrapper[4631]: I1128 13:21:04.844072 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 13:21:04 crc kubenswrapper[4631]: I1128 13:21:04.844708 4631 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Nov 28 13:21:04 crc kubenswrapper[4631]: I1128 13:21:04.845537 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 13:21:04 crc kubenswrapper[4631]: I1128 13:21:04.845573 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 13:21:04 crc kubenswrapper[4631]: I1128 13:21:04.845585 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 13:21:04 crc kubenswrapper[4631]: I1128 13:21:04.845601 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 13:21:04 crc kubenswrapper[4631]: I1128 13:21:04.845611 4631 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T13:21:04Z","lastTransitionTime":"2025-11-28T13:21:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 13:21:04 crc kubenswrapper[4631]: I1128 13:21:04.847772 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 28 13:21:04 crc kubenswrapper[4631]: I1128 13:21:04.857062 4631 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Nov 28 13:21:04 crc kubenswrapper[4631]: E1128 13:21:04.862070 4631 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148060Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608860Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T13:21:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T13:21:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:04Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T13:21:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T13:21:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:04Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4d70419e-7663-4d3f-a15c-3c3ee0ecadaa\\\",\\\"systemUUID\\\":\\\"0670c19c-6aaa-48af-8d2a-ee47e63640da\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 28 13:21:04 crc kubenswrapper[4631]: E1128 13:21:04.862174 4631 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 28 13:21:04 crc kubenswrapper[4631]: I1128 13:21:04.863707 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 13:21:04 crc kubenswrapper[4631]: I1128 13:21:04.864034 4631 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Nov 28 13:21:04 crc kubenswrapper[4631]: I1128 13:21:04.866524 4631 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Nov 28 13:21:04 crc kubenswrapper[4631]: I1128 13:21:04.869790 4631 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Nov 28 13:21:04 crc kubenswrapper[4631]: I1128 13:21:04.871476 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 13:21:04 crc kubenswrapper[4631]: I1128 13:21:04.871506 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 13:21:04 crc kubenswrapper[4631]: I1128 13:21:04.871515 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 13:21:04 crc kubenswrapper[4631]: I1128 13:21:04.871529 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 13:21:04 crc kubenswrapper[4631]: I1128 13:21:04.871538 4631 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T13:21:04Z","lastTransitionTime":"2025-11-28T13:21:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 13:21:04 crc kubenswrapper[4631]: I1128 13:21:04.873560 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Nov 28 13:21:04 crc kubenswrapper[4631]: I1128 13:21:04.884755 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Nov 28 13:21:04 crc kubenswrapper[4631]: I1128 13:21:04.888625 4631 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Nov 28 13:21:04 crc kubenswrapper[4631]: I1128 13:21:04.894165 4631 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Nov 28 13:21:04 crc kubenswrapper[4631]: I1128 13:21:04.897309 4631 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Nov 28 13:21:04 crc kubenswrapper[4631]: I1128 13:21:04.898350 4631 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Nov 28 13:21:04 crc kubenswrapper[4631]: I1128 13:21:04.899849 4631 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Nov 28 13:21:04 crc kubenswrapper[4631]: I1128 13:21:04.900516 4631 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Nov 28 13:21:04 crc kubenswrapper[4631]: I1128 13:21:04.901602 4631 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Nov 28 13:21:04 crc kubenswrapper[4631]: I1128 13:21:04.909018 4631 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Nov 28 13:21:04 crc kubenswrapper[4631]: I1128 13:21:04.909776 4631 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Nov 28 13:21:04 crc kubenswrapper[4631]: I1128 13:21:04.910506 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Nov 28 13:21:04 crc kubenswrapper[4631]: I1128 13:21:04.910764 4631 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Nov 28 13:21:04 crc kubenswrapper[4631]: I1128 13:21:04.911416 4631 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Nov 28 13:21:04 crc kubenswrapper[4631]: I1128 13:21:04.912437 4631 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Nov 28 13:21:04 crc kubenswrapper[4631]: I1128 13:21:04.913366 4631 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Nov 28 13:21:04 crc kubenswrapper[4631]: I1128 13:21:04.914248 4631 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Nov 28 13:21:04 crc kubenswrapper[4631]: I1128 13:21:04.914719 4631 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Nov 28 13:21:04 crc kubenswrapper[4631]: I1128 13:21:04.915222 4631 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Nov 28 13:21:04 crc kubenswrapper[4631]: I1128 13:21:04.918111 4631 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Nov 28 13:21:04 crc kubenswrapper[4631]: I1128 13:21:04.918769 4631 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Nov 28 13:21:04 crc kubenswrapper[4631]: I1128 13:21:04.921603 4631 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Nov 28 13:21:04 crc kubenswrapper[4631]: E1128 13:21:04.922520 4631 kubelet.go:2526] "Housekeeping took longer than expected" err="housekeeping took too long" expected="1s" actual="1.41s" Nov 28 13:21:04 crc kubenswrapper[4631]: I1128 13:21:04.922668 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 13:21:04 crc kubenswrapper[4631]: I1128 13:21:04.922717 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 13:21:04 crc kubenswrapper[4631]: I1128 13:21:04.922769 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 13:21:04 crc kubenswrapper[4631]: E1128 13:21:04.922775 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 13:21:04 crc kubenswrapper[4631]: E1128 13:21:04.922854 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 13:21:04 crc kubenswrapper[4631]: E1128 13:21:04.922925 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 13:21:04 crc kubenswrapper[4631]: I1128 13:21:04.929535 4631 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 28 13:21:04 crc kubenswrapper[4631]: I1128 13:21:04.929562 4631 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Nov 28 13:21:04 crc kubenswrapper[4631]: I1128 13:21:04.973961 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 13:21:04 crc kubenswrapper[4631]: I1128 13:21:04.974000 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 13:21:04 crc kubenswrapper[4631]: I1128 13:21:04.974032 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 13:21:04 crc kubenswrapper[4631]: I1128 13:21:04.974052 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 13:21:04 crc kubenswrapper[4631]: I1128 13:21:04.974064 4631 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T13:21:04Z","lastTransitionTime":"2025-11-28T13:21:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 13:21:05 crc kubenswrapper[4631]: I1128 13:21:05.076590 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 13:21:05 crc kubenswrapper[4631]: I1128 13:21:05.076637 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 13:21:05 crc kubenswrapper[4631]: I1128 13:21:05.076653 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 13:21:05 crc kubenswrapper[4631]: I1128 13:21:05.076673 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 13:21:05 crc kubenswrapper[4631]: I1128 13:21:05.076689 4631 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T13:21:05Z","lastTransitionTime":"2025-11-28T13:21:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 13:21:05 crc kubenswrapper[4631]: I1128 13:21:05.178701 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 13:21:05 crc kubenswrapper[4631]: I1128 13:21:05.178739 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 13:21:05 crc kubenswrapper[4631]: I1128 13:21:05.178747 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 13:21:05 crc kubenswrapper[4631]: I1128 13:21:05.178760 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 13:21:05 crc kubenswrapper[4631]: I1128 13:21:05.178770 4631 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T13:21:05Z","lastTransitionTime":"2025-11-28T13:21:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 13:21:05 crc kubenswrapper[4631]: I1128 13:21:05.239404 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-wbkjm"] Nov 28 13:21:05 crc kubenswrapper[4631]: I1128 13:21:05.239782 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-wbkjm" Nov 28 13:21:05 crc kubenswrapper[4631]: I1128 13:21:05.241690 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Nov 28 13:21:05 crc kubenswrapper[4631]: I1128 13:21:05.242185 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Nov 28 13:21:05 crc kubenswrapper[4631]: I1128 13:21:05.243382 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Nov 28 13:21:05 crc kubenswrapper[4631]: I1128 13:21:05.252508 4631 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6d7f49b-96d5-4d70-a0e3-9d9f37f59073\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:20:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:20:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:20:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f11c70496a148e22446931e8183b52f16bd431cc7bd17e9678f875df424d6a03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T13:20:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35d912d041da9a7ed307ebea6e999cf087dacf84a3cefbfd0f3a9c891c979408\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T13:20:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d396f0adc112df728f80c8e1167526cdf1458a21e4802e2b70104f1a68086ff0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T13:20:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7cd0089f761e1315fe18e82695a8cf9170990bb535ea7235ee39b68edd79f23\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T13:20:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T13:20:43Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 28 13:21:05 crc kubenswrapper[4631]: I1128 13:21:05.265462 4631 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:02Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 28 13:21:05 crc kubenswrapper[4631]: I1128 13:21:05.273679 4631 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:02Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 28 13:21:05 crc kubenswrapper[4631]: I1128 13:21:05.280993 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 13:21:05 crc kubenswrapper[4631]: I1128 13:21:05.281033 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 13:21:05 crc kubenswrapper[4631]: I1128 13:21:05.281043 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 13:21:05 crc kubenswrapper[4631]: I1128 13:21:05.281060 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 13:21:05 crc kubenswrapper[4631]: I1128 13:21:05.281159 4631 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T13:21:05Z","lastTransitionTime":"2025-11-28T13:21:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 13:21:05 crc kubenswrapper[4631]: I1128 13:21:05.281746 4631 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-wbkjm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa609d49-5fa2-4014-956c-dc31c9b35554\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:05Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:05Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wwngs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T13:21:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-wbkjm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 28 13:21:05 crc kubenswrapper[4631]: I1128 13:21:05.294568 4631 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"07b4914f-d3c8-4e6a-909c-3af046e1759c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:20:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:20:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:20:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4f323ff537cd79159a51b592401e92fc06d54c490ff005486f27f41d42a35f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T13:20:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eff71f12d67a15adda3183a742f2869b104d773904a6c5954d1ccb6a2c64dd71\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T13:20:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://53ea681ea681973dd5177ec7f2925ef48fec88ac004b41fca867ce671f5a76d1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T13:20:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6138e1e9121a843fd730def8c3f019dd68a6f38d963ddb00124f2c532762e012\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T13:20:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ac30b44d2fe8ddb4ede530d4b99dab4dc641a097e21d4ae69cfcbeddc38802d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T13:20:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0b61d56ed070810021d253875a7bb5b9cb94a3446c56c4a92fb5418022dd643\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f0b61d56ed070810021d253875a7bb5b9cb94a3446c56c4a92fb5418022dd643\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T13:20:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T13:20:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T13:20:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 28 13:21:05 crc kubenswrapper[4631]: I1128 13:21:05.305188 4631 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:02Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 28 13:21:05 crc kubenswrapper[4631]: I1128 13:21:05.321807 4631 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:02Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 28 13:21:05 crc kubenswrapper[4631]: I1128 13:21:05.333257 4631 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:02Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 28 13:21:05 crc kubenswrapper[4631]: I1128 13:21:05.333671 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/aa609d49-5fa2-4014-956c-dc31c9b35554-hosts-file\") pod \"node-resolver-wbkjm\" (UID: \"aa609d49-5fa2-4014-956c-dc31c9b35554\") " pod="openshift-dns/node-resolver-wbkjm" Nov 28 13:21:05 crc kubenswrapper[4631]: I1128 13:21:05.333706 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wwngs\" (UniqueName: \"kubernetes.io/projected/aa609d49-5fa2-4014-956c-dc31c9b35554-kube-api-access-wwngs\") pod \"node-resolver-wbkjm\" (UID: \"aa609d49-5fa2-4014-956c-dc31c9b35554\") " pod="openshift-dns/node-resolver-wbkjm" Nov 28 13:21:05 crc kubenswrapper[4631]: I1128 13:21:05.348399 4631 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:02Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 28 13:21:05 crc kubenswrapper[4631]: I1128 13:21:05.383577 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 13:21:05 crc kubenswrapper[4631]: I1128 13:21:05.383612 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 13:21:05 crc kubenswrapper[4631]: I1128 13:21:05.383620 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 13:21:05 crc kubenswrapper[4631]: I1128 13:21:05.383636 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 13:21:05 crc kubenswrapper[4631]: I1128 13:21:05.383646 4631 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T13:21:05Z","lastTransitionTime":"2025-11-28T13:21:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 13:21:05 crc kubenswrapper[4631]: I1128 13:21:05.434591 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/aa609d49-5fa2-4014-956c-dc31c9b35554-hosts-file\") pod \"node-resolver-wbkjm\" (UID: \"aa609d49-5fa2-4014-956c-dc31c9b35554\") " pod="openshift-dns/node-resolver-wbkjm" Nov 28 13:21:05 crc kubenswrapper[4631]: I1128 13:21:05.434634 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wwngs\" (UniqueName: \"kubernetes.io/projected/aa609d49-5fa2-4014-956c-dc31c9b35554-kube-api-access-wwngs\") pod \"node-resolver-wbkjm\" (UID: \"aa609d49-5fa2-4014-956c-dc31c9b35554\") " pod="openshift-dns/node-resolver-wbkjm" Nov 28 13:21:05 crc kubenswrapper[4631]: I1128 13:21:05.434945 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/aa609d49-5fa2-4014-956c-dc31c9b35554-hosts-file\") pod \"node-resolver-wbkjm\" (UID: \"aa609d49-5fa2-4014-956c-dc31c9b35554\") " pod="openshift-dns/node-resolver-wbkjm" Nov 28 13:21:05 crc kubenswrapper[4631]: I1128 13:21:05.456471 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wwngs\" (UniqueName: \"kubernetes.io/projected/aa609d49-5fa2-4014-956c-dc31c9b35554-kube-api-access-wwngs\") pod \"node-resolver-wbkjm\" (UID: \"aa609d49-5fa2-4014-956c-dc31c9b35554\") " pod="openshift-dns/node-resolver-wbkjm" Nov 28 13:21:05 crc kubenswrapper[4631]: I1128 13:21:05.486266 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 13:21:05 crc kubenswrapper[4631]: I1128 13:21:05.486338 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 13:21:05 crc kubenswrapper[4631]: I1128 13:21:05.486351 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 13:21:05 crc kubenswrapper[4631]: I1128 13:21:05.486370 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 13:21:05 crc kubenswrapper[4631]: I1128 13:21:05.486385 4631 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T13:21:05Z","lastTransitionTime":"2025-11-28T13:21:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 13:21:05 crc kubenswrapper[4631]: I1128 13:21:05.516456 4631 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Nov 28 13:21:05 crc kubenswrapper[4631]: I1128 13:21:05.550588 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-wbkjm" Nov 28 13:21:05 crc kubenswrapper[4631]: W1128 13:21:05.560725 4631 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podaa609d49_5fa2_4014_956c_dc31c9b35554.slice/crio-3785b235f79d1eef3b2b960418ba36c938e855817a09f93166e29dab8907db91 WatchSource:0}: Error finding container 3785b235f79d1eef3b2b960418ba36c938e855817a09f93166e29dab8907db91: Status 404 returned error can't find the container with id 3785b235f79d1eef3b2b960418ba36c938e855817a09f93166e29dab8907db91 Nov 28 13:21:05 crc kubenswrapper[4631]: I1128 13:21:05.589462 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 13:21:05 crc kubenswrapper[4631]: I1128 13:21:05.589498 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 13:21:05 crc kubenswrapper[4631]: I1128 13:21:05.589536 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 13:21:05 crc kubenswrapper[4631]: I1128 13:21:05.589559 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 13:21:05 crc kubenswrapper[4631]: I1128 13:21:05.589571 4631 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T13:21:05Z","lastTransitionTime":"2025-11-28T13:21:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 13:21:05 crc kubenswrapper[4631]: I1128 13:21:05.633617 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"4050f233444d842037508368dcbeb859d3a3690c02ed92a80f126f9f002410e3"} Nov 28 13:21:05 crc kubenswrapper[4631]: I1128 13:21:05.634737 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"54272daa5ae6c51c3bba4e043cdc2a3ba5c3f468bbc8c434c6349e0e03447793"} Nov 28 13:21:05 crc kubenswrapper[4631]: I1128 13:21:05.636076 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-wbkjm" event={"ID":"aa609d49-5fa2-4014-956c-dc31c9b35554","Type":"ContainerStarted","Data":"3785b235f79d1eef3b2b960418ba36c938e855817a09f93166e29dab8907db91"} Nov 28 13:21:05 crc kubenswrapper[4631]: I1128 13:21:05.637685 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"52446f93ddc7dd40ffdd35102ee559453f083ff89984e10c156ddfe43a4c113a"} Nov 28 13:21:05 crc kubenswrapper[4631]: I1128 13:21:05.692311 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 13:21:05 crc kubenswrapper[4631]: I1128 13:21:05.692348 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 13:21:05 crc kubenswrapper[4631]: I1128 13:21:05.692362 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 13:21:05 crc kubenswrapper[4631]: I1128 13:21:05.692376 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 13:21:05 crc kubenswrapper[4631]: I1128 13:21:05.692385 4631 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T13:21:05Z","lastTransitionTime":"2025-11-28T13:21:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 13:21:05 crc kubenswrapper[4631]: I1128 13:21:05.713999 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-8gtw6"] Nov 28 13:21:05 crc kubenswrapper[4631]: I1128 13:21:05.714554 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-8gtw6" Nov 28 13:21:05 crc kubenswrapper[4631]: I1128 13:21:05.722650 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Nov 28 13:21:05 crc kubenswrapper[4631]: I1128 13:21:05.722699 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Nov 28 13:21:05 crc kubenswrapper[4631]: I1128 13:21:05.722898 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Nov 28 13:21:05 crc kubenswrapper[4631]: I1128 13:21:05.723081 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Nov 28 13:21:05 crc kubenswrapper[4631]: I1128 13:21:05.723229 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Nov 28 13:21:05 crc kubenswrapper[4631]: I1128 13:21:05.726613 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-47sxw"] Nov 28 13:21:05 crc kubenswrapper[4631]: I1128 13:21:05.727123 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" Nov 28 13:21:05 crc kubenswrapper[4631]: I1128 13:21:05.760433 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-j8d66"] Nov 28 13:21:05 crc kubenswrapper[4631]: I1128 13:21:05.771326 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Nov 28 13:21:05 crc kubenswrapper[4631]: I1128 13:21:05.771462 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-j8d66" Nov 28 13:21:05 crc kubenswrapper[4631]: I1128 13:21:05.771526 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Nov 28 13:21:05 crc kubenswrapper[4631]: I1128 13:21:05.772009 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Nov 28 13:21:05 crc kubenswrapper[4631]: I1128 13:21:05.772156 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Nov 28 13:21:05 crc kubenswrapper[4631]: I1128 13:21:05.772741 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Nov 28 13:21:05 crc kubenswrapper[4631]: I1128 13:21:05.796530 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Nov 28 13:21:05 crc kubenswrapper[4631]: I1128 13:21:05.796728 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Nov 28 13:21:05 crc kubenswrapper[4631]: I1128 13:21:05.798157 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 13:21:05 crc kubenswrapper[4631]: I1128 13:21:05.798180 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 13:21:05 crc kubenswrapper[4631]: I1128 13:21:05.798189 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 13:21:05 crc kubenswrapper[4631]: I1128 13:21:05.798204 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 13:21:05 crc kubenswrapper[4631]: I1128 13:21:05.798215 4631 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T13:21:05Z","lastTransitionTime":"2025-11-28T13:21:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 13:21:05 crc kubenswrapper[4631]: I1128 13:21:05.802486 4631 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:02Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 28 13:21:05 crc kubenswrapper[4631]: I1128 13:21:05.839195 4631 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:02Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 28 13:21:05 crc kubenswrapper[4631]: I1128 13:21:05.839616 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/9a346894-644f-4359-baa5-23bb2d0acc5f-host-run-netns\") pod \"multus-j8d66\" (UID: \"9a346894-644f-4359-baa5-23bb2d0acc5f\") " pod="openshift-multus/multus-j8d66" Nov 28 13:21:05 crc kubenswrapper[4631]: I1128 13:21:05.839650 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/9a346894-644f-4359-baa5-23bb2d0acc5f-multus-conf-dir\") pod \"multus-j8d66\" (UID: \"9a346894-644f-4359-baa5-23bb2d0acc5f\") " pod="openshift-multus/multus-j8d66" Nov 28 13:21:05 crc kubenswrapper[4631]: I1128 13:21:05.839689 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/a0fde998-9a55-4734-9008-136b93f29d0c-system-cni-dir\") pod \"multus-additional-cni-plugins-8gtw6\" (UID: \"a0fde998-9a55-4734-9008-136b93f29d0c\") " pod="openshift-multus/multus-additional-cni-plugins-8gtw6" Nov 28 13:21:05 crc kubenswrapper[4631]: I1128 13:21:05.839708 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/a0fde998-9a55-4734-9008-136b93f29d0c-tuning-conf-dir\") pod \"multus-additional-cni-plugins-8gtw6\" (UID: \"a0fde998-9a55-4734-9008-136b93f29d0c\") " pod="openshift-multus/multus-additional-cni-plugins-8gtw6" Nov 28 13:21:05 crc kubenswrapper[4631]: I1128 13:21:05.839721 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/9a346894-644f-4359-baa5-23bb2d0acc5f-host-var-lib-kubelet\") pod \"multus-j8d66\" (UID: \"9a346894-644f-4359-baa5-23bb2d0acc5f\") " pod="openshift-multus/multus-j8d66" Nov 28 13:21:05 crc kubenswrapper[4631]: I1128 13:21:05.839734 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9a346894-644f-4359-baa5-23bb2d0acc5f-etc-kubernetes\") pod \"multus-j8d66\" (UID: \"9a346894-644f-4359-baa5-23bb2d0acc5f\") " pod="openshift-multus/multus-j8d66" Nov 28 13:21:05 crc kubenswrapper[4631]: I1128 13:21:05.839749 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wtbl6\" (UniqueName: \"kubernetes.io/projected/8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1-kube-api-access-wtbl6\") pod \"machine-config-daemon-47sxw\" (UID: \"8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1\") " pod="openshift-machine-config-operator/machine-config-daemon-47sxw" Nov 28 13:21:05 crc kubenswrapper[4631]: I1128 13:21:05.839763 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/9a346894-644f-4359-baa5-23bb2d0acc5f-host-run-k8s-cni-cncf-io\") pod \"multus-j8d66\" (UID: \"9a346894-644f-4359-baa5-23bb2d0acc5f\") " pod="openshift-multus/multus-j8d66" Nov 28 13:21:05 crc kubenswrapper[4631]: I1128 13:21:05.839776 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/9a346894-644f-4359-baa5-23bb2d0acc5f-host-var-lib-cni-multus\") pod \"multus-j8d66\" (UID: \"9a346894-644f-4359-baa5-23bb2d0acc5f\") " pod="openshift-multus/multus-j8d66" Nov 28 13:21:05 crc kubenswrapper[4631]: I1128 13:21:05.839791 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/9a346894-644f-4359-baa5-23bb2d0acc5f-multus-daemon-config\") pod \"multus-j8d66\" (UID: \"9a346894-644f-4359-baa5-23bb2d0acc5f\") " pod="openshift-multus/multus-j8d66" Nov 28 13:21:05 crc kubenswrapper[4631]: I1128 13:21:05.839814 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/9a346894-644f-4359-baa5-23bb2d0acc5f-cni-binary-copy\") pod \"multus-j8d66\" (UID: \"9a346894-644f-4359-baa5-23bb2d0acc5f\") " pod="openshift-multus/multus-j8d66" Nov 28 13:21:05 crc kubenswrapper[4631]: I1128 13:21:05.839828 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/9a346894-644f-4359-baa5-23bb2d0acc5f-host-var-lib-cni-bin\") pod \"multus-j8d66\" (UID: \"9a346894-644f-4359-baa5-23bb2d0acc5f\") " pod="openshift-multus/multus-j8d66" Nov 28 13:21:05 crc kubenswrapper[4631]: I1128 13:21:05.839851 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/a0fde998-9a55-4734-9008-136b93f29d0c-os-release\") pod \"multus-additional-cni-plugins-8gtw6\" (UID: \"a0fde998-9a55-4734-9008-136b93f29d0c\") " pod="openshift-multus/multus-additional-cni-plugins-8gtw6" Nov 28 13:21:05 crc kubenswrapper[4631]: I1128 13:21:05.839865 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/a0fde998-9a55-4734-9008-136b93f29d0c-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-8gtw6\" (UID: \"a0fde998-9a55-4734-9008-136b93f29d0c\") " pod="openshift-multus/multus-additional-cni-plugins-8gtw6" Nov 28 13:21:05 crc kubenswrapper[4631]: I1128 13:21:05.839889 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/9a346894-644f-4359-baa5-23bb2d0acc5f-system-cni-dir\") pod \"multus-j8d66\" (UID: \"9a346894-644f-4359-baa5-23bb2d0acc5f\") " pod="openshift-multus/multus-j8d66" Nov 28 13:21:05 crc kubenswrapper[4631]: I1128 13:21:05.839903 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/9a346894-644f-4359-baa5-23bb2d0acc5f-hostroot\") pod \"multus-j8d66\" (UID: \"9a346894-644f-4359-baa5-23bb2d0acc5f\") " pod="openshift-multus/multus-j8d66" Nov 28 13:21:05 crc kubenswrapper[4631]: I1128 13:21:05.839917 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/a0fde998-9a55-4734-9008-136b93f29d0c-cnibin\") pod \"multus-additional-cni-plugins-8gtw6\" (UID: \"a0fde998-9a55-4734-9008-136b93f29d0c\") " pod="openshift-multus/multus-additional-cni-plugins-8gtw6" Nov 28 13:21:05 crc kubenswrapper[4631]: I1128 13:21:05.839933 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/9a346894-644f-4359-baa5-23bb2d0acc5f-cnibin\") pod \"multus-j8d66\" (UID: \"9a346894-644f-4359-baa5-23bb2d0acc5f\") " pod="openshift-multus/multus-j8d66" Nov 28 13:21:05 crc kubenswrapper[4631]: I1128 13:21:05.839946 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/9a346894-644f-4359-baa5-23bb2d0acc5f-os-release\") pod \"multus-j8d66\" (UID: \"9a346894-644f-4359-baa5-23bb2d0acc5f\") " pod="openshift-multus/multus-j8d66" Nov 28 13:21:05 crc kubenswrapper[4631]: I1128 13:21:05.839960 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/9a346894-644f-4359-baa5-23bb2d0acc5f-multus-socket-dir-parent\") pod \"multus-j8d66\" (UID: \"9a346894-644f-4359-baa5-23bb2d0acc5f\") " pod="openshift-multus/multus-j8d66" Nov 28 13:21:05 crc kubenswrapper[4631]: I1128 13:21:05.839988 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/9a346894-644f-4359-baa5-23bb2d0acc5f-multus-cni-dir\") pod \"multus-j8d66\" (UID: \"9a346894-644f-4359-baa5-23bb2d0acc5f\") " pod="openshift-multus/multus-j8d66" Nov 28 13:21:05 crc kubenswrapper[4631]: I1128 13:21:05.840003 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1-proxy-tls\") pod \"machine-config-daemon-47sxw\" (UID: \"8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1\") " pod="openshift-machine-config-operator/machine-config-daemon-47sxw" Nov 28 13:21:05 crc kubenswrapper[4631]: I1128 13:21:05.840017 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/9a346894-644f-4359-baa5-23bb2d0acc5f-host-run-multus-certs\") pod \"multus-j8d66\" (UID: \"9a346894-644f-4359-baa5-23bb2d0acc5f\") " pod="openshift-multus/multus-j8d66" Nov 28 13:21:05 crc kubenswrapper[4631]: I1128 13:21:05.840032 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2blnl\" (UniqueName: \"kubernetes.io/projected/9a346894-644f-4359-baa5-23bb2d0acc5f-kube-api-access-2blnl\") pod \"multus-j8d66\" (UID: \"9a346894-644f-4359-baa5-23bb2d0acc5f\") " pod="openshift-multus/multus-j8d66" Nov 28 13:21:05 crc kubenswrapper[4631]: I1128 13:21:05.840046 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h89l2\" (UniqueName: \"kubernetes.io/projected/a0fde998-9a55-4734-9008-136b93f29d0c-kube-api-access-h89l2\") pod \"multus-additional-cni-plugins-8gtw6\" (UID: \"a0fde998-9a55-4734-9008-136b93f29d0c\") " pod="openshift-multus/multus-additional-cni-plugins-8gtw6" Nov 28 13:21:05 crc kubenswrapper[4631]: I1128 13:21:05.840062 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1-rootfs\") pod \"machine-config-daemon-47sxw\" (UID: \"8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1\") " pod="openshift-machine-config-operator/machine-config-daemon-47sxw" Nov 28 13:21:05 crc kubenswrapper[4631]: I1128 13:21:05.840075 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/a0fde998-9a55-4734-9008-136b93f29d0c-cni-binary-copy\") pod \"multus-additional-cni-plugins-8gtw6\" (UID: \"a0fde998-9a55-4734-9008-136b93f29d0c\") " pod="openshift-multus/multus-additional-cni-plugins-8gtw6" Nov 28 13:21:05 crc kubenswrapper[4631]: I1128 13:21:05.840089 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1-mcd-auth-proxy-config\") pod \"machine-config-daemon-47sxw\" (UID: \"8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1\") " pod="openshift-machine-config-operator/machine-config-daemon-47sxw" Nov 28 13:21:05 crc kubenswrapper[4631]: I1128 13:21:05.861131 4631 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:02Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 28 13:21:05 crc kubenswrapper[4631]: I1128 13:21:05.900199 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 13:21:05 crc kubenswrapper[4631]: I1128 13:21:05.900234 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 13:21:05 crc kubenswrapper[4631]: I1128 13:21:05.900245 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 13:21:05 crc kubenswrapper[4631]: I1128 13:21:05.900263 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 13:21:05 crc kubenswrapper[4631]: I1128 13:21:05.900301 4631 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T13:21:05Z","lastTransitionTime":"2025-11-28T13:21:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 13:21:05 crc kubenswrapper[4631]: I1128 13:21:05.902978 4631 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:02Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 28 13:21:05 crc kubenswrapper[4631]: I1128 13:21:05.919175 4631 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-wbkjm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa609d49-5fa2-4014-956c-dc31c9b35554\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:05Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:05Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wwngs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T13:21:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-wbkjm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 28 13:21:05 crc kubenswrapper[4631]: I1128 13:21:05.935678 4631 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"07b4914f-d3c8-4e6a-909c-3af046e1759c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:20:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:20:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:20:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4f323ff537cd79159a51b592401e92fc06d54c490ff005486f27f41d42a35f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T13:20:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eff71f12d67a15adda3183a742f2869b104d773904a6c5954d1ccb6a2c64dd71\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T13:20:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://53ea681ea681973dd5177ec7f2925ef48fec88ac004b41fca867ce671f5a76d1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T13:20:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6138e1e9121a843fd730def8c3f019dd68a6f38d963ddb00124f2c532762e012\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T13:20:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ac30b44d2fe8ddb4ede530d4b99dab4dc641a097e21d4ae69cfcbeddc38802d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T13:20:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0b61d56ed070810021d253875a7bb5b9cb94a3446c56c4a92fb5418022dd643\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f0b61d56ed070810021d253875a7bb5b9cb94a3446c56c4a92fb5418022dd643\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T13:20:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T13:20:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T13:20:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 28 13:21:05 crc kubenswrapper[4631]: I1128 13:21:05.941007 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/9a346894-644f-4359-baa5-23bb2d0acc5f-multus-cni-dir\") pod \"multus-j8d66\" (UID: \"9a346894-644f-4359-baa5-23bb2d0acc5f\") " pod="openshift-multus/multus-j8d66" Nov 28 13:21:05 crc kubenswrapper[4631]: I1128 13:21:05.941055 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1-proxy-tls\") pod \"machine-config-daemon-47sxw\" (UID: \"8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1\") " pod="openshift-machine-config-operator/machine-config-daemon-47sxw" Nov 28 13:21:05 crc kubenswrapper[4631]: I1128 13:21:05.941073 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/9a346894-644f-4359-baa5-23bb2d0acc5f-host-run-multus-certs\") pod \"multus-j8d66\" (UID: \"9a346894-644f-4359-baa5-23bb2d0acc5f\") " pod="openshift-multus/multus-j8d66" Nov 28 13:21:05 crc kubenswrapper[4631]: I1128 13:21:05.941091 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2blnl\" (UniqueName: \"kubernetes.io/projected/9a346894-644f-4359-baa5-23bb2d0acc5f-kube-api-access-2blnl\") pod \"multus-j8d66\" (UID: \"9a346894-644f-4359-baa5-23bb2d0acc5f\") " pod="openshift-multus/multus-j8d66" Nov 28 13:21:05 crc kubenswrapper[4631]: I1128 13:21:05.941111 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h89l2\" (UniqueName: \"kubernetes.io/projected/a0fde998-9a55-4734-9008-136b93f29d0c-kube-api-access-h89l2\") pod \"multus-additional-cni-plugins-8gtw6\" (UID: \"a0fde998-9a55-4734-9008-136b93f29d0c\") " pod="openshift-multus/multus-additional-cni-plugins-8gtw6" Nov 28 13:21:05 crc kubenswrapper[4631]: I1128 13:21:05.941131 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1-rootfs\") pod \"machine-config-daemon-47sxw\" (UID: \"8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1\") " pod="openshift-machine-config-operator/machine-config-daemon-47sxw" Nov 28 13:21:05 crc kubenswrapper[4631]: I1128 13:21:05.941151 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/a0fde998-9a55-4734-9008-136b93f29d0c-cni-binary-copy\") pod \"multus-additional-cni-plugins-8gtw6\" (UID: \"a0fde998-9a55-4734-9008-136b93f29d0c\") " pod="openshift-multus/multus-additional-cni-plugins-8gtw6" Nov 28 13:21:05 crc kubenswrapper[4631]: I1128 13:21:05.941169 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1-mcd-auth-proxy-config\") pod \"machine-config-daemon-47sxw\" (UID: \"8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1\") " pod="openshift-machine-config-operator/machine-config-daemon-47sxw" Nov 28 13:21:05 crc kubenswrapper[4631]: I1128 13:21:05.941187 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/9a346894-644f-4359-baa5-23bb2d0acc5f-host-run-netns\") pod \"multus-j8d66\" (UID: \"9a346894-644f-4359-baa5-23bb2d0acc5f\") " pod="openshift-multus/multus-j8d66" Nov 28 13:21:05 crc kubenswrapper[4631]: I1128 13:21:05.941203 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/9a346894-644f-4359-baa5-23bb2d0acc5f-multus-conf-dir\") pod \"multus-j8d66\" (UID: \"9a346894-644f-4359-baa5-23bb2d0acc5f\") " pod="openshift-multus/multus-j8d66" Nov 28 13:21:05 crc kubenswrapper[4631]: I1128 13:21:05.941221 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/a0fde998-9a55-4734-9008-136b93f29d0c-system-cni-dir\") pod \"multus-additional-cni-plugins-8gtw6\" (UID: \"a0fde998-9a55-4734-9008-136b93f29d0c\") " pod="openshift-multus/multus-additional-cni-plugins-8gtw6" Nov 28 13:21:05 crc kubenswrapper[4631]: I1128 13:21:05.941221 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/9a346894-644f-4359-baa5-23bb2d0acc5f-multus-cni-dir\") pod \"multus-j8d66\" (UID: \"9a346894-644f-4359-baa5-23bb2d0acc5f\") " pod="openshift-multus/multus-j8d66" Nov 28 13:21:05 crc kubenswrapper[4631]: I1128 13:21:05.941238 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/a0fde998-9a55-4734-9008-136b93f29d0c-tuning-conf-dir\") pod \"multus-additional-cni-plugins-8gtw6\" (UID: \"a0fde998-9a55-4734-9008-136b93f29d0c\") " pod="openshift-multus/multus-additional-cni-plugins-8gtw6" Nov 28 13:21:05 crc kubenswrapper[4631]: I1128 13:21:05.941317 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1-rootfs\") pod \"machine-config-daemon-47sxw\" (UID: \"8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1\") " pod="openshift-machine-config-operator/machine-config-daemon-47sxw" Nov 28 13:21:05 crc kubenswrapper[4631]: I1128 13:21:05.941349 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/9a346894-644f-4359-baa5-23bb2d0acc5f-host-var-lib-kubelet\") pod \"multus-j8d66\" (UID: \"9a346894-644f-4359-baa5-23bb2d0acc5f\") " pod="openshift-multus/multus-j8d66" Nov 28 13:21:05 crc kubenswrapper[4631]: I1128 13:21:05.941380 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wtbl6\" (UniqueName: \"kubernetes.io/projected/8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1-kube-api-access-wtbl6\") pod \"machine-config-daemon-47sxw\" (UID: \"8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1\") " pod="openshift-machine-config-operator/machine-config-daemon-47sxw" Nov 28 13:21:05 crc kubenswrapper[4631]: I1128 13:21:05.941400 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/9a346894-644f-4359-baa5-23bb2d0acc5f-host-run-k8s-cni-cncf-io\") pod \"multus-j8d66\" (UID: \"9a346894-644f-4359-baa5-23bb2d0acc5f\") " pod="openshift-multus/multus-j8d66" Nov 28 13:21:05 crc kubenswrapper[4631]: I1128 13:21:05.941420 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/9a346894-644f-4359-baa5-23bb2d0acc5f-host-var-lib-cni-multus\") pod \"multus-j8d66\" (UID: \"9a346894-644f-4359-baa5-23bb2d0acc5f\") " pod="openshift-multus/multus-j8d66" Nov 28 13:21:05 crc kubenswrapper[4631]: I1128 13:21:05.941440 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/9a346894-644f-4359-baa5-23bb2d0acc5f-multus-daemon-config\") pod \"multus-j8d66\" (UID: \"9a346894-644f-4359-baa5-23bb2d0acc5f\") " pod="openshift-multus/multus-j8d66" Nov 28 13:21:05 crc kubenswrapper[4631]: I1128 13:21:05.941456 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9a346894-644f-4359-baa5-23bb2d0acc5f-etc-kubernetes\") pod \"multus-j8d66\" (UID: \"9a346894-644f-4359-baa5-23bb2d0acc5f\") " pod="openshift-multus/multus-j8d66" Nov 28 13:21:05 crc kubenswrapper[4631]: I1128 13:21:05.941497 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/9a346894-644f-4359-baa5-23bb2d0acc5f-cni-binary-copy\") pod \"multus-j8d66\" (UID: \"9a346894-644f-4359-baa5-23bb2d0acc5f\") " pod="openshift-multus/multus-j8d66" Nov 28 13:21:05 crc kubenswrapper[4631]: I1128 13:21:05.941517 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/9a346894-644f-4359-baa5-23bb2d0acc5f-host-var-lib-cni-bin\") pod \"multus-j8d66\" (UID: \"9a346894-644f-4359-baa5-23bb2d0acc5f\") " pod="openshift-multus/multus-j8d66" Nov 28 13:21:05 crc kubenswrapper[4631]: I1128 13:21:05.941534 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/a0fde998-9a55-4734-9008-136b93f29d0c-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-8gtw6\" (UID: \"a0fde998-9a55-4734-9008-136b93f29d0c\") " pod="openshift-multus/multus-additional-cni-plugins-8gtw6" Nov 28 13:21:05 crc kubenswrapper[4631]: I1128 13:21:05.941597 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/a0fde998-9a55-4734-9008-136b93f29d0c-os-release\") pod \"multus-additional-cni-plugins-8gtw6\" (UID: \"a0fde998-9a55-4734-9008-136b93f29d0c\") " pod="openshift-multus/multus-additional-cni-plugins-8gtw6" Nov 28 13:21:05 crc kubenswrapper[4631]: I1128 13:21:05.941617 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/9a346894-644f-4359-baa5-23bb2d0acc5f-system-cni-dir\") pod \"multus-j8d66\" (UID: \"9a346894-644f-4359-baa5-23bb2d0acc5f\") " pod="openshift-multus/multus-j8d66" Nov 28 13:21:05 crc kubenswrapper[4631]: I1128 13:21:05.941636 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/9a346894-644f-4359-baa5-23bb2d0acc5f-hostroot\") pod \"multus-j8d66\" (UID: \"9a346894-644f-4359-baa5-23bb2d0acc5f\") " pod="openshift-multus/multus-j8d66" Nov 28 13:21:05 crc kubenswrapper[4631]: I1128 13:21:05.941659 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/9a346894-644f-4359-baa5-23bb2d0acc5f-cnibin\") pod \"multus-j8d66\" (UID: \"9a346894-644f-4359-baa5-23bb2d0acc5f\") " pod="openshift-multus/multus-j8d66" Nov 28 13:21:05 crc kubenswrapper[4631]: I1128 13:21:05.941675 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/9a346894-644f-4359-baa5-23bb2d0acc5f-os-release\") pod \"multus-j8d66\" (UID: \"9a346894-644f-4359-baa5-23bb2d0acc5f\") " pod="openshift-multus/multus-j8d66" Nov 28 13:21:05 crc kubenswrapper[4631]: I1128 13:21:05.941693 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/9a346894-644f-4359-baa5-23bb2d0acc5f-multus-socket-dir-parent\") pod \"multus-j8d66\" (UID: \"9a346894-644f-4359-baa5-23bb2d0acc5f\") " pod="openshift-multus/multus-j8d66" Nov 28 13:21:05 crc kubenswrapper[4631]: I1128 13:21:05.941713 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/a0fde998-9a55-4734-9008-136b93f29d0c-cnibin\") pod \"multus-additional-cni-plugins-8gtw6\" (UID: \"a0fde998-9a55-4734-9008-136b93f29d0c\") " pod="openshift-multus/multus-additional-cni-plugins-8gtw6" Nov 28 13:21:05 crc kubenswrapper[4631]: I1128 13:21:05.941850 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/a0fde998-9a55-4734-9008-136b93f29d0c-cnibin\") pod \"multus-additional-cni-plugins-8gtw6\" (UID: \"a0fde998-9a55-4734-9008-136b93f29d0c\") " pod="openshift-multus/multus-additional-cni-plugins-8gtw6" Nov 28 13:21:05 crc kubenswrapper[4631]: I1128 13:21:05.941877 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/9a346894-644f-4359-baa5-23bb2d0acc5f-host-var-lib-kubelet\") pod \"multus-j8d66\" (UID: \"9a346894-644f-4359-baa5-23bb2d0acc5f\") " pod="openshift-multus/multus-j8d66" Nov 28 13:21:05 crc kubenswrapper[4631]: I1128 13:21:05.941950 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/9a346894-644f-4359-baa5-23bb2d0acc5f-multus-conf-dir\") pod \"multus-j8d66\" (UID: \"9a346894-644f-4359-baa5-23bb2d0acc5f\") " pod="openshift-multus/multus-j8d66" Nov 28 13:21:05 crc kubenswrapper[4631]: I1128 13:21:05.942026 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/9a346894-644f-4359-baa5-23bb2d0acc5f-host-run-netns\") pod \"multus-j8d66\" (UID: \"9a346894-644f-4359-baa5-23bb2d0acc5f\") " pod="openshift-multus/multus-j8d66" Nov 28 13:21:05 crc kubenswrapper[4631]: I1128 13:21:05.942060 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/a0fde998-9a55-4734-9008-136b93f29d0c-system-cni-dir\") pod \"multus-additional-cni-plugins-8gtw6\" (UID: \"a0fde998-9a55-4734-9008-136b93f29d0c\") " pod="openshift-multus/multus-additional-cni-plugins-8gtw6" Nov 28 13:21:05 crc kubenswrapper[4631]: I1128 13:21:05.942275 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/a0fde998-9a55-4734-9008-136b93f29d0c-os-release\") pod \"multus-additional-cni-plugins-8gtw6\" (UID: \"a0fde998-9a55-4734-9008-136b93f29d0c\") " pod="openshift-multus/multus-additional-cni-plugins-8gtw6" Nov 28 13:21:05 crc kubenswrapper[4631]: I1128 13:21:05.942347 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/9a346894-644f-4359-baa5-23bb2d0acc5f-system-cni-dir\") pod \"multus-j8d66\" (UID: \"9a346894-644f-4359-baa5-23bb2d0acc5f\") " pod="openshift-multus/multus-j8d66" Nov 28 13:21:05 crc kubenswrapper[4631]: I1128 13:21:05.942374 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/9a346894-644f-4359-baa5-23bb2d0acc5f-hostroot\") pod \"multus-j8d66\" (UID: \"9a346894-644f-4359-baa5-23bb2d0acc5f\") " pod="openshift-multus/multus-j8d66" Nov 28 13:21:05 crc kubenswrapper[4631]: I1128 13:21:05.941639 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/a0fde998-9a55-4734-9008-136b93f29d0c-tuning-conf-dir\") pod \"multus-additional-cni-plugins-8gtw6\" (UID: \"a0fde998-9a55-4734-9008-136b93f29d0c\") " pod="openshift-multus/multus-additional-cni-plugins-8gtw6" Nov 28 13:21:05 crc kubenswrapper[4631]: I1128 13:21:05.942426 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/9a346894-644f-4359-baa5-23bb2d0acc5f-cnibin\") pod \"multus-j8d66\" (UID: \"9a346894-644f-4359-baa5-23bb2d0acc5f\") " pod="openshift-multus/multus-j8d66" Nov 28 13:21:05 crc kubenswrapper[4631]: I1128 13:21:05.942459 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/9a346894-644f-4359-baa5-23bb2d0acc5f-os-release\") pod \"multus-j8d66\" (UID: \"9a346894-644f-4359-baa5-23bb2d0acc5f\") " pod="openshift-multus/multus-j8d66" Nov 28 13:21:05 crc kubenswrapper[4631]: I1128 13:21:05.942469 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1-mcd-auth-proxy-config\") pod \"machine-config-daemon-47sxw\" (UID: \"8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1\") " pod="openshift-machine-config-operator/machine-config-daemon-47sxw" Nov 28 13:21:05 crc kubenswrapper[4631]: I1128 13:21:05.942508 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/9a346894-644f-4359-baa5-23bb2d0acc5f-multus-socket-dir-parent\") pod \"multus-j8d66\" (UID: \"9a346894-644f-4359-baa5-23bb2d0acc5f\") " pod="openshift-multus/multus-j8d66" Nov 28 13:21:05 crc kubenswrapper[4631]: I1128 13:21:05.942465 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/a0fde998-9a55-4734-9008-136b93f29d0c-cni-binary-copy\") pod \"multus-additional-cni-plugins-8gtw6\" (UID: \"a0fde998-9a55-4734-9008-136b93f29d0c\") " pod="openshift-multus/multus-additional-cni-plugins-8gtw6" Nov 28 13:21:05 crc kubenswrapper[4631]: I1128 13:21:05.942540 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9a346894-644f-4359-baa5-23bb2d0acc5f-etc-kubernetes\") pod \"multus-j8d66\" (UID: \"9a346894-644f-4359-baa5-23bb2d0acc5f\") " pod="openshift-multus/multus-j8d66" Nov 28 13:21:05 crc kubenswrapper[4631]: I1128 13:21:05.942574 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/9a346894-644f-4359-baa5-23bb2d0acc5f-host-var-lib-cni-bin\") pod \"multus-j8d66\" (UID: \"9a346894-644f-4359-baa5-23bb2d0acc5f\") " pod="openshift-multus/multus-j8d66" Nov 28 13:21:05 crc kubenswrapper[4631]: I1128 13:21:05.942577 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/9a346894-644f-4359-baa5-23bb2d0acc5f-host-run-multus-certs\") pod \"multus-j8d66\" (UID: \"9a346894-644f-4359-baa5-23bb2d0acc5f\") " pod="openshift-multus/multus-j8d66" Nov 28 13:21:05 crc kubenswrapper[4631]: I1128 13:21:05.942610 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/9a346894-644f-4359-baa5-23bb2d0acc5f-host-run-k8s-cni-cncf-io\") pod \"multus-j8d66\" (UID: \"9a346894-644f-4359-baa5-23bb2d0acc5f\") " pod="openshift-multus/multus-j8d66" Nov 28 13:21:05 crc kubenswrapper[4631]: I1128 13:21:05.942611 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/9a346894-644f-4359-baa5-23bb2d0acc5f-host-var-lib-cni-multus\") pod \"multus-j8d66\" (UID: \"9a346894-644f-4359-baa5-23bb2d0acc5f\") " pod="openshift-multus/multus-j8d66" Nov 28 13:21:05 crc kubenswrapper[4631]: I1128 13:21:05.942712 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/a0fde998-9a55-4734-9008-136b93f29d0c-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-8gtw6\" (UID: \"a0fde998-9a55-4734-9008-136b93f29d0c\") " pod="openshift-multus/multus-additional-cni-plugins-8gtw6" Nov 28 13:21:05 crc kubenswrapper[4631]: I1128 13:21:05.943046 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/9a346894-644f-4359-baa5-23bb2d0acc5f-cni-binary-copy\") pod \"multus-j8d66\" (UID: \"9a346894-644f-4359-baa5-23bb2d0acc5f\") " pod="openshift-multus/multus-j8d66" Nov 28 13:21:05 crc kubenswrapper[4631]: I1128 13:21:05.943217 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/9a346894-644f-4359-baa5-23bb2d0acc5f-multus-daemon-config\") pod \"multus-j8d66\" (UID: \"9a346894-644f-4359-baa5-23bb2d0acc5f\") " pod="openshift-multus/multus-j8d66" Nov 28 13:21:05 crc kubenswrapper[4631]: I1128 13:21:05.946002 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1-proxy-tls\") pod \"machine-config-daemon-47sxw\" (UID: \"8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1\") " pod="openshift-machine-config-operator/machine-config-daemon-47sxw" Nov 28 13:21:05 crc kubenswrapper[4631]: I1128 13:21:05.953668 4631 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:02Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 28 13:21:05 crc kubenswrapper[4631]: I1128 13:21:05.965612 4631 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-8gtw6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a0fde998-9a55-4734-9008-136b93f29d0c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:05Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h89l2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h89l2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h89l2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h89l2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h89l2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h89l2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h89l2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T13:21:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-8gtw6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 28 13:21:05 crc kubenswrapper[4631]: I1128 13:21:05.966601 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h89l2\" (UniqueName: \"kubernetes.io/projected/a0fde998-9a55-4734-9008-136b93f29d0c-kube-api-access-h89l2\") pod \"multus-additional-cni-plugins-8gtw6\" (UID: \"a0fde998-9a55-4734-9008-136b93f29d0c\") " pod="openshift-multus/multus-additional-cni-plugins-8gtw6" Nov 28 13:21:05 crc kubenswrapper[4631]: I1128 13:21:05.969497 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2blnl\" (UniqueName: \"kubernetes.io/projected/9a346894-644f-4359-baa5-23bb2d0acc5f-kube-api-access-2blnl\") pod \"multus-j8d66\" (UID: \"9a346894-644f-4359-baa5-23bb2d0acc5f\") " pod="openshift-multus/multus-j8d66" Nov 28 13:21:05 crc kubenswrapper[4631]: I1128 13:21:05.974273 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wtbl6\" (UniqueName: \"kubernetes.io/projected/8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1-kube-api-access-wtbl6\") pod \"machine-config-daemon-47sxw\" (UID: \"8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1\") " pod="openshift-machine-config-operator/machine-config-daemon-47sxw" Nov 28 13:21:05 crc kubenswrapper[4631]: I1128 13:21:05.979167 4631 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6d7f49b-96d5-4d70-a0e3-9d9f37f59073\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:20:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:20:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:20:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f11c70496a148e22446931e8183b52f16bd431cc7bd17e9678f875df424d6a03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T13:20:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35d912d041da9a7ed307ebea6e999cf087dacf84a3cefbfd0f3a9c891c979408\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T13:20:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d396f0adc112df728f80c8e1167526cdf1458a21e4802e2b70104f1a68086ff0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T13:20:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7cd0089f761e1315fe18e82695a8cf9170990bb535ea7235ee39b68edd79f23\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T13:20:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T13:20:43Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 28 13:21:05 crc kubenswrapper[4631]: I1128 13:21:05.993765 4631 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:02Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 28 13:21:06 crc kubenswrapper[4631]: I1128 13:21:06.003053 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 13:21:06 crc kubenswrapper[4631]: I1128 13:21:06.003112 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 13:21:06 crc kubenswrapper[4631]: I1128 13:21:06.003131 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 13:21:06 crc kubenswrapper[4631]: I1128 13:21:06.003155 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 13:21:06 crc kubenswrapper[4631]: I1128 13:21:06.003167 4631 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T13:21:06Z","lastTransitionTime":"2025-11-28T13:21:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 13:21:06 crc kubenswrapper[4631]: I1128 13:21:06.005929 4631 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:02Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 28 13:21:06 crc kubenswrapper[4631]: I1128 13:21:06.017226 4631 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-j8d66" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a346894-644f-4359-baa5-23bb2d0acc5f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2blnl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T13:21:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-j8d66\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 28 13:21:06 crc kubenswrapper[4631]: I1128 13:21:06.031269 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-8gtw6" Nov 28 13:21:06 crc kubenswrapper[4631]: I1128 13:21:06.032478 4631 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-8gtw6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a0fde998-9a55-4734-9008-136b93f29d0c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:05Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h89l2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h89l2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h89l2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h89l2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h89l2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h89l2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h89l2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T13:21:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-8gtw6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 28 13:21:06 crc kubenswrapper[4631]: I1128 13:21:06.039409 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" Nov 28 13:21:06 crc kubenswrapper[4631]: I1128 13:21:06.044589 4631 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:05Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:05Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wtbl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wtbl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T13:21:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-47sxw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 28 13:21:06 crc kubenswrapper[4631]: W1128 13:21:06.045014 4631 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda0fde998_9a55_4734_9008_136b93f29d0c.slice/crio-61c2f1dc159b4952e9adf9f8848e22c2a5b45fde03d2ef6777af3ddb855da91b WatchSource:0}: Error finding container 61c2f1dc159b4952e9adf9f8848e22c2a5b45fde03d2ef6777af3ddb855da91b: Status 404 returned error can't find the container with id 61c2f1dc159b4952e9adf9f8848e22c2a5b45fde03d2ef6777af3ddb855da91b Nov 28 13:21:06 crc kubenswrapper[4631]: I1128 13:21:06.056564 4631 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:02Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 28 13:21:06 crc kubenswrapper[4631]: I1128 13:21:06.066281 4631 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:02Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 28 13:21:06 crc kubenswrapper[4631]: I1128 13:21:06.081708 4631 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:02Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 28 13:21:06 crc kubenswrapper[4631]: I1128 13:21:06.092570 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-j8d66" Nov 28 13:21:06 crc kubenswrapper[4631]: I1128 13:21:06.094147 4631 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-wbkjm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa609d49-5fa2-4014-956c-dc31c9b35554\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:05Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:05Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wwngs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T13:21:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-wbkjm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 28 13:21:06 crc kubenswrapper[4631]: I1128 13:21:06.106316 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 13:21:06 crc kubenswrapper[4631]: I1128 13:21:06.106352 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 13:21:06 crc kubenswrapper[4631]: I1128 13:21:06.106363 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 13:21:06 crc kubenswrapper[4631]: I1128 13:21:06.106381 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 13:21:06 crc kubenswrapper[4631]: I1128 13:21:06.106393 4631 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T13:21:06Z","lastTransitionTime":"2025-11-28T13:21:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 13:21:06 crc kubenswrapper[4631]: I1128 13:21:06.115544 4631 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"07b4914f-d3c8-4e6a-909c-3af046e1759c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:20:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:20:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:20:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4f323ff537cd79159a51b592401e92fc06d54c490ff005486f27f41d42a35f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T13:20:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eff71f12d67a15adda3183a742f2869b104d773904a6c5954d1ccb6a2c64dd71\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T13:20:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://53ea681ea681973dd5177ec7f2925ef48fec88ac004b41fca867ce671f5a76d1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T13:20:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6138e1e9121a843fd730def8c3f019dd68a6f38d963ddb00124f2c532762e012\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T13:20:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ac30b44d2fe8ddb4ede530d4b99dab4dc641a097e21d4ae69cfcbeddc38802d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T13:20:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0b61d56ed070810021d253875a7bb5b9cb94a3446c56c4a92fb5418022dd643\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f0b61d56ed070810021d253875a7bb5b9cb94a3446c56c4a92fb5418022dd643\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T13:20:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T13:20:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T13:20:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 28 13:21:06 crc kubenswrapper[4631]: I1128 13:21:06.129097 4631 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:02Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 28 13:21:06 crc kubenswrapper[4631]: I1128 13:21:06.133090 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-268tw"] Nov 28 13:21:06 crc kubenswrapper[4631]: I1128 13:21:06.139785 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-268tw" Nov 28 13:21:06 crc kubenswrapper[4631]: I1128 13:21:06.146050 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Nov 28 13:21:06 crc kubenswrapper[4631]: I1128 13:21:06.146267 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Nov 28 13:21:06 crc kubenswrapper[4631]: I1128 13:21:06.146354 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Nov 28 13:21:06 crc kubenswrapper[4631]: I1128 13:21:06.146489 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Nov 28 13:21:06 crc kubenswrapper[4631]: I1128 13:21:06.146647 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Nov 28 13:21:06 crc kubenswrapper[4631]: I1128 13:21:06.146515 4631 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6d7f49b-96d5-4d70-a0e3-9d9f37f59073\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:20:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:20:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:20:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f11c70496a148e22446931e8183b52f16bd431cc7bd17e9678f875df424d6a03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T13:20:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35d912d041da9a7ed307ebea6e999cf087dacf84a3cefbfd0f3a9c891c979408\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T13:20:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d396f0adc112df728f80c8e1167526cdf1458a21e4802e2b70104f1a68086ff0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T13:20:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7cd0089f761e1315fe18e82695a8cf9170990bb535ea7235ee39b68edd79f23\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T13:20:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T13:20:43Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 28 13:21:06 crc kubenswrapper[4631]: I1128 13:21:06.146926 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Nov 28 13:21:06 crc kubenswrapper[4631]: I1128 13:21:06.147130 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Nov 28 13:21:06 crc kubenswrapper[4631]: I1128 13:21:06.170963 4631 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:02Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 28 13:21:06 crc kubenswrapper[4631]: I1128 13:21:06.196818 4631 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-8gtw6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a0fde998-9a55-4734-9008-136b93f29d0c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:05Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h89l2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h89l2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h89l2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h89l2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h89l2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h89l2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h89l2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T13:21:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-8gtw6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 28 13:21:06 crc kubenswrapper[4631]: I1128 13:21:06.210375 4631 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:05Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:05Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wtbl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wtbl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T13:21:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-47sxw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 28 13:21:06 crc kubenswrapper[4631]: I1128 13:21:06.213319 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 13:21:06 crc kubenswrapper[4631]: I1128 13:21:06.213358 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 13:21:06 crc kubenswrapper[4631]: I1128 13:21:06.213367 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 13:21:06 crc kubenswrapper[4631]: I1128 13:21:06.213382 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 13:21:06 crc kubenswrapper[4631]: I1128 13:21:06.213391 4631 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T13:21:06Z","lastTransitionTime":"2025-11-28T13:21:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 13:21:06 crc kubenswrapper[4631]: I1128 13:21:06.223407 4631 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"07b4914f-d3c8-4e6a-909c-3af046e1759c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:20:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:20:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:20:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4f323ff537cd79159a51b592401e92fc06d54c490ff005486f27f41d42a35f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T13:20:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eff71f12d67a15adda3183a742f2869b104d773904a6c5954d1ccb6a2c64dd71\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T13:20:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://53ea681ea681973dd5177ec7f2925ef48fec88ac004b41fca867ce671f5a76d1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T13:20:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6138e1e9121a843fd730def8c3f019dd68a6f38d963ddb00124f2c532762e012\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T13:20:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ac30b44d2fe8ddb4ede530d4b99dab4dc641a097e21d4ae69cfcbeddc38802d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T13:20:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0b61d56ed070810021d253875a7bb5b9cb94a3446c56c4a92fb5418022dd643\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f0b61d56ed070810021d253875a7bb5b9cb94a3446c56c4a92fb5418022dd643\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T13:20:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T13:20:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T13:20:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 28 13:21:06 crc kubenswrapper[4631]: I1128 13:21:06.235048 4631 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:02Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 28 13:21:06 crc kubenswrapper[4631]: I1128 13:21:06.246818 4631 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:02Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 28 13:21:06 crc kubenswrapper[4631]: I1128 13:21:06.246927 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 13:21:06 crc kubenswrapper[4631]: I1128 13:21:06.247075 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/a205f475-340d-4ec5-a860-d4a5c2ceb39e-host-run-ovn-kubernetes\") pod \"ovnkube-node-268tw\" (UID: \"a205f475-340d-4ec5-a860-d4a5c2ceb39e\") " pod="openshift-ovn-kubernetes/ovnkube-node-268tw" Nov 28 13:21:06 crc kubenswrapper[4631]: I1128 13:21:06.247102 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/a205f475-340d-4ec5-a860-d4a5c2ceb39e-ovn-node-metrics-cert\") pod \"ovnkube-node-268tw\" (UID: \"a205f475-340d-4ec5-a860-d4a5c2ceb39e\") " pod="openshift-ovn-kubernetes/ovnkube-node-268tw" Nov 28 13:21:06 crc kubenswrapper[4631]: I1128 13:21:06.247121 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/a205f475-340d-4ec5-a860-d4a5c2ceb39e-env-overrides\") pod \"ovnkube-node-268tw\" (UID: \"a205f475-340d-4ec5-a860-d4a5c2ceb39e\") " pod="openshift-ovn-kubernetes/ovnkube-node-268tw" Nov 28 13:21:06 crc kubenswrapper[4631]: I1128 13:21:06.247145 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/a205f475-340d-4ec5-a860-d4a5c2ceb39e-log-socket\") pod \"ovnkube-node-268tw\" (UID: \"a205f475-340d-4ec5-a860-d4a5c2ceb39e\") " pod="openshift-ovn-kubernetes/ovnkube-node-268tw" Nov 28 13:21:06 crc kubenswrapper[4631]: I1128 13:21:06.247166 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/a205f475-340d-4ec5-a860-d4a5c2ceb39e-host-kubelet\") pod \"ovnkube-node-268tw\" (UID: \"a205f475-340d-4ec5-a860-d4a5c2ceb39e\") " pod="openshift-ovn-kubernetes/ovnkube-node-268tw" Nov 28 13:21:06 crc kubenswrapper[4631]: I1128 13:21:06.247183 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/a205f475-340d-4ec5-a860-d4a5c2ceb39e-host-cni-bin\") pod \"ovnkube-node-268tw\" (UID: \"a205f475-340d-4ec5-a860-d4a5c2ceb39e\") " pod="openshift-ovn-kubernetes/ovnkube-node-268tw" Nov 28 13:21:06 crc kubenswrapper[4631]: I1128 13:21:06.247205 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a205f475-340d-4ec5-a860-d4a5c2ceb39e-var-lib-openvswitch\") pod \"ovnkube-node-268tw\" (UID: \"a205f475-340d-4ec5-a860-d4a5c2ceb39e\") " pod="openshift-ovn-kubernetes/ovnkube-node-268tw" Nov 28 13:21:06 crc kubenswrapper[4631]: I1128 13:21:06.247223 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/a205f475-340d-4ec5-a860-d4a5c2ceb39e-systemd-units\") pod \"ovnkube-node-268tw\" (UID: \"a205f475-340d-4ec5-a860-d4a5c2ceb39e\") " pod="openshift-ovn-kubernetes/ovnkube-node-268tw" Nov 28 13:21:06 crc kubenswrapper[4631]: I1128 13:21:06.247248 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qdzgl\" (UniqueName: \"kubernetes.io/projected/a205f475-340d-4ec5-a860-d4a5c2ceb39e-kube-api-access-qdzgl\") pod \"ovnkube-node-268tw\" (UID: \"a205f475-340d-4ec5-a860-d4a5c2ceb39e\") " pod="openshift-ovn-kubernetes/ovnkube-node-268tw" Nov 28 13:21:06 crc kubenswrapper[4631]: I1128 13:21:06.247268 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/a205f475-340d-4ec5-a860-d4a5c2ceb39e-host-cni-netd\") pod \"ovnkube-node-268tw\" (UID: \"a205f475-340d-4ec5-a860-d4a5c2ceb39e\") " pod="openshift-ovn-kubernetes/ovnkube-node-268tw" Nov 28 13:21:06 crc kubenswrapper[4631]: I1128 13:21:06.247330 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 13:21:06 crc kubenswrapper[4631]: E1128 13:21:06.247363 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 13:21:10.247325331 +0000 UTC m=+27.054628675 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 13:21:06 crc kubenswrapper[4631]: E1128 13:21:06.247435 4631 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 28 13:21:06 crc kubenswrapper[4631]: E1128 13:21:06.247636 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-28 13:21:10.247604238 +0000 UTC m=+27.054907582 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 28 13:21:06 crc kubenswrapper[4631]: I1128 13:21:06.247530 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/a205f475-340d-4ec5-a860-d4a5c2ceb39e-host-slash\") pod \"ovnkube-node-268tw\" (UID: \"a205f475-340d-4ec5-a860-d4a5c2ceb39e\") " pod="openshift-ovn-kubernetes/ovnkube-node-268tw" Nov 28 13:21:06 crc kubenswrapper[4631]: I1128 13:21:06.247707 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a205f475-340d-4ec5-a860-d4a5c2ceb39e-etc-openvswitch\") pod \"ovnkube-node-268tw\" (UID: \"a205f475-340d-4ec5-a860-d4a5c2ceb39e\") " pod="openshift-ovn-kubernetes/ovnkube-node-268tw" Nov 28 13:21:06 crc kubenswrapper[4631]: I1128 13:21:06.247758 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a205f475-340d-4ec5-a860-d4a5c2ceb39e-run-openvswitch\") pod \"ovnkube-node-268tw\" (UID: \"a205f475-340d-4ec5-a860-d4a5c2ceb39e\") " pod="openshift-ovn-kubernetes/ovnkube-node-268tw" Nov 28 13:21:06 crc kubenswrapper[4631]: I1128 13:21:06.247797 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/a205f475-340d-4ec5-a860-d4a5c2ceb39e-run-ovn\") pod \"ovnkube-node-268tw\" (UID: \"a205f475-340d-4ec5-a860-d4a5c2ceb39e\") " pod="openshift-ovn-kubernetes/ovnkube-node-268tw" Nov 28 13:21:06 crc kubenswrapper[4631]: I1128 13:21:06.247868 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 13:21:06 crc kubenswrapper[4631]: I1128 13:21:06.247929 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/a205f475-340d-4ec5-a860-d4a5c2ceb39e-node-log\") pod \"ovnkube-node-268tw\" (UID: \"a205f475-340d-4ec5-a860-d4a5c2ceb39e\") " pod="openshift-ovn-kubernetes/ovnkube-node-268tw" Nov 28 13:21:06 crc kubenswrapper[4631]: I1128 13:21:06.247952 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/a205f475-340d-4ec5-a860-d4a5c2ceb39e-ovnkube-script-lib\") pod \"ovnkube-node-268tw\" (UID: \"a205f475-340d-4ec5-a860-d4a5c2ceb39e\") " pod="openshift-ovn-kubernetes/ovnkube-node-268tw" Nov 28 13:21:06 crc kubenswrapper[4631]: I1128 13:21:06.247994 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/a205f475-340d-4ec5-a860-d4a5c2ceb39e-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-268tw\" (UID: \"a205f475-340d-4ec5-a860-d4a5c2ceb39e\") " pod="openshift-ovn-kubernetes/ovnkube-node-268tw" Nov 28 13:21:06 crc kubenswrapper[4631]: I1128 13:21:06.248038 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 13:21:06 crc kubenswrapper[4631]: I1128 13:21:06.248086 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/a205f475-340d-4ec5-a860-d4a5c2ceb39e-ovnkube-config\") pod \"ovnkube-node-268tw\" (UID: \"a205f475-340d-4ec5-a860-d4a5c2ceb39e\") " pod="openshift-ovn-kubernetes/ovnkube-node-268tw" Nov 28 13:21:06 crc kubenswrapper[4631]: E1128 13:21:06.248103 4631 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 28 13:21:06 crc kubenswrapper[4631]: I1128 13:21:06.248117 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 13:21:06 crc kubenswrapper[4631]: E1128 13:21:06.248134 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-28 13:21:10.248126188 +0000 UTC m=+27.055429522 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 28 13:21:06 crc kubenswrapper[4631]: I1128 13:21:06.248246 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/a205f475-340d-4ec5-a860-d4a5c2ceb39e-host-run-netns\") pod \"ovnkube-node-268tw\" (UID: \"a205f475-340d-4ec5-a860-d4a5c2ceb39e\") " pod="openshift-ovn-kubernetes/ovnkube-node-268tw" Nov 28 13:21:06 crc kubenswrapper[4631]: E1128 13:21:06.248313 4631 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 28 13:21:06 crc kubenswrapper[4631]: E1128 13:21:06.248334 4631 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 28 13:21:06 crc kubenswrapper[4631]: E1128 13:21:06.248348 4631 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 28 13:21:06 crc kubenswrapper[4631]: I1128 13:21:06.248357 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/a205f475-340d-4ec5-a860-d4a5c2ceb39e-run-systemd\") pod \"ovnkube-node-268tw\" (UID: \"a205f475-340d-4ec5-a860-d4a5c2ceb39e\") " pod="openshift-ovn-kubernetes/ovnkube-node-268tw" Nov 28 13:21:06 crc kubenswrapper[4631]: E1128 13:21:06.248430 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-28 13:21:10.248421164 +0000 UTC m=+27.055724508 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 28 13:21:06 crc kubenswrapper[4631]: E1128 13:21:06.248515 4631 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 28 13:21:06 crc kubenswrapper[4631]: E1128 13:21:06.248527 4631 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 28 13:21:06 crc kubenswrapper[4631]: E1128 13:21:06.248536 4631 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 28 13:21:06 crc kubenswrapper[4631]: E1128 13:21:06.248603 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-28 13:21:10.248594678 +0000 UTC m=+27.055898252 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 28 13:21:06 crc kubenswrapper[4631]: I1128 13:21:06.259164 4631 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:02Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 28 13:21:06 crc kubenswrapper[4631]: I1128 13:21:06.270829 4631 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-wbkjm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa609d49-5fa2-4014-956c-dc31c9b35554\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:05Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:05Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wwngs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T13:21:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-wbkjm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 28 13:21:06 crc kubenswrapper[4631]: I1128 13:21:06.282605 4631 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6d7f49b-96d5-4d70-a0e3-9d9f37f59073\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:20:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:20:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:20:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f11c70496a148e22446931e8183b52f16bd431cc7bd17e9678f875df424d6a03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T13:20:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35d912d041da9a7ed307ebea6e999cf087dacf84a3cefbfd0f3a9c891c979408\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T13:20:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d396f0adc112df728f80c8e1167526cdf1458a21e4802e2b70104f1a68086ff0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T13:20:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7cd0089f761e1315fe18e82695a8cf9170990bb535ea7235ee39b68edd79f23\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T13:20:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T13:20:43Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 28 13:21:06 crc kubenswrapper[4631]: I1128 13:21:06.293809 4631 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:02Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 28 13:21:06 crc kubenswrapper[4631]: I1128 13:21:06.307027 4631 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:02Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 28 13:21:06 crc kubenswrapper[4631]: I1128 13:21:06.315208 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 13:21:06 crc kubenswrapper[4631]: I1128 13:21:06.315271 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 13:21:06 crc kubenswrapper[4631]: I1128 13:21:06.315307 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 13:21:06 crc kubenswrapper[4631]: I1128 13:21:06.315332 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 13:21:06 crc kubenswrapper[4631]: I1128 13:21:06.315346 4631 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T13:21:06Z","lastTransitionTime":"2025-11-28T13:21:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 13:21:06 crc kubenswrapper[4631]: I1128 13:21:06.318650 4631 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:02Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 28 13:21:06 crc kubenswrapper[4631]: I1128 13:21:06.332914 4631 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-j8d66" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a346894-644f-4359-baa5-23bb2d0acc5f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2blnl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T13:21:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-j8d66\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 28 13:21:06 crc kubenswrapper[4631]: I1128 13:21:06.349000 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a205f475-340d-4ec5-a860-d4a5c2ceb39e-var-lib-openvswitch\") pod \"ovnkube-node-268tw\" (UID: \"a205f475-340d-4ec5-a860-d4a5c2ceb39e\") " pod="openshift-ovn-kubernetes/ovnkube-node-268tw" Nov 28 13:21:06 crc kubenswrapper[4631]: I1128 13:21:06.349059 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/a205f475-340d-4ec5-a860-d4a5c2ceb39e-systemd-units\") pod \"ovnkube-node-268tw\" (UID: \"a205f475-340d-4ec5-a860-d4a5c2ceb39e\") " pod="openshift-ovn-kubernetes/ovnkube-node-268tw" Nov 28 13:21:06 crc kubenswrapper[4631]: I1128 13:21:06.349083 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qdzgl\" (UniqueName: \"kubernetes.io/projected/a205f475-340d-4ec5-a860-d4a5c2ceb39e-kube-api-access-qdzgl\") pod \"ovnkube-node-268tw\" (UID: \"a205f475-340d-4ec5-a860-d4a5c2ceb39e\") " pod="openshift-ovn-kubernetes/ovnkube-node-268tw" Nov 28 13:21:06 crc kubenswrapper[4631]: I1128 13:21:06.349120 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/a205f475-340d-4ec5-a860-d4a5c2ceb39e-host-slash\") pod \"ovnkube-node-268tw\" (UID: \"a205f475-340d-4ec5-a860-d4a5c2ceb39e\") " pod="openshift-ovn-kubernetes/ovnkube-node-268tw" Nov 28 13:21:06 crc kubenswrapper[4631]: I1128 13:21:06.349144 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a205f475-340d-4ec5-a860-d4a5c2ceb39e-etc-openvswitch\") pod \"ovnkube-node-268tw\" (UID: \"a205f475-340d-4ec5-a860-d4a5c2ceb39e\") " pod="openshift-ovn-kubernetes/ovnkube-node-268tw" Nov 28 13:21:06 crc kubenswrapper[4631]: I1128 13:21:06.349169 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a205f475-340d-4ec5-a860-d4a5c2ceb39e-run-openvswitch\") pod \"ovnkube-node-268tw\" (UID: \"a205f475-340d-4ec5-a860-d4a5c2ceb39e\") " pod="openshift-ovn-kubernetes/ovnkube-node-268tw" Nov 28 13:21:06 crc kubenswrapper[4631]: I1128 13:21:06.349190 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/a205f475-340d-4ec5-a860-d4a5c2ceb39e-host-cni-netd\") pod \"ovnkube-node-268tw\" (UID: \"a205f475-340d-4ec5-a860-d4a5c2ceb39e\") " pod="openshift-ovn-kubernetes/ovnkube-node-268tw" Nov 28 13:21:06 crc kubenswrapper[4631]: I1128 13:21:06.349197 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/a205f475-340d-4ec5-a860-d4a5c2ceb39e-host-slash\") pod \"ovnkube-node-268tw\" (UID: \"a205f475-340d-4ec5-a860-d4a5c2ceb39e\") " pod="openshift-ovn-kubernetes/ovnkube-node-268tw" Nov 28 13:21:06 crc kubenswrapper[4631]: I1128 13:21:06.349240 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/a205f475-340d-4ec5-a860-d4a5c2ceb39e-run-ovn\") pod \"ovnkube-node-268tw\" (UID: \"a205f475-340d-4ec5-a860-d4a5c2ceb39e\") " pod="openshift-ovn-kubernetes/ovnkube-node-268tw" Nov 28 13:21:06 crc kubenswrapper[4631]: I1128 13:21:06.349319 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/a205f475-340d-4ec5-a860-d4a5c2ceb39e-systemd-units\") pod \"ovnkube-node-268tw\" (UID: \"a205f475-340d-4ec5-a860-d4a5c2ceb39e\") " pod="openshift-ovn-kubernetes/ovnkube-node-268tw" Nov 28 13:21:06 crc kubenswrapper[4631]: I1128 13:21:06.349362 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/a205f475-340d-4ec5-a860-d4a5c2ceb39e-node-log\") pod \"ovnkube-node-268tw\" (UID: \"a205f475-340d-4ec5-a860-d4a5c2ceb39e\") " pod="openshift-ovn-kubernetes/ovnkube-node-268tw" Nov 28 13:21:06 crc kubenswrapper[4631]: I1128 13:21:06.349342 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/a205f475-340d-4ec5-a860-d4a5c2ceb39e-node-log\") pod \"ovnkube-node-268tw\" (UID: \"a205f475-340d-4ec5-a860-d4a5c2ceb39e\") " pod="openshift-ovn-kubernetes/ovnkube-node-268tw" Nov 28 13:21:06 crc kubenswrapper[4631]: I1128 13:21:06.349383 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a205f475-340d-4ec5-a860-d4a5c2ceb39e-run-openvswitch\") pod \"ovnkube-node-268tw\" (UID: \"a205f475-340d-4ec5-a860-d4a5c2ceb39e\") " pod="openshift-ovn-kubernetes/ovnkube-node-268tw" Nov 28 13:21:06 crc kubenswrapper[4631]: I1128 13:21:06.349156 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a205f475-340d-4ec5-a860-d4a5c2ceb39e-var-lib-openvswitch\") pod \"ovnkube-node-268tw\" (UID: \"a205f475-340d-4ec5-a860-d4a5c2ceb39e\") " pod="openshift-ovn-kubernetes/ovnkube-node-268tw" Nov 28 13:21:06 crc kubenswrapper[4631]: I1128 13:21:06.349404 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/a205f475-340d-4ec5-a860-d4a5c2ceb39e-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-268tw\" (UID: \"a205f475-340d-4ec5-a860-d4a5c2ceb39e\") " pod="openshift-ovn-kubernetes/ovnkube-node-268tw" Nov 28 13:21:06 crc kubenswrapper[4631]: I1128 13:21:06.349338 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/a205f475-340d-4ec5-a860-d4a5c2ceb39e-run-ovn\") pod \"ovnkube-node-268tw\" (UID: \"a205f475-340d-4ec5-a860-d4a5c2ceb39e\") " pod="openshift-ovn-kubernetes/ovnkube-node-268tw" Nov 28 13:21:06 crc kubenswrapper[4631]: I1128 13:21:06.349429 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/a205f475-340d-4ec5-a860-d4a5c2ceb39e-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-268tw\" (UID: \"a205f475-340d-4ec5-a860-d4a5c2ceb39e\") " pod="openshift-ovn-kubernetes/ovnkube-node-268tw" Nov 28 13:21:06 crc kubenswrapper[4631]: I1128 13:21:06.349364 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a205f475-340d-4ec5-a860-d4a5c2ceb39e-etc-openvswitch\") pod \"ovnkube-node-268tw\" (UID: \"a205f475-340d-4ec5-a860-d4a5c2ceb39e\") " pod="openshift-ovn-kubernetes/ovnkube-node-268tw" Nov 28 13:21:06 crc kubenswrapper[4631]: I1128 13:21:06.349444 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/a205f475-340d-4ec5-a860-d4a5c2ceb39e-ovnkube-script-lib\") pod \"ovnkube-node-268tw\" (UID: \"a205f475-340d-4ec5-a860-d4a5c2ceb39e\") " pod="openshift-ovn-kubernetes/ovnkube-node-268tw" Nov 28 13:21:06 crc kubenswrapper[4631]: I1128 13:21:06.349625 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/a205f475-340d-4ec5-a860-d4a5c2ceb39e-host-run-netns\") pod \"ovnkube-node-268tw\" (UID: \"a205f475-340d-4ec5-a860-d4a5c2ceb39e\") " pod="openshift-ovn-kubernetes/ovnkube-node-268tw" Nov 28 13:21:06 crc kubenswrapper[4631]: I1128 13:21:06.349612 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/a205f475-340d-4ec5-a860-d4a5c2ceb39e-host-run-netns\") pod \"ovnkube-node-268tw\" (UID: \"a205f475-340d-4ec5-a860-d4a5c2ceb39e\") " pod="openshift-ovn-kubernetes/ovnkube-node-268tw" Nov 28 13:21:06 crc kubenswrapper[4631]: I1128 13:21:06.349661 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/a205f475-340d-4ec5-a860-d4a5c2ceb39e-ovnkube-config\") pod \"ovnkube-node-268tw\" (UID: \"a205f475-340d-4ec5-a860-d4a5c2ceb39e\") " pod="openshift-ovn-kubernetes/ovnkube-node-268tw" Nov 28 13:21:06 crc kubenswrapper[4631]: I1128 13:21:06.349828 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/a205f475-340d-4ec5-a860-d4a5c2ceb39e-host-cni-netd\") pod \"ovnkube-node-268tw\" (UID: \"a205f475-340d-4ec5-a860-d4a5c2ceb39e\") " pod="openshift-ovn-kubernetes/ovnkube-node-268tw" Nov 28 13:21:06 crc kubenswrapper[4631]: I1128 13:21:06.349914 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/a205f475-340d-4ec5-a860-d4a5c2ceb39e-run-systemd\") pod \"ovnkube-node-268tw\" (UID: \"a205f475-340d-4ec5-a860-d4a5c2ceb39e\") " pod="openshift-ovn-kubernetes/ovnkube-node-268tw" Nov 28 13:21:06 crc kubenswrapper[4631]: I1128 13:21:06.350020 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/a205f475-340d-4ec5-a860-d4a5c2ceb39e-host-run-ovn-kubernetes\") pod \"ovnkube-node-268tw\" (UID: \"a205f475-340d-4ec5-a860-d4a5c2ceb39e\") " pod="openshift-ovn-kubernetes/ovnkube-node-268tw" Nov 28 13:21:06 crc kubenswrapper[4631]: I1128 13:21:06.350098 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/a205f475-340d-4ec5-a860-d4a5c2ceb39e-env-overrides\") pod \"ovnkube-node-268tw\" (UID: \"a205f475-340d-4ec5-a860-d4a5c2ceb39e\") " pod="openshift-ovn-kubernetes/ovnkube-node-268tw" Nov 28 13:21:06 crc kubenswrapper[4631]: I1128 13:21:06.350169 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/a205f475-340d-4ec5-a860-d4a5c2ceb39e-ovn-node-metrics-cert\") pod \"ovnkube-node-268tw\" (UID: \"a205f475-340d-4ec5-a860-d4a5c2ceb39e\") " pod="openshift-ovn-kubernetes/ovnkube-node-268tw" Nov 28 13:21:06 crc kubenswrapper[4631]: I1128 13:21:06.350264 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/a205f475-340d-4ec5-a860-d4a5c2ceb39e-log-socket\") pod \"ovnkube-node-268tw\" (UID: \"a205f475-340d-4ec5-a860-d4a5c2ceb39e\") " pod="openshift-ovn-kubernetes/ovnkube-node-268tw" Nov 28 13:21:06 crc kubenswrapper[4631]: I1128 13:21:06.350472 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/a205f475-340d-4ec5-a860-d4a5c2ceb39e-host-kubelet\") pod \"ovnkube-node-268tw\" (UID: \"a205f475-340d-4ec5-a860-d4a5c2ceb39e\") " pod="openshift-ovn-kubernetes/ovnkube-node-268tw" Nov 28 13:21:06 crc kubenswrapper[4631]: I1128 13:21:06.350613 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/a205f475-340d-4ec5-a860-d4a5c2ceb39e-host-cni-bin\") pod \"ovnkube-node-268tw\" (UID: \"a205f475-340d-4ec5-a860-d4a5c2ceb39e\") " pod="openshift-ovn-kubernetes/ovnkube-node-268tw" Nov 28 13:21:06 crc kubenswrapper[4631]: I1128 13:21:06.350418 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/a205f475-340d-4ec5-a860-d4a5c2ceb39e-log-socket\") pod \"ovnkube-node-268tw\" (UID: \"a205f475-340d-4ec5-a860-d4a5c2ceb39e\") " pod="openshift-ovn-kubernetes/ovnkube-node-268tw" Nov 28 13:21:06 crc kubenswrapper[4631]: I1128 13:21:06.349938 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/a205f475-340d-4ec5-a860-d4a5c2ceb39e-run-systemd\") pod \"ovnkube-node-268tw\" (UID: \"a205f475-340d-4ec5-a860-d4a5c2ceb39e\") " pod="openshift-ovn-kubernetes/ovnkube-node-268tw" Nov 28 13:21:06 crc kubenswrapper[4631]: I1128 13:21:06.350579 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/a205f475-340d-4ec5-a860-d4a5c2ceb39e-host-kubelet\") pod \"ovnkube-node-268tw\" (UID: \"a205f475-340d-4ec5-a860-d4a5c2ceb39e\") " pod="openshift-ovn-kubernetes/ovnkube-node-268tw" Nov 28 13:21:06 crc kubenswrapper[4631]: I1128 13:21:06.350634 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/a205f475-340d-4ec5-a860-d4a5c2ceb39e-env-overrides\") pod \"ovnkube-node-268tw\" (UID: \"a205f475-340d-4ec5-a860-d4a5c2ceb39e\") " pod="openshift-ovn-kubernetes/ovnkube-node-268tw" Nov 28 13:21:06 crc kubenswrapper[4631]: I1128 13:21:06.350412 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/a205f475-340d-4ec5-a860-d4a5c2ceb39e-ovnkube-config\") pod \"ovnkube-node-268tw\" (UID: \"a205f475-340d-4ec5-a860-d4a5c2ceb39e\") " pod="openshift-ovn-kubernetes/ovnkube-node-268tw" Nov 28 13:21:06 crc kubenswrapper[4631]: I1128 13:21:06.350665 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/a205f475-340d-4ec5-a860-d4a5c2ceb39e-host-cni-bin\") pod \"ovnkube-node-268tw\" (UID: \"a205f475-340d-4ec5-a860-d4a5c2ceb39e\") " pod="openshift-ovn-kubernetes/ovnkube-node-268tw" Nov 28 13:21:06 crc kubenswrapper[4631]: I1128 13:21:06.350736 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/a205f475-340d-4ec5-a860-d4a5c2ceb39e-ovnkube-script-lib\") pod \"ovnkube-node-268tw\" (UID: \"a205f475-340d-4ec5-a860-d4a5c2ceb39e\") " pod="openshift-ovn-kubernetes/ovnkube-node-268tw" Nov 28 13:21:06 crc kubenswrapper[4631]: I1128 13:21:06.350099 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/a205f475-340d-4ec5-a860-d4a5c2ceb39e-host-run-ovn-kubernetes\") pod \"ovnkube-node-268tw\" (UID: \"a205f475-340d-4ec5-a860-d4a5c2ceb39e\") " pod="openshift-ovn-kubernetes/ovnkube-node-268tw" Nov 28 13:21:06 crc kubenswrapper[4631]: I1128 13:21:06.352233 4631 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-268tw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a205f475-340d-4ec5-a860-d4a5c2ceb39e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:06Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdzgl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdzgl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdzgl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdzgl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdzgl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdzgl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdzgl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdzgl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdzgl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T13:21:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-268tw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 28 13:21:06 crc kubenswrapper[4631]: I1128 13:21:06.418588 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 13:21:06 crc kubenswrapper[4631]: I1128 13:21:06.418641 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 13:21:06 crc kubenswrapper[4631]: I1128 13:21:06.418652 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 13:21:06 crc kubenswrapper[4631]: I1128 13:21:06.418741 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 13:21:06 crc kubenswrapper[4631]: I1128 13:21:06.418756 4631 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T13:21:06Z","lastTransitionTime":"2025-11-28T13:21:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 13:21:06 crc kubenswrapper[4631]: I1128 13:21:06.504932 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/a205f475-340d-4ec5-a860-d4a5c2ceb39e-ovn-node-metrics-cert\") pod \"ovnkube-node-268tw\" (UID: \"a205f475-340d-4ec5-a860-d4a5c2ceb39e\") " pod="openshift-ovn-kubernetes/ovnkube-node-268tw" Nov 28 13:21:06 crc kubenswrapper[4631]: I1128 13:21:06.505855 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qdzgl\" (UniqueName: \"kubernetes.io/projected/a205f475-340d-4ec5-a860-d4a5c2ceb39e-kube-api-access-qdzgl\") pod \"ovnkube-node-268tw\" (UID: \"a205f475-340d-4ec5-a860-d4a5c2ceb39e\") " pod="openshift-ovn-kubernetes/ovnkube-node-268tw" Nov 28 13:21:06 crc kubenswrapper[4631]: I1128 13:21:06.513045 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 13:21:06 crc kubenswrapper[4631]: I1128 13:21:06.513245 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 13:21:06 crc kubenswrapper[4631]: I1128 13:21:06.513096 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 13:21:06 crc kubenswrapper[4631]: E1128 13:21:06.513464 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 13:21:06 crc kubenswrapper[4631]: E1128 13:21:06.513584 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 13:21:06 crc kubenswrapper[4631]: E1128 13:21:06.513728 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 13:21:06 crc kubenswrapper[4631]: I1128 13:21:06.523131 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 13:21:06 crc kubenswrapper[4631]: I1128 13:21:06.523275 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 13:21:06 crc kubenswrapper[4631]: I1128 13:21:06.523382 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 13:21:06 crc kubenswrapper[4631]: I1128 13:21:06.523445 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 13:21:06 crc kubenswrapper[4631]: I1128 13:21:06.523513 4631 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T13:21:06Z","lastTransitionTime":"2025-11-28T13:21:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 13:21:06 crc kubenswrapper[4631]: I1128 13:21:06.626350 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 13:21:06 crc kubenswrapper[4631]: I1128 13:21:06.626389 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 13:21:06 crc kubenswrapper[4631]: I1128 13:21:06.626399 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 13:21:06 crc kubenswrapper[4631]: I1128 13:21:06.626419 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 13:21:06 crc kubenswrapper[4631]: I1128 13:21:06.626431 4631 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T13:21:06Z","lastTransitionTime":"2025-11-28T13:21:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 13:21:06 crc kubenswrapper[4631]: I1128 13:21:06.643874 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"c19e346e44c18c1937710aca15059221d7f4a46252fa90098870f100dd4157da"} Nov 28 13:21:06 crc kubenswrapper[4631]: I1128 13:21:06.646243 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-j8d66" event={"ID":"9a346894-644f-4359-baa5-23bb2d0acc5f","Type":"ContainerStarted","Data":"1e499b5d5f702b05bb24f616fb7d5ba1799d9b03e06c9efbfbbdcfd86e00e753"} Nov 28 13:21:06 crc kubenswrapper[4631]: I1128 13:21:06.646449 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-j8d66" event={"ID":"9a346894-644f-4359-baa5-23bb2d0acc5f","Type":"ContainerStarted","Data":"95ef7a20d6b6e57f7110525e13b38cbe56f74ff537e579b53d7e6c19fae05ee8"} Nov 28 13:21:06 crc kubenswrapper[4631]: I1128 13:21:06.650609 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" event={"ID":"8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1","Type":"ContainerStarted","Data":"7f613562449a8f5ebf17cd881967c47d44fb0296171712af40f42263f69f8136"} Nov 28 13:21:06 crc kubenswrapper[4631]: I1128 13:21:06.650734 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" event={"ID":"8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1","Type":"ContainerStarted","Data":"ec827042c58c204e08cc96b8d31f9c3ecff92049b65bdf29c8ae6577b44cd034"} Nov 28 13:21:06 crc kubenswrapper[4631]: I1128 13:21:06.656324 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-8gtw6" event={"ID":"a0fde998-9a55-4734-9008-136b93f29d0c","Type":"ContainerStarted","Data":"af2091e57fbae619a1f0bdbe59e8b35b23ae732d5d4dbe736284fbe566cb0c48"} Nov 28 13:21:06 crc kubenswrapper[4631]: I1128 13:21:06.656377 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-8gtw6" event={"ID":"a0fde998-9a55-4734-9008-136b93f29d0c","Type":"ContainerStarted","Data":"61c2f1dc159b4952e9adf9f8848e22c2a5b45fde03d2ef6777af3ddb855da91b"} Nov 28 13:21:06 crc kubenswrapper[4631]: I1128 13:21:06.657606 4631 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"07b4914f-d3c8-4e6a-909c-3af046e1759c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:20:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:20:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:20:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4f323ff537cd79159a51b592401e92fc06d54c490ff005486f27f41d42a35f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T13:20:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eff71f12d67a15adda3183a742f2869b104d773904a6c5954d1ccb6a2c64dd71\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T13:20:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://53ea681ea681973dd5177ec7f2925ef48fec88ac004b41fca867ce671f5a76d1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T13:20:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6138e1e9121a843fd730def8c3f019dd68a6f38d963ddb00124f2c532762e012\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T13:20:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ac30b44d2fe8ddb4ede530d4b99dab4dc641a097e21d4ae69cfcbeddc38802d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T13:20:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0b61d56ed070810021d253875a7bb5b9cb94a3446c56c4a92fb5418022dd643\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f0b61d56ed070810021d253875a7bb5b9cb94a3446c56c4a92fb5418022dd643\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T13:20:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T13:20:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T13:20:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 28 13:21:06 crc kubenswrapper[4631]: I1128 13:21:06.659776 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"754e658d99a10b103b24fc0244ad606402d50d072cf3c2bb93518afa4ab66935"} Nov 28 13:21:06 crc kubenswrapper[4631]: I1128 13:21:06.661409 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-wbkjm" event={"ID":"aa609d49-5fa2-4014-956c-dc31c9b35554","Type":"ContainerStarted","Data":"a2c72077eba60b5be9ecab590888229efc29a451f66f5588315b43b9bea6d37c"} Nov 28 13:21:06 crc kubenswrapper[4631]: I1128 13:21:06.673770 4631 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c19e346e44c18c1937710aca15059221d7f4a46252fa90098870f100dd4157da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T13:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 28 13:21:06 crc kubenswrapper[4631]: I1128 13:21:06.692340 4631 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:02Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T13:21:06Z is after 2025-08-24T17:21:41Z" Nov 28 13:21:06 crc kubenswrapper[4631]: I1128 13:21:06.707413 4631 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:02Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T13:21:06Z is after 2025-08-24T17:21:41Z" Nov 28 13:21:06 crc kubenswrapper[4631]: I1128 13:21:06.722896 4631 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-wbkjm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa609d49-5fa2-4014-956c-dc31c9b35554\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:05Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:05Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wwngs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T13:21:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-wbkjm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T13:21:06Z is after 2025-08-24T17:21:41Z" Nov 28 13:21:06 crc kubenswrapper[4631]: I1128 13:21:06.738597 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 13:21:06 crc kubenswrapper[4631]: I1128 13:21:06.738890 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 13:21:06 crc kubenswrapper[4631]: I1128 13:21:06.738984 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 13:21:06 crc kubenswrapper[4631]: I1128 13:21:06.739063 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 13:21:06 crc kubenswrapper[4631]: I1128 13:21:06.739135 4631 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T13:21:06Z","lastTransitionTime":"2025-11-28T13:21:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 13:21:06 crc kubenswrapper[4631]: I1128 13:21:06.745990 4631 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6d7f49b-96d5-4d70-a0e3-9d9f37f59073\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:20:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:20:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:20:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f11c70496a148e22446931e8183b52f16bd431cc7bd17e9678f875df424d6a03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T13:20:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35d912d041da9a7ed307ebea6e999cf087dacf84a3cefbfd0f3a9c891c979408\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T13:20:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d396f0adc112df728f80c8e1167526cdf1458a21e4802e2b70104f1a68086ff0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T13:20:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7cd0089f761e1315fe18e82695a8cf9170990bb535ea7235ee39b68edd79f23\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T13:20:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T13:20:43Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T13:21:06Z is after 2025-08-24T17:21:41Z" Nov 28 13:21:06 crc kubenswrapper[4631]: I1128 13:21:06.762674 4631 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:02Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T13:21:06Z is after 2025-08-24T17:21:41Z" Nov 28 13:21:06 crc kubenswrapper[4631]: I1128 13:21:06.766563 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-268tw" Nov 28 13:21:06 crc kubenswrapper[4631]: I1128 13:21:06.777954 4631 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:02Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T13:21:06Z is after 2025-08-24T17:21:41Z" Nov 28 13:21:06 crc kubenswrapper[4631]: I1128 13:21:06.796522 4631 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:02Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T13:21:06Z is after 2025-08-24T17:21:41Z" Nov 28 13:21:06 crc kubenswrapper[4631]: I1128 13:21:06.810630 4631 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-j8d66" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a346894-644f-4359-baa5-23bb2d0acc5f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2blnl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T13:21:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-j8d66\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T13:21:06Z is after 2025-08-24T17:21:41Z" Nov 28 13:21:06 crc kubenswrapper[4631]: I1128 13:21:06.834996 4631 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-268tw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a205f475-340d-4ec5-a860-d4a5c2ceb39e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:06Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdzgl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdzgl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdzgl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdzgl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdzgl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdzgl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdzgl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdzgl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdzgl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T13:21:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-268tw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T13:21:06Z is after 2025-08-24T17:21:41Z" Nov 28 13:21:06 crc kubenswrapper[4631]: I1128 13:21:06.841492 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 13:21:06 crc kubenswrapper[4631]: I1128 13:21:06.841654 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 13:21:06 crc kubenswrapper[4631]: I1128 13:21:06.841764 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 13:21:06 crc kubenswrapper[4631]: I1128 13:21:06.841882 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 13:21:06 crc kubenswrapper[4631]: I1128 13:21:06.841987 4631 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T13:21:06Z","lastTransitionTime":"2025-11-28T13:21:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 13:21:06 crc kubenswrapper[4631]: I1128 13:21:06.856182 4631 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-8gtw6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a0fde998-9a55-4734-9008-136b93f29d0c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:05Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h89l2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h89l2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h89l2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h89l2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h89l2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h89l2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h89l2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T13:21:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-8gtw6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T13:21:06Z is after 2025-08-24T17:21:41Z" Nov 28 13:21:06 crc kubenswrapper[4631]: I1128 13:21:06.869483 4631 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:05Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:05Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wtbl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wtbl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T13:21:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-47sxw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T13:21:06Z is after 2025-08-24T17:21:41Z" Nov 28 13:21:06 crc kubenswrapper[4631]: I1128 13:21:06.885816 4631 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:02Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T13:21:06Z is after 2025-08-24T17:21:41Z" Nov 28 13:21:06 crc kubenswrapper[4631]: I1128 13:21:06.898528 4631 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-wbkjm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa609d49-5fa2-4014-956c-dc31c9b35554\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a2c72077eba60b5be9ecab590888229efc29a451f66f5588315b43b9bea6d37c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T13:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wwngs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T13:21:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-wbkjm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T13:21:06Z is after 2025-08-24T17:21:41Z" Nov 28 13:21:06 crc kubenswrapper[4631]: I1128 13:21:06.914569 4631 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"07b4914f-d3c8-4e6a-909c-3af046e1759c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:20:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:20:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:20:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4f323ff537cd79159a51b592401e92fc06d54c490ff005486f27f41d42a35f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T13:20:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eff71f12d67a15adda3183a742f2869b104d773904a6c5954d1ccb6a2c64dd71\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T13:20:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://53ea681ea681973dd5177ec7f2925ef48fec88ac004b41fca867ce671f5a76d1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T13:20:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6138e1e9121a843fd730def8c3f019dd68a6f38d963ddb00124f2c532762e012\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T13:20:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ac30b44d2fe8ddb4ede530d4b99dab4dc641a097e21d4ae69cfcbeddc38802d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T13:20:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0b61d56ed070810021d253875a7bb5b9cb94a3446c56c4a92fb5418022dd643\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f0b61d56ed070810021d253875a7bb5b9cb94a3446c56c4a92fb5418022dd643\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T13:20:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T13:20:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T13:20:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T13:21:06Z is after 2025-08-24T17:21:41Z" Nov 28 13:21:06 crc kubenswrapper[4631]: I1128 13:21:06.927665 4631 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c19e346e44c18c1937710aca15059221d7f4a46252fa90098870f100dd4157da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T13:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T13:21:06Z is after 2025-08-24T17:21:41Z" Nov 28 13:21:06 crc kubenswrapper[4631]: I1128 13:21:06.938542 4631 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:02Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T13:21:06Z is after 2025-08-24T17:21:41Z" Nov 28 13:21:06 crc kubenswrapper[4631]: I1128 13:21:06.944085 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 13:21:06 crc kubenswrapper[4631]: I1128 13:21:06.944330 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 13:21:06 crc kubenswrapper[4631]: I1128 13:21:06.944459 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 13:21:06 crc kubenswrapper[4631]: I1128 13:21:06.944576 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 13:21:06 crc kubenswrapper[4631]: I1128 13:21:06.944692 4631 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T13:21:06Z","lastTransitionTime":"2025-11-28T13:21:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 13:21:06 crc kubenswrapper[4631]: I1128 13:21:06.954410 4631 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6d7f49b-96d5-4d70-a0e3-9d9f37f59073\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:20:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:20:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:20:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f11c70496a148e22446931e8183b52f16bd431cc7bd17e9678f875df424d6a03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T13:20:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35d912d041da9a7ed307ebea6e999cf087dacf84a3cefbfd0f3a9c891c979408\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T13:20:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d396f0adc112df728f80c8e1167526cdf1458a21e4802e2b70104f1a68086ff0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T13:20:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7cd0089f761e1315fe18e82695a8cf9170990bb535ea7235ee39b68edd79f23\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T13:20:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T13:20:43Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T13:21:06Z is after 2025-08-24T17:21:41Z" Nov 28 13:21:06 crc kubenswrapper[4631]: I1128 13:21:06.966054 4631 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:02Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T13:21:06Z is after 2025-08-24T17:21:41Z" Nov 28 13:21:06 crc kubenswrapper[4631]: I1128 13:21:06.976746 4631 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:02Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T13:21:06Z is after 2025-08-24T17:21:41Z" Nov 28 13:21:06 crc kubenswrapper[4631]: I1128 13:21:06.987770 4631 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:02Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T13:21:06Z is after 2025-08-24T17:21:41Z" Nov 28 13:21:06 crc kubenswrapper[4631]: I1128 13:21:06.999557 4631 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-j8d66" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a346894-644f-4359-baa5-23bb2d0acc5f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e499b5d5f702b05bb24f616fb7d5ba1799d9b03e06c9efbfbbdcfd86e00e753\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T13:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2blnl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T13:21:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-j8d66\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T13:21:06Z is after 2025-08-24T17:21:41Z" Nov 28 13:21:07 crc kubenswrapper[4631]: I1128 13:21:07.016091 4631 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-268tw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a205f475-340d-4ec5-a860-d4a5c2ceb39e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:06Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdzgl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdzgl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdzgl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdzgl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdzgl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdzgl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdzgl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdzgl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdzgl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T13:21:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-268tw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T13:21:07Z is after 2025-08-24T17:21:41Z" Nov 28 13:21:07 crc kubenswrapper[4631]: I1128 13:21:07.029260 4631 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-8gtw6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a0fde998-9a55-4734-9008-136b93f29d0c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:05Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h89l2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h89l2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h89l2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h89l2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h89l2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h89l2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h89l2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T13:21:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-8gtw6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T13:21:07Z is after 2025-08-24T17:21:41Z" Nov 28 13:21:07 crc kubenswrapper[4631]: I1128 13:21:07.038991 4631 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:05Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:05Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wtbl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wtbl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T13:21:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-47sxw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T13:21:07Z is after 2025-08-24T17:21:41Z" Nov 28 13:21:07 crc kubenswrapper[4631]: I1128 13:21:07.047811 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 13:21:07 crc kubenswrapper[4631]: I1128 13:21:07.048110 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 13:21:07 crc kubenswrapper[4631]: I1128 13:21:07.048353 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 13:21:07 crc kubenswrapper[4631]: I1128 13:21:07.048571 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 13:21:07 crc kubenswrapper[4631]: I1128 13:21:07.048760 4631 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T13:21:07Z","lastTransitionTime":"2025-11-28T13:21:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 13:21:07 crc kubenswrapper[4631]: I1128 13:21:07.150801 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 13:21:07 crc kubenswrapper[4631]: I1128 13:21:07.150846 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 13:21:07 crc kubenswrapper[4631]: I1128 13:21:07.150855 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 13:21:07 crc kubenswrapper[4631]: I1128 13:21:07.150871 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 13:21:07 crc kubenswrapper[4631]: I1128 13:21:07.150883 4631 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T13:21:07Z","lastTransitionTime":"2025-11-28T13:21:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 13:21:07 crc kubenswrapper[4631]: I1128 13:21:07.253523 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 13:21:07 crc kubenswrapper[4631]: I1128 13:21:07.253578 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 13:21:07 crc kubenswrapper[4631]: I1128 13:21:07.253591 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 13:21:07 crc kubenswrapper[4631]: I1128 13:21:07.253611 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 13:21:07 crc kubenswrapper[4631]: I1128 13:21:07.253627 4631 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T13:21:07Z","lastTransitionTime":"2025-11-28T13:21:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 13:21:07 crc kubenswrapper[4631]: I1128 13:21:07.270587 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-cbcjn"] Nov 28 13:21:07 crc kubenswrapper[4631]: I1128 13:21:07.271084 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-cbcjn" Nov 28 13:21:07 crc kubenswrapper[4631]: I1128 13:21:07.272756 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Nov 28 13:21:07 crc kubenswrapper[4631]: I1128 13:21:07.274390 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Nov 28 13:21:07 crc kubenswrapper[4631]: I1128 13:21:07.274505 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Nov 28 13:21:07 crc kubenswrapper[4631]: I1128 13:21:07.275174 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Nov 28 13:21:07 crc kubenswrapper[4631]: I1128 13:21:07.296225 4631 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-8gtw6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a0fde998-9a55-4734-9008-136b93f29d0c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:05Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h89l2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h89l2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h89l2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h89l2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h89l2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h89l2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h89l2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T13:21:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-8gtw6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T13:21:07Z is after 2025-08-24T17:21:41Z" Nov 28 13:21:07 crc kubenswrapper[4631]: I1128 13:21:07.321265 4631 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:05Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:05Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wtbl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wtbl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T13:21:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-47sxw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T13:21:07Z is after 2025-08-24T17:21:41Z" Nov 28 13:21:07 crc kubenswrapper[4631]: I1128 13:21:07.343636 4631 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:02Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T13:21:07Z is after 2025-08-24T17:21:41Z" Nov 28 13:21:07 crc kubenswrapper[4631]: I1128 13:21:07.356454 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 13:21:07 crc kubenswrapper[4631]: I1128 13:21:07.356495 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 13:21:07 crc kubenswrapper[4631]: I1128 13:21:07.356505 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 13:21:07 crc kubenswrapper[4631]: I1128 13:21:07.356522 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 13:21:07 crc kubenswrapper[4631]: I1128 13:21:07.356533 4631 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T13:21:07Z","lastTransitionTime":"2025-11-28T13:21:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 13:21:07 crc kubenswrapper[4631]: I1128 13:21:07.358392 4631 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-wbkjm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa609d49-5fa2-4014-956c-dc31c9b35554\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a2c72077eba60b5be9ecab590888229efc29a451f66f5588315b43b9bea6d37c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T13:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wwngs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T13:21:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-wbkjm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T13:21:07Z is after 2025-08-24T17:21:41Z" Nov 28 13:21:07 crc kubenswrapper[4631]: I1128 13:21:07.364875 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vk9nm\" (UniqueName: \"kubernetes.io/projected/fe4682b5-79b2-456b-b2b8-4fecbe51e5da-kube-api-access-vk9nm\") pod \"node-ca-cbcjn\" (UID: \"fe4682b5-79b2-456b-b2b8-4fecbe51e5da\") " pod="openshift-image-registry/node-ca-cbcjn" Nov 28 13:21:07 crc kubenswrapper[4631]: I1128 13:21:07.364956 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/fe4682b5-79b2-456b-b2b8-4fecbe51e5da-host\") pod \"node-ca-cbcjn\" (UID: \"fe4682b5-79b2-456b-b2b8-4fecbe51e5da\") " pod="openshift-image-registry/node-ca-cbcjn" Nov 28 13:21:07 crc kubenswrapper[4631]: I1128 13:21:07.365015 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/fe4682b5-79b2-456b-b2b8-4fecbe51e5da-serviceca\") pod \"node-ca-cbcjn\" (UID: \"fe4682b5-79b2-456b-b2b8-4fecbe51e5da\") " pod="openshift-image-registry/node-ca-cbcjn" Nov 28 13:21:07 crc kubenswrapper[4631]: I1128 13:21:07.386100 4631 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"07b4914f-d3c8-4e6a-909c-3af046e1759c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:20:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:20:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:20:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4f323ff537cd79159a51b592401e92fc06d54c490ff005486f27f41d42a35f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T13:20:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eff71f12d67a15adda3183a742f2869b104d773904a6c5954d1ccb6a2c64dd71\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T13:20:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://53ea681ea681973dd5177ec7f2925ef48fec88ac004b41fca867ce671f5a76d1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T13:20:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6138e1e9121a843fd730def8c3f019dd68a6f38d963ddb00124f2c532762e012\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T13:20:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ac30b44d2fe8ddb4ede530d4b99dab4dc641a097e21d4ae69cfcbeddc38802d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T13:20:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0b61d56ed070810021d253875a7bb5b9cb94a3446c56c4a92fb5418022dd643\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f0b61d56ed070810021d253875a7bb5b9cb94a3446c56c4a92fb5418022dd643\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T13:20:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T13:20:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T13:20:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T13:21:07Z is after 2025-08-24T17:21:41Z" Nov 28 13:21:07 crc kubenswrapper[4631]: I1128 13:21:07.397462 4631 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c19e346e44c18c1937710aca15059221d7f4a46252fa90098870f100dd4157da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T13:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T13:21:07Z is after 2025-08-24T17:21:41Z" Nov 28 13:21:07 crc kubenswrapper[4631]: I1128 13:21:07.414014 4631 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:02Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T13:21:07Z is after 2025-08-24T17:21:41Z" Nov 28 13:21:07 crc kubenswrapper[4631]: I1128 13:21:07.428431 4631 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6d7f49b-96d5-4d70-a0e3-9d9f37f59073\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:20:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:20:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:20:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f11c70496a148e22446931e8183b52f16bd431cc7bd17e9678f875df424d6a03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T13:20:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35d912d041da9a7ed307ebea6e999cf087dacf84a3cefbfd0f3a9c891c979408\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T13:20:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d396f0adc112df728f80c8e1167526cdf1458a21e4802e2b70104f1a68086ff0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T13:20:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7cd0089f761e1315fe18e82695a8cf9170990bb535ea7235ee39b68edd79f23\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T13:20:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T13:20:43Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T13:21:07Z is after 2025-08-24T17:21:41Z" Nov 28 13:21:07 crc kubenswrapper[4631]: I1128 13:21:07.441619 4631 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:02Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T13:21:07Z is after 2025-08-24T17:21:41Z" Nov 28 13:21:07 crc kubenswrapper[4631]: I1128 13:21:07.459987 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 13:21:07 crc kubenswrapper[4631]: I1128 13:21:07.460032 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 13:21:07 crc kubenswrapper[4631]: I1128 13:21:07.460068 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 13:21:07 crc kubenswrapper[4631]: I1128 13:21:07.460087 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 13:21:07 crc kubenswrapper[4631]: I1128 13:21:07.460097 4631 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T13:21:07Z","lastTransitionTime":"2025-11-28T13:21:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 13:21:07 crc kubenswrapper[4631]: I1128 13:21:07.465824 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/fe4682b5-79b2-456b-b2b8-4fecbe51e5da-serviceca\") pod \"node-ca-cbcjn\" (UID: \"fe4682b5-79b2-456b-b2b8-4fecbe51e5da\") " pod="openshift-image-registry/node-ca-cbcjn" Nov 28 13:21:07 crc kubenswrapper[4631]: I1128 13:21:07.465889 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vk9nm\" (UniqueName: \"kubernetes.io/projected/fe4682b5-79b2-456b-b2b8-4fecbe51e5da-kube-api-access-vk9nm\") pod \"node-ca-cbcjn\" (UID: \"fe4682b5-79b2-456b-b2b8-4fecbe51e5da\") " pod="openshift-image-registry/node-ca-cbcjn" Nov 28 13:21:07 crc kubenswrapper[4631]: I1128 13:21:07.465928 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/fe4682b5-79b2-456b-b2b8-4fecbe51e5da-host\") pod \"node-ca-cbcjn\" (UID: \"fe4682b5-79b2-456b-b2b8-4fecbe51e5da\") " pod="openshift-image-registry/node-ca-cbcjn" Nov 28 13:21:07 crc kubenswrapper[4631]: I1128 13:21:07.466025 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/fe4682b5-79b2-456b-b2b8-4fecbe51e5da-host\") pod \"node-ca-cbcjn\" (UID: \"fe4682b5-79b2-456b-b2b8-4fecbe51e5da\") " pod="openshift-image-registry/node-ca-cbcjn" Nov 28 13:21:07 crc kubenswrapper[4631]: I1128 13:21:07.466859 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/fe4682b5-79b2-456b-b2b8-4fecbe51e5da-serviceca\") pod \"node-ca-cbcjn\" (UID: \"fe4682b5-79b2-456b-b2b8-4fecbe51e5da\") " pod="openshift-image-registry/node-ca-cbcjn" Nov 28 13:21:07 crc kubenswrapper[4631]: I1128 13:21:07.482858 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vk9nm\" (UniqueName: \"kubernetes.io/projected/fe4682b5-79b2-456b-b2b8-4fecbe51e5da-kube-api-access-vk9nm\") pod \"node-ca-cbcjn\" (UID: \"fe4682b5-79b2-456b-b2b8-4fecbe51e5da\") " pod="openshift-image-registry/node-ca-cbcjn" Nov 28 13:21:07 crc kubenswrapper[4631]: I1128 13:21:07.483983 4631 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:02Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T13:21:07Z is after 2025-08-24T17:21:41Z" Nov 28 13:21:07 crc kubenswrapper[4631]: I1128 13:21:07.497809 4631 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:02Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T13:21:07Z is after 2025-08-24T17:21:41Z" Nov 28 13:21:07 crc kubenswrapper[4631]: I1128 13:21:07.509992 4631 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-j8d66" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a346894-644f-4359-baa5-23bb2d0acc5f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e499b5d5f702b05bb24f616fb7d5ba1799d9b03e06c9efbfbbdcfd86e00e753\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T13:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2blnl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T13:21:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-j8d66\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T13:21:07Z is after 2025-08-24T17:21:41Z" Nov 28 13:21:07 crc kubenswrapper[4631]: I1128 13:21:07.535245 4631 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-268tw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a205f475-340d-4ec5-a860-d4a5c2ceb39e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:06Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdzgl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdzgl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdzgl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdzgl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdzgl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdzgl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdzgl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdzgl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdzgl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T13:21:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-268tw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T13:21:07Z is after 2025-08-24T17:21:41Z" Nov 28 13:21:07 crc kubenswrapper[4631]: I1128 13:21:07.553739 4631 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-cbcjn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fe4682b5-79b2-456b-b2b8-4fecbe51e5da\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:07Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:07Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vk9nm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T13:21:07Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-cbcjn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T13:21:07Z is after 2025-08-24T17:21:41Z" Nov 28 13:21:07 crc kubenswrapper[4631]: I1128 13:21:07.562629 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 13:21:07 crc kubenswrapper[4631]: I1128 13:21:07.562669 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 13:21:07 crc kubenswrapper[4631]: I1128 13:21:07.562679 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 13:21:07 crc kubenswrapper[4631]: I1128 13:21:07.562695 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 13:21:07 crc kubenswrapper[4631]: I1128 13:21:07.562708 4631 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T13:21:07Z","lastTransitionTime":"2025-11-28T13:21:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 13:21:07 crc kubenswrapper[4631]: I1128 13:21:07.666341 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 13:21:07 crc kubenswrapper[4631]: I1128 13:21:07.666401 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 13:21:07 crc kubenswrapper[4631]: I1128 13:21:07.666414 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 13:21:07 crc kubenswrapper[4631]: I1128 13:21:07.666440 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 13:21:07 crc kubenswrapper[4631]: I1128 13:21:07.666458 4631 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T13:21:07Z","lastTransitionTime":"2025-11-28T13:21:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 13:21:07 crc kubenswrapper[4631]: I1128 13:21:07.666677 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" event={"ID":"8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1","Type":"ContainerStarted","Data":"7fb5d8adf7fbc8de46da5e68668ac8f1a3083ca93ac373f4626c6ba1cb6e07d8"} Nov 28 13:21:07 crc kubenswrapper[4631]: I1128 13:21:07.668103 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"8235c1ea6445a3e88ee3e0f57ce76e9840dfa95421832d59a3bc2e19952850db"} Nov 28 13:21:07 crc kubenswrapper[4631]: I1128 13:21:07.669739 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"f4f7a0832e3d84f17976d4edf685abcfe1c4e4a8333d232899a591e329e5eaf4"} Nov 28 13:21:07 crc kubenswrapper[4631]: I1128 13:21:07.671003 4631 generic.go:334] "Generic (PLEG): container finished" podID="a205f475-340d-4ec5-a860-d4a5c2ceb39e" containerID="542662c37344400bc1d05230e709812104866735429d37ed0d7894050e7a54e8" exitCode=0 Nov 28 13:21:07 crc kubenswrapper[4631]: I1128 13:21:07.671082 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-268tw" event={"ID":"a205f475-340d-4ec5-a860-d4a5c2ceb39e","Type":"ContainerDied","Data":"542662c37344400bc1d05230e709812104866735429d37ed0d7894050e7a54e8"} Nov 28 13:21:07 crc kubenswrapper[4631]: I1128 13:21:07.671124 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-268tw" event={"ID":"a205f475-340d-4ec5-a860-d4a5c2ceb39e","Type":"ContainerStarted","Data":"11131235ce93081bf60d5d43d90a0fe910b89daf18e38b53128dcd441a1c0194"} Nov 28 13:21:07 crc kubenswrapper[4631]: I1128 13:21:07.672890 4631 generic.go:334] "Generic (PLEG): container finished" podID="a0fde998-9a55-4734-9008-136b93f29d0c" containerID="af2091e57fbae619a1f0bdbe59e8b35b23ae732d5d4dbe736284fbe566cb0c48" exitCode=0 Nov 28 13:21:07 crc kubenswrapper[4631]: I1128 13:21:07.672994 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-8gtw6" event={"ID":"a0fde998-9a55-4734-9008-136b93f29d0c","Type":"ContainerDied","Data":"af2091e57fbae619a1f0bdbe59e8b35b23ae732d5d4dbe736284fbe566cb0c48"} Nov 28 13:21:07 crc kubenswrapper[4631]: I1128 13:21:07.724627 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-cbcjn" Nov 28 13:21:07 crc kubenswrapper[4631]: I1128 13:21:07.726428 4631 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-268tw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a205f475-340d-4ec5-a860-d4a5c2ceb39e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:06Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdzgl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdzgl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdzgl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdzgl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdzgl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdzgl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdzgl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdzgl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdzgl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T13:21:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-268tw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T13:21:07Z is after 2025-08-24T17:21:41Z" Nov 28 13:21:07 crc kubenswrapper[4631]: I1128 13:21:07.750967 4631 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-cbcjn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fe4682b5-79b2-456b-b2b8-4fecbe51e5da\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:07Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:07Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vk9nm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T13:21:07Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-cbcjn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T13:21:07Z is after 2025-08-24T17:21:41Z" Nov 28 13:21:07 crc kubenswrapper[4631]: I1128 13:21:07.776907 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 13:21:07 crc kubenswrapper[4631]: I1128 13:21:07.777392 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 13:21:07 crc kubenswrapper[4631]: I1128 13:21:07.777406 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 13:21:07 crc kubenswrapper[4631]: I1128 13:21:07.777427 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 13:21:07 crc kubenswrapper[4631]: I1128 13:21:07.777438 4631 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T13:21:07Z","lastTransitionTime":"2025-11-28T13:21:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 13:21:07 crc kubenswrapper[4631]: I1128 13:21:07.810810 4631 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:02Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T13:21:07Z is after 2025-08-24T17:21:41Z" Nov 28 13:21:07 crc kubenswrapper[4631]: I1128 13:21:07.838914 4631 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-j8d66" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a346894-644f-4359-baa5-23bb2d0acc5f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e499b5d5f702b05bb24f616fb7d5ba1799d9b03e06c9efbfbbdcfd86e00e753\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T13:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2blnl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T13:21:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-j8d66\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T13:21:07Z is after 2025-08-24T17:21:41Z" Nov 28 13:21:07 crc kubenswrapper[4631]: I1128 13:21:07.873910 4631 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-8gtw6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a0fde998-9a55-4734-9008-136b93f29d0c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:05Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h89l2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h89l2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h89l2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h89l2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h89l2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h89l2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h89l2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T13:21:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-8gtw6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T13:21:07Z is after 2025-08-24T17:21:41Z" Nov 28 13:21:07 crc kubenswrapper[4631]: I1128 13:21:07.883968 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 13:21:07 crc kubenswrapper[4631]: I1128 13:21:07.884003 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 13:21:07 crc kubenswrapper[4631]: I1128 13:21:07.884011 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 13:21:07 crc kubenswrapper[4631]: I1128 13:21:07.884024 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 13:21:07 crc kubenswrapper[4631]: I1128 13:21:07.884033 4631 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T13:21:07Z","lastTransitionTime":"2025-11-28T13:21:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 13:21:07 crc kubenswrapper[4631]: I1128 13:21:07.914140 4631 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fb5d8adf7fbc8de46da5e68668ac8f1a3083ca93ac373f4626c6ba1cb6e07d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T13:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wtbl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f613562449a8f5ebf17cd881967c47d44fb0296171712af40f42263f69f8136\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T13:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wtbl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T13:21:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-47sxw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T13:21:07Z is after 2025-08-24T17:21:41Z" Nov 28 13:21:07 crc kubenswrapper[4631]: I1128 13:21:07.939976 4631 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"07b4914f-d3c8-4e6a-909c-3af046e1759c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:20:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:20:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:20:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4f323ff537cd79159a51b592401e92fc06d54c490ff005486f27f41d42a35f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T13:20:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eff71f12d67a15adda3183a742f2869b104d773904a6c5954d1ccb6a2c64dd71\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T13:20:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://53ea681ea681973dd5177ec7f2925ef48fec88ac004b41fca867ce671f5a76d1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T13:20:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6138e1e9121a843fd730def8c3f019dd68a6f38d963ddb00124f2c532762e012\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T13:20:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ac30b44d2fe8ddb4ede530d4b99dab4dc641a097e21d4ae69cfcbeddc38802d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T13:20:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0b61d56ed070810021d253875a7bb5b9cb94a3446c56c4a92fb5418022dd643\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f0b61d56ed070810021d253875a7bb5b9cb94a3446c56c4a92fb5418022dd643\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T13:20:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T13:20:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T13:20:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T13:21:07Z is after 2025-08-24T17:21:41Z" Nov 28 13:21:07 crc kubenswrapper[4631]: I1128 13:21:07.961418 4631 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c19e346e44c18c1937710aca15059221d7f4a46252fa90098870f100dd4157da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T13:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T13:21:07Z is after 2025-08-24T17:21:41Z" Nov 28 13:21:07 crc kubenswrapper[4631]: I1128 13:21:07.977256 4631 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:02Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T13:21:07Z is after 2025-08-24T17:21:41Z" Nov 28 13:21:07 crc kubenswrapper[4631]: I1128 13:21:07.985784 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 13:21:07 crc kubenswrapper[4631]: I1128 13:21:07.985806 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 13:21:07 crc kubenswrapper[4631]: I1128 13:21:07.985813 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 13:21:07 crc kubenswrapper[4631]: I1128 13:21:07.985826 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 13:21:07 crc kubenswrapper[4631]: I1128 13:21:07.985835 4631 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T13:21:07Z","lastTransitionTime":"2025-11-28T13:21:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 13:21:07 crc kubenswrapper[4631]: I1128 13:21:07.990915 4631 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:02Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T13:21:07Z is after 2025-08-24T17:21:41Z" Nov 28 13:21:08 crc kubenswrapper[4631]: I1128 13:21:08.005204 4631 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-wbkjm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa609d49-5fa2-4014-956c-dc31c9b35554\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a2c72077eba60b5be9ecab590888229efc29a451f66f5588315b43b9bea6d37c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T13:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wwngs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T13:21:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-wbkjm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T13:21:08Z is after 2025-08-24T17:21:41Z" Nov 28 13:21:08 crc kubenswrapper[4631]: I1128 13:21:08.022801 4631 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:02Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T13:21:08Z is after 2025-08-24T17:21:41Z" Nov 28 13:21:08 crc kubenswrapper[4631]: I1128 13:21:08.044310 4631 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:02Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T13:21:08Z is after 2025-08-24T17:21:41Z" Nov 28 13:21:08 crc kubenswrapper[4631]: I1128 13:21:08.057710 4631 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6d7f49b-96d5-4d70-a0e3-9d9f37f59073\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:20:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:20:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:20:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f11c70496a148e22446931e8183b52f16bd431cc7bd17e9678f875df424d6a03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T13:20:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35d912d041da9a7ed307ebea6e999cf087dacf84a3cefbfd0f3a9c891c979408\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T13:20:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d396f0adc112df728f80c8e1167526cdf1458a21e4802e2b70104f1a68086ff0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T13:20:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7cd0089f761e1315fe18e82695a8cf9170990bb535ea7235ee39b68edd79f23\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T13:20:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T13:20:43Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T13:21:08Z is after 2025-08-24T17:21:41Z" Nov 28 13:21:08 crc kubenswrapper[4631]: I1128 13:21:08.078446 4631 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-8gtw6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a0fde998-9a55-4734-9008-136b93f29d0c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:05Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h89l2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af2091e57fbae619a1f0bdbe59e8b35b23ae732d5d4dbe736284fbe566cb0c48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af2091e57fbae619a1f0bdbe59e8b35b23ae732d5d4dbe736284fbe566cb0c48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T13:21:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T13:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h89l2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h89l2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h89l2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h89l2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h89l2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h89l2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T13:21:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-8gtw6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T13:21:08Z is after 2025-08-24T17:21:41Z" Nov 28 13:21:08 crc kubenswrapper[4631]: I1128 13:21:08.090622 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 13:21:08 crc kubenswrapper[4631]: I1128 13:21:08.090656 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 13:21:08 crc kubenswrapper[4631]: I1128 13:21:08.090670 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 13:21:08 crc kubenswrapper[4631]: I1128 13:21:08.090688 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 13:21:08 crc kubenswrapper[4631]: I1128 13:21:08.090701 4631 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T13:21:08Z","lastTransitionTime":"2025-11-28T13:21:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 13:21:08 crc kubenswrapper[4631]: I1128 13:21:08.111500 4631 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fb5d8adf7fbc8de46da5e68668ac8f1a3083ca93ac373f4626c6ba1cb6e07d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T13:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wtbl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f613562449a8f5ebf17cd881967c47d44fb0296171712af40f42263f69f8136\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T13:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wtbl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T13:21:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-47sxw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T13:21:08Z is after 2025-08-24T17:21:41Z" Nov 28 13:21:08 crc kubenswrapper[4631]: I1128 13:21:08.140932 4631 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c19e346e44c18c1937710aca15059221d7f4a46252fa90098870f100dd4157da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T13:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T13:21:08Z is after 2025-08-24T17:21:41Z" Nov 28 13:21:08 crc kubenswrapper[4631]: I1128 13:21:08.155974 4631 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:02Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T13:21:08Z is after 2025-08-24T17:21:41Z" Nov 28 13:21:08 crc kubenswrapper[4631]: I1128 13:21:08.174335 4631 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8235c1ea6445a3e88ee3e0f57ce76e9840dfa95421832d59a3bc2e19952850db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T13:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://754e658d99a10b103b24fc0244ad606402d50d072cf3c2bb93518afa4ab66935\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T13:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T13:21:08Z is after 2025-08-24T17:21:41Z" Nov 28 13:21:08 crc kubenswrapper[4631]: I1128 13:21:08.190176 4631 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-wbkjm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa609d49-5fa2-4014-956c-dc31c9b35554\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a2c72077eba60b5be9ecab590888229efc29a451f66f5588315b43b9bea6d37c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T13:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wwngs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T13:21:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-wbkjm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T13:21:08Z is after 2025-08-24T17:21:41Z" Nov 28 13:21:08 crc kubenswrapper[4631]: I1128 13:21:08.192834 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 13:21:08 crc kubenswrapper[4631]: I1128 13:21:08.192862 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 13:21:08 crc kubenswrapper[4631]: I1128 13:21:08.192873 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 13:21:08 crc kubenswrapper[4631]: I1128 13:21:08.192889 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 13:21:08 crc kubenswrapper[4631]: I1128 13:21:08.192900 4631 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T13:21:08Z","lastTransitionTime":"2025-11-28T13:21:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 13:21:08 crc kubenswrapper[4631]: I1128 13:21:08.214765 4631 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"07b4914f-d3c8-4e6a-909c-3af046e1759c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:20:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:20:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:20:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4f323ff537cd79159a51b592401e92fc06d54c490ff005486f27f41d42a35f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T13:20:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eff71f12d67a15adda3183a742f2869b104d773904a6c5954d1ccb6a2c64dd71\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T13:20:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://53ea681ea681973dd5177ec7f2925ef48fec88ac004b41fca867ce671f5a76d1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T13:20:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6138e1e9121a843fd730def8c3f019dd68a6f38d963ddb00124f2c532762e012\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T13:20:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ac30b44d2fe8ddb4ede530d4b99dab4dc641a097e21d4ae69cfcbeddc38802d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T13:20:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0b61d56ed070810021d253875a7bb5b9cb94a3446c56c4a92fb5418022dd643\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f0b61d56ed070810021d253875a7bb5b9cb94a3446c56c4a92fb5418022dd643\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T13:20:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T13:20:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T13:20:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T13:21:08Z is after 2025-08-24T17:21:41Z" Nov 28 13:21:08 crc kubenswrapper[4631]: I1128 13:21:08.230104 4631 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4f7a0832e3d84f17976d4edf685abcfe1c4e4a8333d232899a591e329e5eaf4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T13:21:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T13:21:08Z is after 2025-08-24T17:21:41Z" Nov 28 13:21:08 crc kubenswrapper[4631]: I1128 13:21:08.244809 4631 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6d7f49b-96d5-4d70-a0e3-9d9f37f59073\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:20:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:20:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:20:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f11c70496a148e22446931e8183b52f16bd431cc7bd17e9678f875df424d6a03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T13:20:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35d912d041da9a7ed307ebea6e999cf087dacf84a3cefbfd0f3a9c891c979408\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T13:20:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d396f0adc112df728f80c8e1167526cdf1458a21e4802e2b70104f1a68086ff0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T13:20:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7cd0089f761e1315fe18e82695a8cf9170990bb535ea7235ee39b68edd79f23\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T13:20:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T13:20:43Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T13:21:08Z is after 2025-08-24T17:21:41Z" Nov 28 13:21:08 crc kubenswrapper[4631]: I1128 13:21:08.259042 4631 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:02Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T13:21:08Z is after 2025-08-24T17:21:41Z" Nov 28 13:21:08 crc kubenswrapper[4631]: I1128 13:21:08.286164 4631 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-cbcjn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fe4682b5-79b2-456b-b2b8-4fecbe51e5da\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:07Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:07Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vk9nm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T13:21:07Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-cbcjn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T13:21:08Z is after 2025-08-24T17:21:41Z" Nov 28 13:21:08 crc kubenswrapper[4631]: I1128 13:21:08.294882 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 13:21:08 crc kubenswrapper[4631]: I1128 13:21:08.294918 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 13:21:08 crc kubenswrapper[4631]: I1128 13:21:08.294948 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 13:21:08 crc kubenswrapper[4631]: I1128 13:21:08.294963 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 13:21:08 crc kubenswrapper[4631]: I1128 13:21:08.294984 4631 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T13:21:08Z","lastTransitionTime":"2025-11-28T13:21:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 13:21:08 crc kubenswrapper[4631]: I1128 13:21:08.327191 4631 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:02Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T13:21:08Z is after 2025-08-24T17:21:41Z" Nov 28 13:21:08 crc kubenswrapper[4631]: I1128 13:21:08.372781 4631 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-j8d66" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a346894-644f-4359-baa5-23bb2d0acc5f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e499b5d5f702b05bb24f616fb7d5ba1799d9b03e06c9efbfbbdcfd86e00e753\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T13:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2blnl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T13:21:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-j8d66\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T13:21:08Z is after 2025-08-24T17:21:41Z" Nov 28 13:21:08 crc kubenswrapper[4631]: I1128 13:21:08.397814 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 13:21:08 crc kubenswrapper[4631]: I1128 13:21:08.397850 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 13:21:08 crc kubenswrapper[4631]: I1128 13:21:08.397860 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 13:21:08 crc kubenswrapper[4631]: I1128 13:21:08.397875 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 13:21:08 crc kubenswrapper[4631]: I1128 13:21:08.397884 4631 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T13:21:08Z","lastTransitionTime":"2025-11-28T13:21:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 13:21:08 crc kubenswrapper[4631]: I1128 13:21:08.413442 4631 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-268tw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a205f475-340d-4ec5-a860-d4a5c2ceb39e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdzgl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdzgl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdzgl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdzgl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdzgl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdzgl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdzgl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdzgl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://542662c37344400bc1d05230e709812104866735429d37ed0d7894050e7a54e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://542662c37344400bc1d05230e709812104866735429d37ed0d7894050e7a54e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T13:21:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T13:21:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdzgl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T13:21:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-268tw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T13:21:08Z is after 2025-08-24T17:21:41Z" Nov 28 13:21:08 crc kubenswrapper[4631]: I1128 13:21:08.500443 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 13:21:08 crc kubenswrapper[4631]: I1128 13:21:08.500507 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 13:21:08 crc kubenswrapper[4631]: I1128 13:21:08.500523 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 13:21:08 crc kubenswrapper[4631]: I1128 13:21:08.500547 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 13:21:08 crc kubenswrapper[4631]: I1128 13:21:08.500562 4631 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T13:21:08Z","lastTransitionTime":"2025-11-28T13:21:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 13:21:08 crc kubenswrapper[4631]: I1128 13:21:08.512905 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 13:21:08 crc kubenswrapper[4631]: I1128 13:21:08.513040 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 13:21:08 crc kubenswrapper[4631]: E1128 13:21:08.513102 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 13:21:08 crc kubenswrapper[4631]: E1128 13:21:08.513260 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 13:21:08 crc kubenswrapper[4631]: I1128 13:21:08.513056 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 13:21:08 crc kubenswrapper[4631]: E1128 13:21:08.513446 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 13:21:08 crc kubenswrapper[4631]: I1128 13:21:08.603945 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 13:21:08 crc kubenswrapper[4631]: I1128 13:21:08.603998 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 13:21:08 crc kubenswrapper[4631]: I1128 13:21:08.604008 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 13:21:08 crc kubenswrapper[4631]: I1128 13:21:08.604028 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 13:21:08 crc kubenswrapper[4631]: I1128 13:21:08.604049 4631 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T13:21:08Z","lastTransitionTime":"2025-11-28T13:21:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 13:21:08 crc kubenswrapper[4631]: I1128 13:21:08.679729 4631 generic.go:334] "Generic (PLEG): container finished" podID="a0fde998-9a55-4734-9008-136b93f29d0c" containerID="87b6625e0c8a70b46f900a5b57e8ef093ba5a192432dfa9708715ec59e8f8981" exitCode=0 Nov 28 13:21:08 crc kubenswrapper[4631]: I1128 13:21:08.679848 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-8gtw6" event={"ID":"a0fde998-9a55-4734-9008-136b93f29d0c","Type":"ContainerDied","Data":"87b6625e0c8a70b46f900a5b57e8ef093ba5a192432dfa9708715ec59e8f8981"} Nov 28 13:21:08 crc kubenswrapper[4631]: I1128 13:21:08.682395 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-cbcjn" event={"ID":"fe4682b5-79b2-456b-b2b8-4fecbe51e5da","Type":"ContainerStarted","Data":"82b212ec45f62923e664878cebfacbd6b7d867235cb1c9bb5b6dd79f8f752c7f"} Nov 28 13:21:08 crc kubenswrapper[4631]: I1128 13:21:08.682429 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-cbcjn" event={"ID":"fe4682b5-79b2-456b-b2b8-4fecbe51e5da","Type":"ContainerStarted","Data":"2454e3bff7706b952aa38d139f4b15ca8b1d1880b0fc7d33a35589a4e9a4f09f"} Nov 28 13:21:08 crc kubenswrapper[4631]: I1128 13:21:08.696690 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-268tw" event={"ID":"a205f475-340d-4ec5-a860-d4a5c2ceb39e","Type":"ContainerStarted","Data":"e5085439ece35186f2692a8cf5be0b011fbf3f212f4f772b674c70649486c89b"} Nov 28 13:21:08 crc kubenswrapper[4631]: I1128 13:21:08.696737 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-268tw" event={"ID":"a205f475-340d-4ec5-a860-d4a5c2ceb39e","Type":"ContainerStarted","Data":"95aeff4871c229569b0c99ff9a6cd17c63f10c1c89d9f04c533dce699c89ce31"} Nov 28 13:21:08 crc kubenswrapper[4631]: I1128 13:21:08.696750 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-268tw" event={"ID":"a205f475-340d-4ec5-a860-d4a5c2ceb39e","Type":"ContainerStarted","Data":"e571a03b3c52184c8fa59e1ab6fff8d9118f114dd6c156638ba36fb7a0d2fe40"} Nov 28 13:21:08 crc kubenswrapper[4631]: I1128 13:21:08.711727 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 13:21:08 crc kubenswrapper[4631]: I1128 13:21:08.711760 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 13:21:08 crc kubenswrapper[4631]: I1128 13:21:08.711770 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 13:21:08 crc kubenswrapper[4631]: I1128 13:21:08.711787 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 13:21:08 crc kubenswrapper[4631]: I1128 13:21:08.711796 4631 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T13:21:08Z","lastTransitionTime":"2025-11-28T13:21:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 13:21:08 crc kubenswrapper[4631]: I1128 13:21:08.713117 4631 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-8gtw6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a0fde998-9a55-4734-9008-136b93f29d0c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:05Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h89l2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af2091e57fbae619a1f0bdbe59e8b35b23ae732d5d4dbe736284fbe566cb0c48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af2091e57fbae619a1f0bdbe59e8b35b23ae732d5d4dbe736284fbe566cb0c48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T13:21:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T13:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h89l2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://87b6625e0c8a70b46f900a5b57e8ef093ba5a192432dfa9708715ec59e8f8981\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://87b6625e0c8a70b46f900a5b57e8ef093ba5a192432dfa9708715ec59e8f8981\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T13:21:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T13:21:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h89l2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h89l2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h89l2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h89l2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h89l2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T13:21:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-8gtw6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T13:21:08Z is after 2025-08-24T17:21:41Z" Nov 28 13:21:08 crc kubenswrapper[4631]: I1128 13:21:08.728516 4631 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fb5d8adf7fbc8de46da5e68668ac8f1a3083ca93ac373f4626c6ba1cb6e07d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T13:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wtbl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f613562449a8f5ebf17cd881967c47d44fb0296171712af40f42263f69f8136\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T13:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wtbl6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T13:21:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-47sxw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T13:21:08Z is after 2025-08-24T17:21:41Z" Nov 28 13:21:08 crc kubenswrapper[4631]: I1128 13:21:08.743805 4631 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c19e346e44c18c1937710aca15059221d7f4a46252fa90098870f100dd4157da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T13:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T13:21:08Z is after 2025-08-24T17:21:41Z" Nov 28 13:21:08 crc kubenswrapper[4631]: I1128 13:21:08.761564 4631 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:02Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T13:21:08Z is after 2025-08-24T17:21:41Z" Nov 28 13:21:08 crc kubenswrapper[4631]: I1128 13:21:08.780621 4631 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T13:21:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8235c1ea6445a3e88ee3e0f57ce76e9840dfa95421832d59a3bc2e19952850db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T13:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://754e658d99a10b103b24fc0244ad606402d50d072cf3c2bb93518afa4ab66935\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T13:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T13:21:08Z is after 2025-08-24T17:21:41Z" Nov 28 13:21:08 crc kubenswrapper[4631]: I1128 13:21:08.815472 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 13:21:08 crc kubenswrapper[4631]: I1128 13:21:08.815503 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 13:21:08 crc kubenswrapper[4631]: I1128 13:21:08.815512 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 13:21:08 crc kubenswrapper[4631]: I1128 13:21:08.815527 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 13:21:08 crc kubenswrapper[4631]: I1128 13:21:08.815536 4631 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T13:21:08Z","lastTransitionTime":"2025-11-28T13:21:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 13:21:08 crc kubenswrapper[4631]: I1128 13:21:08.828838 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-wbkjm" podStartSLOduration=6.82881881 podStartE2EDuration="6.82881881s" podCreationTimestamp="2025-11-28 13:21:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 13:21:08.800908532 +0000 UTC m=+25.608211876" watchObservedRunningTime="2025-11-28 13:21:08.82881881 +0000 UTC m=+25.636122154" Nov 28 13:21:08 crc kubenswrapper[4631]: I1128 13:21:08.842676 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=6.842657647 podStartE2EDuration="6.842657647s" podCreationTimestamp="2025-11-28 13:21:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 13:21:08.829086275 +0000 UTC m=+25.636389619" watchObservedRunningTime="2025-11-28 13:21:08.842657647 +0000 UTC m=+25.649960991" Nov 28 13:21:08 crc kubenswrapper[4631]: I1128 13:21:08.883063 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=6.883046083 podStartE2EDuration="6.883046083s" podCreationTimestamp="2025-11-28 13:21:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 13:21:08.866556512 +0000 UTC m=+25.673859856" watchObservedRunningTime="2025-11-28 13:21:08.883046083 +0000 UTC m=+25.690349427" Nov 28 13:21:08 crc kubenswrapper[4631]: I1128 13:21:08.917805 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 13:21:08 crc kubenswrapper[4631]: I1128 13:21:08.917834 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 13:21:08 crc kubenswrapper[4631]: I1128 13:21:08.917842 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 13:21:08 crc kubenswrapper[4631]: I1128 13:21:08.917855 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 13:21:08 crc kubenswrapper[4631]: I1128 13:21:08.917864 4631 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T13:21:08Z","lastTransitionTime":"2025-11-28T13:21:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 13:21:08 crc kubenswrapper[4631]: I1128 13:21:08.978717 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-j8d66" podStartSLOduration=5.978696885 podStartE2EDuration="5.978696885s" podCreationTimestamp="2025-11-28 13:21:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 13:21:08.941049174 +0000 UTC m=+25.748352518" watchObservedRunningTime="2025-11-28 13:21:08.978696885 +0000 UTC m=+25.786000219" Nov 28 13:21:09 crc kubenswrapper[4631]: I1128 13:21:09.002560 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-cbcjn" podStartSLOduration=6.002539818 podStartE2EDuration="6.002539818s" podCreationTimestamp="2025-11-28 13:21:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 13:21:09.001889956 +0000 UTC m=+25.809193290" watchObservedRunningTime="2025-11-28 13:21:09.002539818 +0000 UTC m=+25.809843162" Nov 28 13:21:09 crc kubenswrapper[4631]: I1128 13:21:09.020272 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 13:21:09 crc kubenswrapper[4631]: I1128 13:21:09.020350 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 13:21:09 crc kubenswrapper[4631]: I1128 13:21:09.020359 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 13:21:09 crc kubenswrapper[4631]: I1128 13:21:09.020378 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 13:21:09 crc kubenswrapper[4631]: I1128 13:21:09.020388 4631 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T13:21:09Z","lastTransitionTime":"2025-11-28T13:21:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 13:21:09 crc kubenswrapper[4631]: I1128 13:21:09.097205 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" podStartSLOduration=6.0971816 podStartE2EDuration="6.0971816s" podCreationTimestamp="2025-11-28 13:21:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 13:21:09.09623378 +0000 UTC m=+25.903537124" watchObservedRunningTime="2025-11-28 13:21:09.0971816 +0000 UTC m=+25.904484944" Nov 28 13:21:09 crc kubenswrapper[4631]: I1128 13:21:09.122627 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 13:21:09 crc kubenswrapper[4631]: I1128 13:21:09.122693 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 13:21:09 crc kubenswrapper[4631]: I1128 13:21:09.122714 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 13:21:09 crc kubenswrapper[4631]: I1128 13:21:09.122733 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 13:21:09 crc kubenswrapper[4631]: I1128 13:21:09.122749 4631 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T13:21:09Z","lastTransitionTime":"2025-11-28T13:21:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 13:21:09 crc kubenswrapper[4631]: I1128 13:21:09.203246 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jt7v2"] Nov 28 13:21:09 crc kubenswrapper[4631]: I1128 13:21:09.203670 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jt7v2" Nov 28 13:21:09 crc kubenswrapper[4631]: I1128 13:21:09.216425 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Nov 28 13:21:09 crc kubenswrapper[4631]: I1128 13:21:09.225247 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 13:21:09 crc kubenswrapper[4631]: I1128 13:21:09.225281 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 13:21:09 crc kubenswrapper[4631]: I1128 13:21:09.225325 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 13:21:09 crc kubenswrapper[4631]: I1128 13:21:09.225341 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 13:21:09 crc kubenswrapper[4631]: I1128 13:21:09.225351 4631 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T13:21:09Z","lastTransitionTime":"2025-11-28T13:21:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 13:21:09 crc kubenswrapper[4631]: I1128 13:21:09.234434 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-dd58r"] Nov 28 13:21:09 crc kubenswrapper[4631]: I1128 13:21:09.234900 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-dd58r" Nov 28 13:21:09 crc kubenswrapper[4631]: E1128 13:21:09.234965 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-dd58r" podUID="8716de2d-f33e-4434-9076-6345669aaff8" Nov 28 13:21:09 crc kubenswrapper[4631]: I1128 13:21:09.236374 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Nov 28 13:21:09 crc kubenswrapper[4631]: I1128 13:21:09.284613 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/20918daf-8eb0-49ff-bf19-f765c0b3eb95-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-jt7v2\" (UID: \"20918daf-8eb0-49ff-bf19-f765c0b3eb95\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jt7v2" Nov 28 13:21:09 crc kubenswrapper[4631]: I1128 13:21:09.284849 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/20918daf-8eb0-49ff-bf19-f765c0b3eb95-env-overrides\") pod \"ovnkube-control-plane-749d76644c-jt7v2\" (UID: \"20918daf-8eb0-49ff-bf19-f765c0b3eb95\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jt7v2" Nov 28 13:21:09 crc kubenswrapper[4631]: I1128 13:21:09.284975 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bcbwf\" (UniqueName: \"kubernetes.io/projected/20918daf-8eb0-49ff-bf19-f765c0b3eb95-kube-api-access-bcbwf\") pod \"ovnkube-control-plane-749d76644c-jt7v2\" (UID: \"20918daf-8eb0-49ff-bf19-f765c0b3eb95\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jt7v2" Nov 28 13:21:09 crc kubenswrapper[4631]: I1128 13:21:09.285057 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/20918daf-8eb0-49ff-bf19-f765c0b3eb95-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-jt7v2\" (UID: \"20918daf-8eb0-49ff-bf19-f765c0b3eb95\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jt7v2" Nov 28 13:21:09 crc kubenswrapper[4631]: I1128 13:21:09.327706 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 13:21:09 crc kubenswrapper[4631]: I1128 13:21:09.327739 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 13:21:09 crc kubenswrapper[4631]: I1128 13:21:09.327748 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 13:21:09 crc kubenswrapper[4631]: I1128 13:21:09.327763 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 13:21:09 crc kubenswrapper[4631]: I1128 13:21:09.327774 4631 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T13:21:09Z","lastTransitionTime":"2025-11-28T13:21:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 13:21:09 crc kubenswrapper[4631]: I1128 13:21:09.385862 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/20918daf-8eb0-49ff-bf19-f765c0b3eb95-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-jt7v2\" (UID: \"20918daf-8eb0-49ff-bf19-f765c0b3eb95\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jt7v2" Nov 28 13:21:09 crc kubenswrapper[4631]: I1128 13:21:09.385911 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/20918daf-8eb0-49ff-bf19-f765c0b3eb95-env-overrides\") pod \"ovnkube-control-plane-749d76644c-jt7v2\" (UID: \"20918daf-8eb0-49ff-bf19-f765c0b3eb95\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jt7v2" Nov 28 13:21:09 crc kubenswrapper[4631]: I1128 13:21:09.385929 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bcbwf\" (UniqueName: \"kubernetes.io/projected/20918daf-8eb0-49ff-bf19-f765c0b3eb95-kube-api-access-bcbwf\") pod \"ovnkube-control-plane-749d76644c-jt7v2\" (UID: \"20918daf-8eb0-49ff-bf19-f765c0b3eb95\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jt7v2" Nov 28 13:21:09 crc kubenswrapper[4631]: I1128 13:21:09.385957 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2gdsf\" (UniqueName: \"kubernetes.io/projected/8716de2d-f33e-4434-9076-6345669aaff8-kube-api-access-2gdsf\") pod \"network-metrics-daemon-dd58r\" (UID: \"8716de2d-f33e-4434-9076-6345669aaff8\") " pod="openshift-multus/network-metrics-daemon-dd58r" Nov 28 13:21:09 crc kubenswrapper[4631]: I1128 13:21:09.385980 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/20918daf-8eb0-49ff-bf19-f765c0b3eb95-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-jt7v2\" (UID: \"20918daf-8eb0-49ff-bf19-f765c0b3eb95\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jt7v2" Nov 28 13:21:09 crc kubenswrapper[4631]: I1128 13:21:09.386012 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8716de2d-f33e-4434-9076-6345669aaff8-metrics-certs\") pod \"network-metrics-daemon-dd58r\" (UID: \"8716de2d-f33e-4434-9076-6345669aaff8\") " pod="openshift-multus/network-metrics-daemon-dd58r" Nov 28 13:21:09 crc kubenswrapper[4631]: I1128 13:21:09.386878 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/20918daf-8eb0-49ff-bf19-f765c0b3eb95-env-overrides\") pod \"ovnkube-control-plane-749d76644c-jt7v2\" (UID: \"20918daf-8eb0-49ff-bf19-f765c0b3eb95\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jt7v2" Nov 28 13:21:09 crc kubenswrapper[4631]: I1128 13:21:09.386989 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/20918daf-8eb0-49ff-bf19-f765c0b3eb95-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-jt7v2\" (UID: \"20918daf-8eb0-49ff-bf19-f765c0b3eb95\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jt7v2" Nov 28 13:21:09 crc kubenswrapper[4631]: I1128 13:21:09.395431 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/20918daf-8eb0-49ff-bf19-f765c0b3eb95-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-jt7v2\" (UID: \"20918daf-8eb0-49ff-bf19-f765c0b3eb95\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jt7v2" Nov 28 13:21:09 crc kubenswrapper[4631]: I1128 13:21:09.403626 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bcbwf\" (UniqueName: \"kubernetes.io/projected/20918daf-8eb0-49ff-bf19-f765c0b3eb95-kube-api-access-bcbwf\") pod \"ovnkube-control-plane-749d76644c-jt7v2\" (UID: \"20918daf-8eb0-49ff-bf19-f765c0b3eb95\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jt7v2" Nov 28 13:21:09 crc kubenswrapper[4631]: I1128 13:21:09.430852 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 13:21:09 crc kubenswrapper[4631]: I1128 13:21:09.430895 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 13:21:09 crc kubenswrapper[4631]: I1128 13:21:09.430905 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 13:21:09 crc kubenswrapper[4631]: I1128 13:21:09.430921 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 13:21:09 crc kubenswrapper[4631]: I1128 13:21:09.430930 4631 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T13:21:09Z","lastTransitionTime":"2025-11-28T13:21:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 13:21:09 crc kubenswrapper[4631]: I1128 13:21:09.486722 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8716de2d-f33e-4434-9076-6345669aaff8-metrics-certs\") pod \"network-metrics-daemon-dd58r\" (UID: \"8716de2d-f33e-4434-9076-6345669aaff8\") " pod="openshift-multus/network-metrics-daemon-dd58r" Nov 28 13:21:09 crc kubenswrapper[4631]: I1128 13:21:09.486784 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2gdsf\" (UniqueName: \"kubernetes.io/projected/8716de2d-f33e-4434-9076-6345669aaff8-kube-api-access-2gdsf\") pod \"network-metrics-daemon-dd58r\" (UID: \"8716de2d-f33e-4434-9076-6345669aaff8\") " pod="openshift-multus/network-metrics-daemon-dd58r" Nov 28 13:21:09 crc kubenswrapper[4631]: E1128 13:21:09.486943 4631 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 28 13:21:09 crc kubenswrapper[4631]: E1128 13:21:09.487046 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8716de2d-f33e-4434-9076-6345669aaff8-metrics-certs podName:8716de2d-f33e-4434-9076-6345669aaff8 nodeName:}" failed. No retries permitted until 2025-11-28 13:21:09.987023615 +0000 UTC m=+26.794327049 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/8716de2d-f33e-4434-9076-6345669aaff8-metrics-certs") pod "network-metrics-daemon-dd58r" (UID: "8716de2d-f33e-4434-9076-6345669aaff8") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 28 13:21:09 crc kubenswrapper[4631]: I1128 13:21:09.510935 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2gdsf\" (UniqueName: \"kubernetes.io/projected/8716de2d-f33e-4434-9076-6345669aaff8-kube-api-access-2gdsf\") pod \"network-metrics-daemon-dd58r\" (UID: \"8716de2d-f33e-4434-9076-6345669aaff8\") " pod="openshift-multus/network-metrics-daemon-dd58r" Nov 28 13:21:09 crc kubenswrapper[4631]: I1128 13:21:09.515709 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jt7v2" Nov 28 13:21:09 crc kubenswrapper[4631]: I1128 13:21:09.538788 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 13:21:09 crc kubenswrapper[4631]: I1128 13:21:09.538824 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 13:21:09 crc kubenswrapper[4631]: I1128 13:21:09.538833 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 13:21:09 crc kubenswrapper[4631]: I1128 13:21:09.538847 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 13:21:09 crc kubenswrapper[4631]: I1128 13:21:09.538856 4631 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T13:21:09Z","lastTransitionTime":"2025-11-28T13:21:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 13:21:09 crc kubenswrapper[4631]: I1128 13:21:09.642713 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 13:21:09 crc kubenswrapper[4631]: I1128 13:21:09.642809 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 13:21:09 crc kubenswrapper[4631]: I1128 13:21:09.642820 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 13:21:09 crc kubenswrapper[4631]: I1128 13:21:09.642834 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 13:21:09 crc kubenswrapper[4631]: I1128 13:21:09.642844 4631 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T13:21:09Z","lastTransitionTime":"2025-11-28T13:21:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 13:21:09 crc kubenswrapper[4631]: I1128 13:21:09.702365 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-268tw" event={"ID":"a205f475-340d-4ec5-a860-d4a5c2ceb39e","Type":"ContainerStarted","Data":"0ab2d69688bf305aeb7ece5c432d87adcb06eaca6c90ef5aa4e73b7fe477246b"} Nov 28 13:21:09 crc kubenswrapper[4631]: I1128 13:21:09.702414 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-268tw" event={"ID":"a205f475-340d-4ec5-a860-d4a5c2ceb39e","Type":"ContainerStarted","Data":"794ab47dc4660d76accb0456995b602cd5cdfc08616f63427947aba9e725eaf7"} Nov 28 13:21:09 crc kubenswrapper[4631]: I1128 13:21:09.702425 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-268tw" event={"ID":"a205f475-340d-4ec5-a860-d4a5c2ceb39e","Type":"ContainerStarted","Data":"016163e216d642018075fa0e9ab61fa630fa9718030246855fd44a8104c3d6e6"} Nov 28 13:21:09 crc kubenswrapper[4631]: I1128 13:21:09.702962 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jt7v2" event={"ID":"20918daf-8eb0-49ff-bf19-f765c0b3eb95","Type":"ContainerStarted","Data":"4b0a89ce3f75ac5e27b048be39d64f3b218f30e692fbbe4b4b85b22ff575f15b"} Nov 28 13:21:09 crc kubenswrapper[4631]: I1128 13:21:09.704364 4631 generic.go:334] "Generic (PLEG): container finished" podID="a0fde998-9a55-4734-9008-136b93f29d0c" containerID="a09285621158d2e147db6f7f2e2d25d281e59bc3c9881a0fccbadabf48a499b0" exitCode=0 Nov 28 13:21:09 crc kubenswrapper[4631]: I1128 13:21:09.704397 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-8gtw6" event={"ID":"a0fde998-9a55-4734-9008-136b93f29d0c","Type":"ContainerDied","Data":"a09285621158d2e147db6f7f2e2d25d281e59bc3c9881a0fccbadabf48a499b0"} Nov 28 13:21:09 crc kubenswrapper[4631]: I1128 13:21:09.745676 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 13:21:09 crc kubenswrapper[4631]: I1128 13:21:09.745726 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 13:21:09 crc kubenswrapper[4631]: I1128 13:21:09.745736 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 13:21:09 crc kubenswrapper[4631]: I1128 13:21:09.745750 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 13:21:09 crc kubenswrapper[4631]: I1128 13:21:09.745760 4631 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T13:21:09Z","lastTransitionTime":"2025-11-28T13:21:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 13:21:09 crc kubenswrapper[4631]: I1128 13:21:09.850178 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 13:21:09 crc kubenswrapper[4631]: I1128 13:21:09.850210 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 13:21:09 crc kubenswrapper[4631]: I1128 13:21:09.850220 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 13:21:09 crc kubenswrapper[4631]: I1128 13:21:09.850252 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 13:21:09 crc kubenswrapper[4631]: I1128 13:21:09.850261 4631 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T13:21:09Z","lastTransitionTime":"2025-11-28T13:21:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 13:21:09 crc kubenswrapper[4631]: I1128 13:21:09.952822 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 13:21:09 crc kubenswrapper[4631]: I1128 13:21:09.952859 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 13:21:09 crc kubenswrapper[4631]: I1128 13:21:09.952867 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 13:21:09 crc kubenswrapper[4631]: I1128 13:21:09.952882 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 13:21:09 crc kubenswrapper[4631]: I1128 13:21:09.952890 4631 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T13:21:09Z","lastTransitionTime":"2025-11-28T13:21:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 13:21:09 crc kubenswrapper[4631]: I1128 13:21:09.991653 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8716de2d-f33e-4434-9076-6345669aaff8-metrics-certs\") pod \"network-metrics-daemon-dd58r\" (UID: \"8716de2d-f33e-4434-9076-6345669aaff8\") " pod="openshift-multus/network-metrics-daemon-dd58r" Nov 28 13:21:09 crc kubenswrapper[4631]: E1128 13:21:09.991808 4631 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 28 13:21:09 crc kubenswrapper[4631]: E1128 13:21:09.991922 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8716de2d-f33e-4434-9076-6345669aaff8-metrics-certs podName:8716de2d-f33e-4434-9076-6345669aaff8 nodeName:}" failed. No retries permitted until 2025-11-28 13:21:10.991900834 +0000 UTC m=+27.799204218 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/8716de2d-f33e-4434-9076-6345669aaff8-metrics-certs") pod "network-metrics-daemon-dd58r" (UID: "8716de2d-f33e-4434-9076-6345669aaff8") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 28 13:21:10 crc kubenswrapper[4631]: I1128 13:21:10.055016 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 13:21:10 crc kubenswrapper[4631]: I1128 13:21:10.055055 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 13:21:10 crc kubenswrapper[4631]: I1128 13:21:10.055065 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 13:21:10 crc kubenswrapper[4631]: I1128 13:21:10.055081 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 13:21:10 crc kubenswrapper[4631]: I1128 13:21:10.055094 4631 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T13:21:10Z","lastTransitionTime":"2025-11-28T13:21:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 13:21:10 crc kubenswrapper[4631]: I1128 13:21:10.157547 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 13:21:10 crc kubenswrapper[4631]: I1128 13:21:10.157574 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 13:21:10 crc kubenswrapper[4631]: I1128 13:21:10.157582 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 13:21:10 crc kubenswrapper[4631]: I1128 13:21:10.157594 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 13:21:10 crc kubenswrapper[4631]: I1128 13:21:10.157603 4631 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T13:21:10Z","lastTransitionTime":"2025-11-28T13:21:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 13:21:10 crc kubenswrapper[4631]: I1128 13:21:10.259964 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 13:21:10 crc kubenswrapper[4631]: I1128 13:21:10.260015 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 13:21:10 crc kubenswrapper[4631]: I1128 13:21:10.260029 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 13:21:10 crc kubenswrapper[4631]: I1128 13:21:10.260047 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 13:21:10 crc kubenswrapper[4631]: I1128 13:21:10.260058 4631 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T13:21:10Z","lastTransitionTime":"2025-11-28T13:21:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 13:21:10 crc kubenswrapper[4631]: I1128 13:21:10.294460 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 13:21:10 crc kubenswrapper[4631]: I1128 13:21:10.294544 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 13:21:10 crc kubenswrapper[4631]: E1128 13:21:10.294618 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 13:21:18.294562834 +0000 UTC m=+35.101866168 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 13:21:10 crc kubenswrapper[4631]: I1128 13:21:10.294660 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 13:21:10 crc kubenswrapper[4631]: I1128 13:21:10.294716 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 13:21:10 crc kubenswrapper[4631]: I1128 13:21:10.294740 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 13:21:10 crc kubenswrapper[4631]: E1128 13:21:10.294881 4631 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 28 13:21:10 crc kubenswrapper[4631]: E1128 13:21:10.294895 4631 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 28 13:21:10 crc kubenswrapper[4631]: E1128 13:21:10.294907 4631 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 28 13:21:10 crc kubenswrapper[4631]: E1128 13:21:10.294945 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-28 13:21:18.294931101 +0000 UTC m=+35.102234445 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 28 13:21:10 crc kubenswrapper[4631]: E1128 13:21:10.294621 4631 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 28 13:21:10 crc kubenswrapper[4631]: E1128 13:21:10.295222 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-28 13:21:18.295215538 +0000 UTC m=+35.102518882 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 28 13:21:10 crc kubenswrapper[4631]: E1128 13:21:10.295268 4631 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 28 13:21:10 crc kubenswrapper[4631]: E1128 13:21:10.295308 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-28 13:21:18.29530014 +0000 UTC m=+35.102603484 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 28 13:21:10 crc kubenswrapper[4631]: E1128 13:21:10.295352 4631 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 28 13:21:10 crc kubenswrapper[4631]: E1128 13:21:10.295363 4631 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 28 13:21:10 crc kubenswrapper[4631]: E1128 13:21:10.295371 4631 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 28 13:21:10 crc kubenswrapper[4631]: E1128 13:21:10.295391 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-28 13:21:18.295385111 +0000 UTC m=+35.102688455 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 28 13:21:10 crc kubenswrapper[4631]: I1128 13:21:10.362642 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 13:21:10 crc kubenswrapper[4631]: I1128 13:21:10.362678 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 13:21:10 crc kubenswrapper[4631]: I1128 13:21:10.362686 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 13:21:10 crc kubenswrapper[4631]: I1128 13:21:10.362701 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 13:21:10 crc kubenswrapper[4631]: I1128 13:21:10.362711 4631 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T13:21:10Z","lastTransitionTime":"2025-11-28T13:21:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 13:21:10 crc kubenswrapper[4631]: I1128 13:21:10.464417 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 13:21:10 crc kubenswrapper[4631]: I1128 13:21:10.464452 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 13:21:10 crc kubenswrapper[4631]: I1128 13:21:10.464460 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 13:21:10 crc kubenswrapper[4631]: I1128 13:21:10.464473 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 13:21:10 crc kubenswrapper[4631]: I1128 13:21:10.464481 4631 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T13:21:10Z","lastTransitionTime":"2025-11-28T13:21:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 13:21:10 crc kubenswrapper[4631]: I1128 13:21:10.513020 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-dd58r" Nov 28 13:21:10 crc kubenswrapper[4631]: I1128 13:21:10.513049 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 13:21:10 crc kubenswrapper[4631]: E1128 13:21:10.513162 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-dd58r" podUID="8716de2d-f33e-4434-9076-6345669aaff8" Nov 28 13:21:10 crc kubenswrapper[4631]: I1128 13:21:10.513181 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 13:21:10 crc kubenswrapper[4631]: I1128 13:21:10.513215 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 13:21:10 crc kubenswrapper[4631]: E1128 13:21:10.513387 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 13:21:10 crc kubenswrapper[4631]: E1128 13:21:10.513529 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 13:21:10 crc kubenswrapper[4631]: E1128 13:21:10.513626 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 13:21:10 crc kubenswrapper[4631]: I1128 13:21:10.566647 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 13:21:10 crc kubenswrapper[4631]: I1128 13:21:10.566681 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 13:21:10 crc kubenswrapper[4631]: I1128 13:21:10.566689 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 13:21:10 crc kubenswrapper[4631]: I1128 13:21:10.566703 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 13:21:10 crc kubenswrapper[4631]: I1128 13:21:10.566712 4631 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T13:21:10Z","lastTransitionTime":"2025-11-28T13:21:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 13:21:10 crc kubenswrapper[4631]: I1128 13:21:10.668665 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 13:21:10 crc kubenswrapper[4631]: I1128 13:21:10.668695 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 13:21:10 crc kubenswrapper[4631]: I1128 13:21:10.668703 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 13:21:10 crc kubenswrapper[4631]: I1128 13:21:10.668716 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 13:21:10 crc kubenswrapper[4631]: I1128 13:21:10.668724 4631 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T13:21:10Z","lastTransitionTime":"2025-11-28T13:21:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 13:21:10 crc kubenswrapper[4631]: I1128 13:21:10.709475 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jt7v2" event={"ID":"20918daf-8eb0-49ff-bf19-f765c0b3eb95","Type":"ContainerStarted","Data":"7d25a3a7d331733498e0f4f45f05cda67814421a2b293bc1734758ce7241f2e6"} Nov 28 13:21:10 crc kubenswrapper[4631]: I1128 13:21:10.709539 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jt7v2" event={"ID":"20918daf-8eb0-49ff-bf19-f765c0b3eb95","Type":"ContainerStarted","Data":"461ce6490e520c9df2c6ec8cbfe1d4d3ce5ae397cb705a4584b72e9eccf15b54"} Nov 28 13:21:10 crc kubenswrapper[4631]: I1128 13:21:10.712435 4631 generic.go:334] "Generic (PLEG): container finished" podID="a0fde998-9a55-4734-9008-136b93f29d0c" containerID="77d13aa80535741eb84f46055ad3b6d9b022dc44727eeaa72dc3dfaab48d1d85" exitCode=0 Nov 28 13:21:10 crc kubenswrapper[4631]: I1128 13:21:10.712477 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-8gtw6" event={"ID":"a0fde998-9a55-4734-9008-136b93f29d0c","Type":"ContainerDied","Data":"77d13aa80535741eb84f46055ad3b6d9b022dc44727eeaa72dc3dfaab48d1d85"} Nov 28 13:21:10 crc kubenswrapper[4631]: I1128 13:21:10.723936 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jt7v2" podStartSLOduration=6.723917819 podStartE2EDuration="6.723917819s" podCreationTimestamp="2025-11-28 13:21:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 13:21:10.723570301 +0000 UTC m=+27.530873645" watchObservedRunningTime="2025-11-28 13:21:10.723917819 +0000 UTC m=+27.531221163" Nov 28 13:21:10 crc kubenswrapper[4631]: I1128 13:21:10.771070 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 13:21:10 crc kubenswrapper[4631]: I1128 13:21:10.771097 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 13:21:10 crc kubenswrapper[4631]: I1128 13:21:10.771104 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 13:21:10 crc kubenswrapper[4631]: I1128 13:21:10.771116 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 13:21:10 crc kubenswrapper[4631]: I1128 13:21:10.771126 4631 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T13:21:10Z","lastTransitionTime":"2025-11-28T13:21:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 13:21:10 crc kubenswrapper[4631]: I1128 13:21:10.875022 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 13:21:10 crc kubenswrapper[4631]: I1128 13:21:10.875064 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 13:21:10 crc kubenswrapper[4631]: I1128 13:21:10.875075 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 13:21:10 crc kubenswrapper[4631]: I1128 13:21:10.875090 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 13:21:10 crc kubenswrapper[4631]: I1128 13:21:10.875102 4631 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T13:21:10Z","lastTransitionTime":"2025-11-28T13:21:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 13:21:10 crc kubenswrapper[4631]: I1128 13:21:10.977008 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 13:21:10 crc kubenswrapper[4631]: I1128 13:21:10.977035 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 13:21:10 crc kubenswrapper[4631]: I1128 13:21:10.977044 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 13:21:10 crc kubenswrapper[4631]: I1128 13:21:10.977057 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 13:21:10 crc kubenswrapper[4631]: I1128 13:21:10.977065 4631 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T13:21:10Z","lastTransitionTime":"2025-11-28T13:21:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 13:21:11 crc kubenswrapper[4631]: I1128 13:21:11.000886 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8716de2d-f33e-4434-9076-6345669aaff8-metrics-certs\") pod \"network-metrics-daemon-dd58r\" (UID: \"8716de2d-f33e-4434-9076-6345669aaff8\") " pod="openshift-multus/network-metrics-daemon-dd58r" Nov 28 13:21:11 crc kubenswrapper[4631]: E1128 13:21:11.001006 4631 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 28 13:21:11 crc kubenswrapper[4631]: E1128 13:21:11.001044 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8716de2d-f33e-4434-9076-6345669aaff8-metrics-certs podName:8716de2d-f33e-4434-9076-6345669aaff8 nodeName:}" failed. No retries permitted until 2025-11-28 13:21:13.00103125 +0000 UTC m=+29.808334594 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/8716de2d-f33e-4434-9076-6345669aaff8-metrics-certs") pod "network-metrics-daemon-dd58r" (UID: "8716de2d-f33e-4434-9076-6345669aaff8") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 28 13:21:11 crc kubenswrapper[4631]: I1128 13:21:11.079238 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 13:21:11 crc kubenswrapper[4631]: I1128 13:21:11.079497 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 13:21:11 crc kubenswrapper[4631]: I1128 13:21:11.079513 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 13:21:11 crc kubenswrapper[4631]: I1128 13:21:11.079528 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 13:21:11 crc kubenswrapper[4631]: I1128 13:21:11.079539 4631 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T13:21:11Z","lastTransitionTime":"2025-11-28T13:21:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 13:21:11 crc kubenswrapper[4631]: I1128 13:21:11.181576 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 13:21:11 crc kubenswrapper[4631]: I1128 13:21:11.181808 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 13:21:11 crc kubenswrapper[4631]: I1128 13:21:11.181877 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 13:21:11 crc kubenswrapper[4631]: I1128 13:21:11.181966 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 13:21:11 crc kubenswrapper[4631]: I1128 13:21:11.182142 4631 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T13:21:11Z","lastTransitionTime":"2025-11-28T13:21:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 13:21:11 crc kubenswrapper[4631]: I1128 13:21:11.284624 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 13:21:11 crc kubenswrapper[4631]: I1128 13:21:11.284867 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 13:21:11 crc kubenswrapper[4631]: I1128 13:21:11.284945 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 13:21:11 crc kubenswrapper[4631]: I1128 13:21:11.285015 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 13:21:11 crc kubenswrapper[4631]: I1128 13:21:11.285074 4631 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T13:21:11Z","lastTransitionTime":"2025-11-28T13:21:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 13:21:11 crc kubenswrapper[4631]: I1128 13:21:11.387824 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 13:21:11 crc kubenswrapper[4631]: I1128 13:21:11.387856 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 13:21:11 crc kubenswrapper[4631]: I1128 13:21:11.387867 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 13:21:11 crc kubenswrapper[4631]: I1128 13:21:11.387882 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 13:21:11 crc kubenswrapper[4631]: I1128 13:21:11.387907 4631 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T13:21:11Z","lastTransitionTime":"2025-11-28T13:21:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 13:21:11 crc kubenswrapper[4631]: I1128 13:21:11.490213 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 13:21:11 crc kubenswrapper[4631]: I1128 13:21:11.490262 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 13:21:11 crc kubenswrapper[4631]: I1128 13:21:11.490276 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 13:21:11 crc kubenswrapper[4631]: I1128 13:21:11.490319 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 13:21:11 crc kubenswrapper[4631]: I1128 13:21:11.490333 4631 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T13:21:11Z","lastTransitionTime":"2025-11-28T13:21:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 13:21:11 crc kubenswrapper[4631]: I1128 13:21:11.592591 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 13:21:11 crc kubenswrapper[4631]: I1128 13:21:11.592646 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 13:21:11 crc kubenswrapper[4631]: I1128 13:21:11.592663 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 13:21:11 crc kubenswrapper[4631]: I1128 13:21:11.592688 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 13:21:11 crc kubenswrapper[4631]: I1128 13:21:11.592706 4631 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T13:21:11Z","lastTransitionTime":"2025-11-28T13:21:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 13:21:11 crc kubenswrapper[4631]: I1128 13:21:11.695228 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 13:21:11 crc kubenswrapper[4631]: I1128 13:21:11.695566 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 13:21:11 crc kubenswrapper[4631]: I1128 13:21:11.695579 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 13:21:11 crc kubenswrapper[4631]: I1128 13:21:11.695597 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 13:21:11 crc kubenswrapper[4631]: I1128 13:21:11.695612 4631 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T13:21:11Z","lastTransitionTime":"2025-11-28T13:21:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 13:21:11 crc kubenswrapper[4631]: I1128 13:21:11.718573 4631 generic.go:334] "Generic (PLEG): container finished" podID="a0fde998-9a55-4734-9008-136b93f29d0c" containerID="8af2894df3a76c3c0d8498863353f010952ff70334510994ecca4332b38a73b7" exitCode=0 Nov 28 13:21:11 crc kubenswrapper[4631]: I1128 13:21:11.719016 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-8gtw6" event={"ID":"a0fde998-9a55-4734-9008-136b93f29d0c","Type":"ContainerDied","Data":"8af2894df3a76c3c0d8498863353f010952ff70334510994ecca4332b38a73b7"} Nov 28 13:21:11 crc kubenswrapper[4631]: I1128 13:21:11.798422 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 13:21:11 crc kubenswrapper[4631]: I1128 13:21:11.798457 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 13:21:11 crc kubenswrapper[4631]: I1128 13:21:11.798468 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 13:21:11 crc kubenswrapper[4631]: I1128 13:21:11.798486 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 13:21:11 crc kubenswrapper[4631]: I1128 13:21:11.798498 4631 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T13:21:11Z","lastTransitionTime":"2025-11-28T13:21:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 13:21:11 crc kubenswrapper[4631]: I1128 13:21:11.900780 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 13:21:11 crc kubenswrapper[4631]: I1128 13:21:11.900816 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 13:21:11 crc kubenswrapper[4631]: I1128 13:21:11.900827 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 13:21:11 crc kubenswrapper[4631]: I1128 13:21:11.900846 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 13:21:11 crc kubenswrapper[4631]: I1128 13:21:11.900857 4631 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T13:21:11Z","lastTransitionTime":"2025-11-28T13:21:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 13:21:12 crc kubenswrapper[4631]: I1128 13:21:12.005047 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 13:21:12 crc kubenswrapper[4631]: I1128 13:21:12.005110 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 13:21:12 crc kubenswrapper[4631]: I1128 13:21:12.005128 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 13:21:12 crc kubenswrapper[4631]: I1128 13:21:12.005156 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 13:21:12 crc kubenswrapper[4631]: I1128 13:21:12.005180 4631 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T13:21:12Z","lastTransitionTime":"2025-11-28T13:21:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 13:21:12 crc kubenswrapper[4631]: I1128 13:21:12.108103 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 13:21:12 crc kubenswrapper[4631]: I1128 13:21:12.108148 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 13:21:12 crc kubenswrapper[4631]: I1128 13:21:12.108160 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 13:21:12 crc kubenswrapper[4631]: I1128 13:21:12.108179 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 13:21:12 crc kubenswrapper[4631]: I1128 13:21:12.108191 4631 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T13:21:12Z","lastTransitionTime":"2025-11-28T13:21:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 13:21:12 crc kubenswrapper[4631]: I1128 13:21:12.210829 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 13:21:12 crc kubenswrapper[4631]: I1128 13:21:12.210870 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 13:21:12 crc kubenswrapper[4631]: I1128 13:21:12.210881 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 13:21:12 crc kubenswrapper[4631]: I1128 13:21:12.210896 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 13:21:12 crc kubenswrapper[4631]: I1128 13:21:12.210905 4631 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T13:21:12Z","lastTransitionTime":"2025-11-28T13:21:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 13:21:12 crc kubenswrapper[4631]: I1128 13:21:12.313810 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 13:21:12 crc kubenswrapper[4631]: I1128 13:21:12.313850 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 13:21:12 crc kubenswrapper[4631]: I1128 13:21:12.313861 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 13:21:12 crc kubenswrapper[4631]: I1128 13:21:12.313875 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 13:21:12 crc kubenswrapper[4631]: I1128 13:21:12.313886 4631 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T13:21:12Z","lastTransitionTime":"2025-11-28T13:21:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 13:21:12 crc kubenswrapper[4631]: I1128 13:21:12.417147 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 13:21:12 crc kubenswrapper[4631]: I1128 13:21:12.417210 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 13:21:12 crc kubenswrapper[4631]: I1128 13:21:12.417225 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 13:21:12 crc kubenswrapper[4631]: I1128 13:21:12.417247 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 13:21:12 crc kubenswrapper[4631]: I1128 13:21:12.417261 4631 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T13:21:12Z","lastTransitionTime":"2025-11-28T13:21:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 13:21:12 crc kubenswrapper[4631]: I1128 13:21:12.512258 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-dd58r" Nov 28 13:21:12 crc kubenswrapper[4631]: I1128 13:21:12.512268 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 13:21:12 crc kubenswrapper[4631]: I1128 13:21:12.512642 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 13:21:12 crc kubenswrapper[4631]: E1128 13:21:12.512868 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-dd58r" podUID="8716de2d-f33e-4434-9076-6345669aaff8" Nov 28 13:21:12 crc kubenswrapper[4631]: I1128 13:21:12.512877 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 13:21:12 crc kubenswrapper[4631]: E1128 13:21:12.513097 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 13:21:12 crc kubenswrapper[4631]: E1128 13:21:12.513203 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 13:21:12 crc kubenswrapper[4631]: E1128 13:21:12.513328 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 13:21:12 crc kubenswrapper[4631]: I1128 13:21:12.519473 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 13:21:12 crc kubenswrapper[4631]: I1128 13:21:12.519500 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 13:21:12 crc kubenswrapper[4631]: I1128 13:21:12.519509 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 13:21:12 crc kubenswrapper[4631]: I1128 13:21:12.519522 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 13:21:12 crc kubenswrapper[4631]: I1128 13:21:12.519532 4631 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T13:21:12Z","lastTransitionTime":"2025-11-28T13:21:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 13:21:12 crc kubenswrapper[4631]: I1128 13:21:12.623143 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 13:21:12 crc kubenswrapper[4631]: I1128 13:21:12.623190 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 13:21:12 crc kubenswrapper[4631]: I1128 13:21:12.623206 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 13:21:12 crc kubenswrapper[4631]: I1128 13:21:12.623226 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 13:21:12 crc kubenswrapper[4631]: I1128 13:21:12.623237 4631 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T13:21:12Z","lastTransitionTime":"2025-11-28T13:21:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 13:21:12 crc kubenswrapper[4631]: I1128 13:21:12.725061 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 13:21:12 crc kubenswrapper[4631]: I1128 13:21:12.725090 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 13:21:12 crc kubenswrapper[4631]: I1128 13:21:12.725100 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 13:21:12 crc kubenswrapper[4631]: I1128 13:21:12.725113 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 13:21:12 crc kubenswrapper[4631]: I1128 13:21:12.725123 4631 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T13:21:12Z","lastTransitionTime":"2025-11-28T13:21:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 13:21:12 crc kubenswrapper[4631]: I1128 13:21:12.726223 4631 generic.go:334] "Generic (PLEG): container finished" podID="a0fde998-9a55-4734-9008-136b93f29d0c" containerID="aa6d834ad4feb21b6769f77bae27819530b93fb6c377a91aa7cd8e2419586fc9" exitCode=0 Nov 28 13:21:12 crc kubenswrapper[4631]: I1128 13:21:12.726329 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-8gtw6" event={"ID":"a0fde998-9a55-4734-9008-136b93f29d0c","Type":"ContainerDied","Data":"aa6d834ad4feb21b6769f77bae27819530b93fb6c377a91aa7cd8e2419586fc9"} Nov 28 13:21:12 crc kubenswrapper[4631]: I1128 13:21:12.731792 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-268tw" event={"ID":"a205f475-340d-4ec5-a860-d4a5c2ceb39e","Type":"ContainerStarted","Data":"9c674448c309477eb36de0eb0312fde682c3248fb66ed1d1d4100dc114fda786"} Nov 28 13:21:12 crc kubenswrapper[4631]: I1128 13:21:12.829661 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 13:21:12 crc kubenswrapper[4631]: I1128 13:21:12.829748 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 13:21:12 crc kubenswrapper[4631]: I1128 13:21:12.829765 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 13:21:12 crc kubenswrapper[4631]: I1128 13:21:12.829814 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 13:21:12 crc kubenswrapper[4631]: I1128 13:21:12.829829 4631 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T13:21:12Z","lastTransitionTime":"2025-11-28T13:21:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 13:21:12 crc kubenswrapper[4631]: I1128 13:21:12.932615 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 13:21:12 crc kubenswrapper[4631]: I1128 13:21:12.932672 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 13:21:12 crc kubenswrapper[4631]: I1128 13:21:12.932688 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 13:21:12 crc kubenswrapper[4631]: I1128 13:21:12.932713 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 13:21:12 crc kubenswrapper[4631]: I1128 13:21:12.932730 4631 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T13:21:12Z","lastTransitionTime":"2025-11-28T13:21:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 13:21:13 crc kubenswrapper[4631]: I1128 13:21:13.022204 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8716de2d-f33e-4434-9076-6345669aaff8-metrics-certs\") pod \"network-metrics-daemon-dd58r\" (UID: \"8716de2d-f33e-4434-9076-6345669aaff8\") " pod="openshift-multus/network-metrics-daemon-dd58r" Nov 28 13:21:13 crc kubenswrapper[4631]: E1128 13:21:13.022438 4631 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 28 13:21:13 crc kubenswrapper[4631]: E1128 13:21:13.022533 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8716de2d-f33e-4434-9076-6345669aaff8-metrics-certs podName:8716de2d-f33e-4434-9076-6345669aaff8 nodeName:}" failed. No retries permitted until 2025-11-28 13:21:17.022505184 +0000 UTC m=+33.829808528 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/8716de2d-f33e-4434-9076-6345669aaff8-metrics-certs") pod "network-metrics-daemon-dd58r" (UID: "8716de2d-f33e-4434-9076-6345669aaff8") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 28 13:21:13 crc kubenswrapper[4631]: I1128 13:21:13.039475 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 13:21:13 crc kubenswrapper[4631]: I1128 13:21:13.039543 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 13:21:13 crc kubenswrapper[4631]: I1128 13:21:13.039564 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 13:21:13 crc kubenswrapper[4631]: I1128 13:21:13.039598 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 13:21:13 crc kubenswrapper[4631]: I1128 13:21:13.039615 4631 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T13:21:13Z","lastTransitionTime":"2025-11-28T13:21:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 13:21:13 crc kubenswrapper[4631]: I1128 13:21:13.143568 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 13:21:13 crc kubenswrapper[4631]: I1128 13:21:13.143622 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 13:21:13 crc kubenswrapper[4631]: I1128 13:21:13.143646 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 13:21:13 crc kubenswrapper[4631]: I1128 13:21:13.143674 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 13:21:13 crc kubenswrapper[4631]: I1128 13:21:13.143694 4631 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T13:21:13Z","lastTransitionTime":"2025-11-28T13:21:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 13:21:13 crc kubenswrapper[4631]: I1128 13:21:13.246594 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 13:21:13 crc kubenswrapper[4631]: I1128 13:21:13.246644 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 13:21:13 crc kubenswrapper[4631]: I1128 13:21:13.246656 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 13:21:13 crc kubenswrapper[4631]: I1128 13:21:13.246680 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 13:21:13 crc kubenswrapper[4631]: I1128 13:21:13.246694 4631 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T13:21:13Z","lastTransitionTime":"2025-11-28T13:21:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 13:21:13 crc kubenswrapper[4631]: I1128 13:21:13.349771 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 13:21:13 crc kubenswrapper[4631]: I1128 13:21:13.349824 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 13:21:13 crc kubenswrapper[4631]: I1128 13:21:13.349841 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 13:21:13 crc kubenswrapper[4631]: I1128 13:21:13.349875 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 13:21:13 crc kubenswrapper[4631]: I1128 13:21:13.349892 4631 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T13:21:13Z","lastTransitionTime":"2025-11-28T13:21:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 13:21:13 crc kubenswrapper[4631]: I1128 13:21:13.453183 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 13:21:13 crc kubenswrapper[4631]: I1128 13:21:13.453245 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 13:21:13 crc kubenswrapper[4631]: I1128 13:21:13.453265 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 13:21:13 crc kubenswrapper[4631]: I1128 13:21:13.453322 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 13:21:13 crc kubenswrapper[4631]: I1128 13:21:13.453342 4631 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T13:21:13Z","lastTransitionTime":"2025-11-28T13:21:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 13:21:13 crc kubenswrapper[4631]: I1128 13:21:13.557251 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 13:21:13 crc kubenswrapper[4631]: I1128 13:21:13.557385 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 13:21:13 crc kubenswrapper[4631]: I1128 13:21:13.557404 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 13:21:13 crc kubenswrapper[4631]: I1128 13:21:13.557436 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 13:21:13 crc kubenswrapper[4631]: I1128 13:21:13.557457 4631 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T13:21:13Z","lastTransitionTime":"2025-11-28T13:21:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 13:21:13 crc kubenswrapper[4631]: I1128 13:21:13.659555 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 13:21:13 crc kubenswrapper[4631]: I1128 13:21:13.659587 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 13:21:13 crc kubenswrapper[4631]: I1128 13:21:13.659597 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 13:21:13 crc kubenswrapper[4631]: I1128 13:21:13.659610 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 13:21:13 crc kubenswrapper[4631]: I1128 13:21:13.659619 4631 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T13:21:13Z","lastTransitionTime":"2025-11-28T13:21:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 13:21:13 crc kubenswrapper[4631]: I1128 13:21:13.741423 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-8gtw6" event={"ID":"a0fde998-9a55-4734-9008-136b93f29d0c","Type":"ContainerStarted","Data":"5e9359a14c3878d8117e0b4d796bd3fd673378d6a1c37fada1e837ac68a25700"} Nov 28 13:21:13 crc kubenswrapper[4631]: I1128 13:21:13.762471 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 13:21:13 crc kubenswrapper[4631]: I1128 13:21:13.762516 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 13:21:13 crc kubenswrapper[4631]: I1128 13:21:13.762529 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 13:21:13 crc kubenswrapper[4631]: I1128 13:21:13.762550 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 13:21:13 crc kubenswrapper[4631]: I1128 13:21:13.762565 4631 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T13:21:13Z","lastTransitionTime":"2025-11-28T13:21:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 13:21:13 crc kubenswrapper[4631]: I1128 13:21:13.775938 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-8gtw6" podStartSLOduration=10.775884812 podStartE2EDuration="10.775884812s" podCreationTimestamp="2025-11-28 13:21:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 13:21:13.773983232 +0000 UTC m=+30.581286616" watchObservedRunningTime="2025-11-28 13:21:13.775884812 +0000 UTC m=+30.583188186" Nov 28 13:21:13 crc kubenswrapper[4631]: I1128 13:21:13.865597 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 13:21:13 crc kubenswrapper[4631]: I1128 13:21:13.866033 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 13:21:13 crc kubenswrapper[4631]: I1128 13:21:13.866056 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 13:21:13 crc kubenswrapper[4631]: I1128 13:21:13.866083 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 13:21:13 crc kubenswrapper[4631]: I1128 13:21:13.866101 4631 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T13:21:13Z","lastTransitionTime":"2025-11-28T13:21:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 13:21:13 crc kubenswrapper[4631]: I1128 13:21:13.972724 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 13:21:13 crc kubenswrapper[4631]: I1128 13:21:13.972797 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 13:21:13 crc kubenswrapper[4631]: I1128 13:21:13.972816 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 13:21:13 crc kubenswrapper[4631]: I1128 13:21:13.972848 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 13:21:13 crc kubenswrapper[4631]: I1128 13:21:13.972872 4631 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T13:21:13Z","lastTransitionTime":"2025-11-28T13:21:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 13:21:14 crc kubenswrapper[4631]: I1128 13:21:14.075782 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 13:21:14 crc kubenswrapper[4631]: I1128 13:21:14.075891 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 13:21:14 crc kubenswrapper[4631]: I1128 13:21:14.075906 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 13:21:14 crc kubenswrapper[4631]: I1128 13:21:14.075932 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 13:21:14 crc kubenswrapper[4631]: I1128 13:21:14.075947 4631 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T13:21:14Z","lastTransitionTime":"2025-11-28T13:21:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 13:21:14 crc kubenswrapper[4631]: I1128 13:21:14.179521 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 13:21:14 crc kubenswrapper[4631]: I1128 13:21:14.179612 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 13:21:14 crc kubenswrapper[4631]: I1128 13:21:14.179636 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 13:21:14 crc kubenswrapper[4631]: I1128 13:21:14.179670 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 13:21:14 crc kubenswrapper[4631]: I1128 13:21:14.179697 4631 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T13:21:14Z","lastTransitionTime":"2025-11-28T13:21:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 13:21:14 crc kubenswrapper[4631]: I1128 13:21:14.283478 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 13:21:14 crc kubenswrapper[4631]: I1128 13:21:14.283555 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 13:21:14 crc kubenswrapper[4631]: I1128 13:21:14.283580 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 13:21:14 crc kubenswrapper[4631]: I1128 13:21:14.283613 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 13:21:14 crc kubenswrapper[4631]: I1128 13:21:14.283640 4631 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T13:21:14Z","lastTransitionTime":"2025-11-28T13:21:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 13:21:14 crc kubenswrapper[4631]: I1128 13:21:14.386649 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 13:21:14 crc kubenswrapper[4631]: I1128 13:21:14.386707 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 13:21:14 crc kubenswrapper[4631]: I1128 13:21:14.386725 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 13:21:14 crc kubenswrapper[4631]: I1128 13:21:14.386763 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 13:21:14 crc kubenswrapper[4631]: I1128 13:21:14.386782 4631 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T13:21:14Z","lastTransitionTime":"2025-11-28T13:21:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 13:21:14 crc kubenswrapper[4631]: I1128 13:21:14.490178 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 13:21:14 crc kubenswrapper[4631]: I1128 13:21:14.490233 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 13:21:14 crc kubenswrapper[4631]: I1128 13:21:14.490248 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 13:21:14 crc kubenswrapper[4631]: I1128 13:21:14.490272 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 13:21:14 crc kubenswrapper[4631]: I1128 13:21:14.490305 4631 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T13:21:14Z","lastTransitionTime":"2025-11-28T13:21:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 13:21:14 crc kubenswrapper[4631]: I1128 13:21:14.512536 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 13:21:14 crc kubenswrapper[4631]: I1128 13:21:14.512588 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 13:21:14 crc kubenswrapper[4631]: I1128 13:21:14.512536 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 13:21:14 crc kubenswrapper[4631]: I1128 13:21:14.512748 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-dd58r" Nov 28 13:21:14 crc kubenswrapper[4631]: E1128 13:21:14.513031 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 13:21:14 crc kubenswrapper[4631]: E1128 13:21:14.513144 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 13:21:14 crc kubenswrapper[4631]: E1128 13:21:14.513245 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 13:21:14 crc kubenswrapper[4631]: E1128 13:21:14.513369 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-dd58r" podUID="8716de2d-f33e-4434-9076-6345669aaff8" Nov 28 13:21:14 crc kubenswrapper[4631]: I1128 13:21:14.593982 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 13:21:14 crc kubenswrapper[4631]: I1128 13:21:14.594028 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 13:21:14 crc kubenswrapper[4631]: I1128 13:21:14.594042 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 13:21:14 crc kubenswrapper[4631]: I1128 13:21:14.594062 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 13:21:14 crc kubenswrapper[4631]: I1128 13:21:14.594078 4631 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T13:21:14Z","lastTransitionTime":"2025-11-28T13:21:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 13:21:14 crc kubenswrapper[4631]: I1128 13:21:14.696641 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 13:21:14 crc kubenswrapper[4631]: I1128 13:21:14.696678 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 13:21:14 crc kubenswrapper[4631]: I1128 13:21:14.696687 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 13:21:14 crc kubenswrapper[4631]: I1128 13:21:14.696702 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 13:21:14 crc kubenswrapper[4631]: I1128 13:21:14.696711 4631 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T13:21:14Z","lastTransitionTime":"2025-11-28T13:21:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 13:21:14 crc kubenswrapper[4631]: I1128 13:21:14.751481 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-268tw" event={"ID":"a205f475-340d-4ec5-a860-d4a5c2ceb39e","Type":"ContainerStarted","Data":"da4a7ce2b9fcb91df9dab6ba86046a998e310a5cc62a8eefcd98ea01a26f8db1"} Nov 28 13:21:14 crc kubenswrapper[4631]: I1128 13:21:14.782077 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-268tw" podStartSLOduration=11.782048595 podStartE2EDuration="11.782048595s" podCreationTimestamp="2025-11-28 13:21:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 13:21:14.780745928 +0000 UTC m=+31.588049312" watchObservedRunningTime="2025-11-28 13:21:14.782048595 +0000 UTC m=+31.589351949" Nov 28 13:21:14 crc kubenswrapper[4631]: I1128 13:21:14.800056 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 13:21:14 crc kubenswrapper[4631]: I1128 13:21:14.800123 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 13:21:14 crc kubenswrapper[4631]: I1128 13:21:14.800150 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 13:21:14 crc kubenswrapper[4631]: I1128 13:21:14.800176 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 13:21:14 crc kubenswrapper[4631]: I1128 13:21:14.800193 4631 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T13:21:14Z","lastTransitionTime":"2025-11-28T13:21:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 13:21:14 crc kubenswrapper[4631]: I1128 13:21:14.902970 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 13:21:14 crc kubenswrapper[4631]: I1128 13:21:14.903015 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 13:21:14 crc kubenswrapper[4631]: I1128 13:21:14.903028 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 13:21:14 crc kubenswrapper[4631]: I1128 13:21:14.903047 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 13:21:14 crc kubenswrapper[4631]: I1128 13:21:14.903060 4631 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T13:21:14Z","lastTransitionTime":"2025-11-28T13:21:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 13:21:15 crc kubenswrapper[4631]: I1128 13:21:15.005583 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 13:21:15 crc kubenswrapper[4631]: I1128 13:21:15.005622 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 13:21:15 crc kubenswrapper[4631]: I1128 13:21:15.005658 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 13:21:15 crc kubenswrapper[4631]: I1128 13:21:15.005686 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 13:21:15 crc kubenswrapper[4631]: I1128 13:21:15.005699 4631 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T13:21:15Z","lastTransitionTime":"2025-11-28T13:21:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 13:21:15 crc kubenswrapper[4631]: I1128 13:21:15.109310 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 13:21:15 crc kubenswrapper[4631]: I1128 13:21:15.109638 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 13:21:15 crc kubenswrapper[4631]: I1128 13:21:15.109649 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 13:21:15 crc kubenswrapper[4631]: I1128 13:21:15.109665 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 13:21:15 crc kubenswrapper[4631]: I1128 13:21:15.109676 4631 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T13:21:15Z","lastTransitionTime":"2025-11-28T13:21:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 13:21:15 crc kubenswrapper[4631]: I1128 13:21:15.202016 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 13:21:15 crc kubenswrapper[4631]: I1128 13:21:15.202050 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 13:21:15 crc kubenswrapper[4631]: I1128 13:21:15.202059 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 13:21:15 crc kubenswrapper[4631]: I1128 13:21:15.202074 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 13:21:15 crc kubenswrapper[4631]: I1128 13:21:15.202085 4631 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T13:21:15Z","lastTransitionTime":"2025-11-28T13:21:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 13:21:15 crc kubenswrapper[4631]: I1128 13:21:15.243210 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-bq6bx"] Nov 28 13:21:15 crc kubenswrapper[4631]: I1128 13:21:15.243601 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-bq6bx" Nov 28 13:21:15 crc kubenswrapper[4631]: I1128 13:21:15.245619 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Nov 28 13:21:15 crc kubenswrapper[4631]: I1128 13:21:15.245731 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Nov 28 13:21:15 crc kubenswrapper[4631]: I1128 13:21:15.247423 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Nov 28 13:21:15 crc kubenswrapper[4631]: I1128 13:21:15.247524 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Nov 28 13:21:15 crc kubenswrapper[4631]: I1128 13:21:15.350147 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/44aff9a4-2d69-413f-aff8-5c01a5117269-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-bq6bx\" (UID: \"44aff9a4-2d69-413f-aff8-5c01a5117269\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-bq6bx" Nov 28 13:21:15 crc kubenswrapper[4631]: I1128 13:21:15.350213 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/44aff9a4-2d69-413f-aff8-5c01a5117269-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-bq6bx\" (UID: \"44aff9a4-2d69-413f-aff8-5c01a5117269\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-bq6bx" Nov 28 13:21:15 crc kubenswrapper[4631]: I1128 13:21:15.350238 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/44aff9a4-2d69-413f-aff8-5c01a5117269-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-bq6bx\" (UID: \"44aff9a4-2d69-413f-aff8-5c01a5117269\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-bq6bx" Nov 28 13:21:15 crc kubenswrapper[4631]: I1128 13:21:15.350261 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/44aff9a4-2d69-413f-aff8-5c01a5117269-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-bq6bx\" (UID: \"44aff9a4-2d69-413f-aff8-5c01a5117269\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-bq6bx" Nov 28 13:21:15 crc kubenswrapper[4631]: I1128 13:21:15.350307 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/44aff9a4-2d69-413f-aff8-5c01a5117269-service-ca\") pod \"cluster-version-operator-5c965bbfc6-bq6bx\" (UID: \"44aff9a4-2d69-413f-aff8-5c01a5117269\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-bq6bx" Nov 28 13:21:15 crc kubenswrapper[4631]: I1128 13:21:15.450837 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/44aff9a4-2d69-413f-aff8-5c01a5117269-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-bq6bx\" (UID: \"44aff9a4-2d69-413f-aff8-5c01a5117269\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-bq6bx" Nov 28 13:21:15 crc kubenswrapper[4631]: I1128 13:21:15.450881 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/44aff9a4-2d69-413f-aff8-5c01a5117269-service-ca\") pod \"cluster-version-operator-5c965bbfc6-bq6bx\" (UID: \"44aff9a4-2d69-413f-aff8-5c01a5117269\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-bq6bx" Nov 28 13:21:15 crc kubenswrapper[4631]: I1128 13:21:15.450941 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/44aff9a4-2d69-413f-aff8-5c01a5117269-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-bq6bx\" (UID: \"44aff9a4-2d69-413f-aff8-5c01a5117269\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-bq6bx" Nov 28 13:21:15 crc kubenswrapper[4631]: I1128 13:21:15.450979 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/44aff9a4-2d69-413f-aff8-5c01a5117269-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-bq6bx\" (UID: \"44aff9a4-2d69-413f-aff8-5c01a5117269\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-bq6bx" Nov 28 13:21:15 crc kubenswrapper[4631]: I1128 13:21:15.451002 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/44aff9a4-2d69-413f-aff8-5c01a5117269-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-bq6bx\" (UID: \"44aff9a4-2d69-413f-aff8-5c01a5117269\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-bq6bx" Nov 28 13:21:15 crc kubenswrapper[4631]: I1128 13:21:15.451059 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/44aff9a4-2d69-413f-aff8-5c01a5117269-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-bq6bx\" (UID: \"44aff9a4-2d69-413f-aff8-5c01a5117269\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-bq6bx" Nov 28 13:21:15 crc kubenswrapper[4631]: I1128 13:21:15.451737 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/44aff9a4-2d69-413f-aff8-5c01a5117269-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-bq6bx\" (UID: \"44aff9a4-2d69-413f-aff8-5c01a5117269\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-bq6bx" Nov 28 13:21:15 crc kubenswrapper[4631]: I1128 13:21:15.452772 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/44aff9a4-2d69-413f-aff8-5c01a5117269-service-ca\") pod \"cluster-version-operator-5c965bbfc6-bq6bx\" (UID: \"44aff9a4-2d69-413f-aff8-5c01a5117269\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-bq6bx" Nov 28 13:21:15 crc kubenswrapper[4631]: I1128 13:21:15.456840 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/44aff9a4-2d69-413f-aff8-5c01a5117269-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-bq6bx\" (UID: \"44aff9a4-2d69-413f-aff8-5c01a5117269\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-bq6bx" Nov 28 13:21:15 crc kubenswrapper[4631]: I1128 13:21:15.479433 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/44aff9a4-2d69-413f-aff8-5c01a5117269-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-bq6bx\" (UID: \"44aff9a4-2d69-413f-aff8-5c01a5117269\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-bq6bx" Nov 28 13:21:15 crc kubenswrapper[4631]: I1128 13:21:15.582480 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-bq6bx" Nov 28 13:21:15 crc kubenswrapper[4631]: I1128 13:21:15.755726 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-bq6bx" event={"ID":"44aff9a4-2d69-413f-aff8-5c01a5117269","Type":"ContainerStarted","Data":"70b6bdbdd1e6e1ed5f3deec4ad641575245b0faacdfbbc35744fef1480fe5080"} Nov 28 13:21:15 crc kubenswrapper[4631]: I1128 13:21:15.755794 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-268tw" Nov 28 13:21:15 crc kubenswrapper[4631]: I1128 13:21:15.755750 4631 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 28 13:21:15 crc kubenswrapper[4631]: I1128 13:21:15.756365 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-268tw" Nov 28 13:21:16 crc kubenswrapper[4631]: I1128 13:21:16.093787 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-268tw" Nov 28 13:21:16 crc kubenswrapper[4631]: I1128 13:21:16.104113 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-268tw" Nov 28 13:21:16 crc kubenswrapper[4631]: I1128 13:21:16.512794 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 13:21:16 crc kubenswrapper[4631]: I1128 13:21:16.512828 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 13:21:16 crc kubenswrapper[4631]: I1128 13:21:16.512961 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 13:21:16 crc kubenswrapper[4631]: E1128 13:21:16.513182 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 13:21:16 crc kubenswrapper[4631]: I1128 13:21:16.513312 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-dd58r" Nov 28 13:21:16 crc kubenswrapper[4631]: E1128 13:21:16.513356 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 13:21:16 crc kubenswrapper[4631]: E1128 13:21:16.513453 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 13:21:16 crc kubenswrapper[4631]: E1128 13:21:16.513636 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-dd58r" podUID="8716de2d-f33e-4434-9076-6345669aaff8" Nov 28 13:21:16 crc kubenswrapper[4631]: I1128 13:21:16.674354 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-dd58r"] Nov 28 13:21:16 crc kubenswrapper[4631]: I1128 13:21:16.759979 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-dd58r" Nov 28 13:21:16 crc kubenswrapper[4631]: E1128 13:21:16.760085 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-dd58r" podUID="8716de2d-f33e-4434-9076-6345669aaff8" Nov 28 13:21:16 crc kubenswrapper[4631]: I1128 13:21:16.761044 4631 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 28 13:21:16 crc kubenswrapper[4631]: I1128 13:21:16.761501 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-bq6bx" event={"ID":"44aff9a4-2d69-413f-aff8-5c01a5117269","Type":"ContainerStarted","Data":"eca5c4e3c68649575bbe5e06603e14292542d3c8e47a47ef928fd7660a02806c"} Nov 28 13:21:17 crc kubenswrapper[4631]: I1128 13:21:17.069856 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8716de2d-f33e-4434-9076-6345669aaff8-metrics-certs\") pod \"network-metrics-daemon-dd58r\" (UID: \"8716de2d-f33e-4434-9076-6345669aaff8\") " pod="openshift-multus/network-metrics-daemon-dd58r" Nov 28 13:21:17 crc kubenswrapper[4631]: E1128 13:21:17.070137 4631 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 28 13:21:17 crc kubenswrapper[4631]: E1128 13:21:17.070270 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8716de2d-f33e-4434-9076-6345669aaff8-metrics-certs podName:8716de2d-f33e-4434-9076-6345669aaff8 nodeName:}" failed. No retries permitted until 2025-11-28 13:21:25.070232906 +0000 UTC m=+41.877536290 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/8716de2d-f33e-4434-9076-6345669aaff8-metrics-certs") pod "network-metrics-daemon-dd58r" (UID: "8716de2d-f33e-4434-9076-6345669aaff8") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 28 13:21:17 crc kubenswrapper[4631]: I1128 13:21:17.763398 4631 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 28 13:21:18 crc kubenswrapper[4631]: I1128 13:21:18.384677 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 13:21:18 crc kubenswrapper[4631]: E1128 13:21:18.385043 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 13:21:34.385021353 +0000 UTC m=+51.192324707 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 13:21:18 crc kubenswrapper[4631]: I1128 13:21:18.385206 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 13:21:18 crc kubenswrapper[4631]: I1128 13:21:18.385237 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 13:21:18 crc kubenswrapper[4631]: I1128 13:21:18.385317 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 13:21:18 crc kubenswrapper[4631]: I1128 13:21:18.385344 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 13:21:18 crc kubenswrapper[4631]: E1128 13:21:18.385618 4631 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 28 13:21:18 crc kubenswrapper[4631]: E1128 13:21:18.385683 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-28 13:21:34.385670497 +0000 UTC m=+51.192973841 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 28 13:21:18 crc kubenswrapper[4631]: E1128 13:21:18.385681 4631 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 28 13:21:18 crc kubenswrapper[4631]: E1128 13:21:18.385709 4631 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 28 13:21:18 crc kubenswrapper[4631]: E1128 13:21:18.385725 4631 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 28 13:21:18 crc kubenswrapper[4631]: E1128 13:21:18.385771 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-28 13:21:34.385756729 +0000 UTC m=+51.193060103 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 28 13:21:18 crc kubenswrapper[4631]: E1128 13:21:18.385804 4631 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 28 13:21:18 crc kubenswrapper[4631]: E1128 13:21:18.385838 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-28 13:21:34.3858296 +0000 UTC m=+51.193132964 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 28 13:21:18 crc kubenswrapper[4631]: E1128 13:21:18.385993 4631 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 28 13:21:18 crc kubenswrapper[4631]: E1128 13:21:18.386120 4631 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 28 13:21:18 crc kubenswrapper[4631]: E1128 13:21:18.386273 4631 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 28 13:21:18 crc kubenswrapper[4631]: E1128 13:21:18.386373 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-28 13:21:34.386361701 +0000 UTC m=+51.193665045 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 28 13:21:18 crc kubenswrapper[4631]: I1128 13:21:18.512649 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-dd58r" Nov 28 13:21:18 crc kubenswrapper[4631]: I1128 13:21:18.512732 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 13:21:18 crc kubenswrapper[4631]: I1128 13:21:18.512661 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 13:21:18 crc kubenswrapper[4631]: E1128 13:21:18.512846 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-dd58r" podUID="8716de2d-f33e-4434-9076-6345669aaff8" Nov 28 13:21:18 crc kubenswrapper[4631]: E1128 13:21:18.512920 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 13:21:18 crc kubenswrapper[4631]: I1128 13:21:18.512935 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 13:21:18 crc kubenswrapper[4631]: E1128 13:21:18.512997 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 13:21:18 crc kubenswrapper[4631]: E1128 13:21:18.513270 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 13:21:18 crc kubenswrapper[4631]: I1128 13:21:18.837534 4631 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Nov 28 13:21:18 crc kubenswrapper[4631]: I1128 13:21:18.837646 4631 kubelet_node_status.go:538] "Fast updating node status as it just became ready" Nov 28 13:21:18 crc kubenswrapper[4631]: I1128 13:21:18.887874 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-bq6bx" podStartSLOduration=15.887855941 podStartE2EDuration="15.887855941s" podCreationTimestamp="2025-11-28 13:21:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 13:21:16.779439312 +0000 UTC m=+33.586742666" watchObservedRunningTime="2025-11-28 13:21:18.887855941 +0000 UTC m=+35.695159295" Nov 28 13:21:18 crc kubenswrapper[4631]: I1128 13:21:18.888549 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-8tbnv"] Nov 28 13:21:18 crc kubenswrapper[4631]: I1128 13:21:18.889456 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-jgx6f"] Nov 28 13:21:18 crc kubenswrapper[4631]: I1128 13:21:18.889608 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8tbnv" Nov 28 13:21:18 crc kubenswrapper[4631]: I1128 13:21:18.890533 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-jgx6f" Nov 28 13:21:18 crc kubenswrapper[4631]: I1128 13:21:18.917137 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Nov 28 13:21:18 crc kubenswrapper[4631]: I1128 13:21:18.917387 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Nov 28 13:21:18 crc kubenswrapper[4631]: I1128 13:21:18.917529 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Nov 28 13:21:18 crc kubenswrapper[4631]: I1128 13:21:18.917685 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Nov 28 13:21:18 crc kubenswrapper[4631]: I1128 13:21:18.917830 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Nov 28 13:21:18 crc kubenswrapper[4631]: I1128 13:21:18.917975 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Nov 28 13:21:18 crc kubenswrapper[4631]: I1128 13:21:18.918119 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Nov 28 13:21:18 crc kubenswrapper[4631]: I1128 13:21:18.918268 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Nov 28 13:21:18 crc kubenswrapper[4631]: I1128 13:21:18.918464 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Nov 28 13:21:18 crc kubenswrapper[4631]: I1128 13:21:18.918509 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Nov 28 13:21:18 crc kubenswrapper[4631]: I1128 13:21:18.918466 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Nov 28 13:21:18 crc kubenswrapper[4631]: I1128 13:21:18.918689 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Nov 28 13:21:18 crc kubenswrapper[4631]: I1128 13:21:18.918707 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Nov 28 13:21:18 crc kubenswrapper[4631]: I1128 13:21:18.918740 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Nov 28 13:21:18 crc kubenswrapper[4631]: I1128 13:21:18.918779 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Nov 28 13:21:18 crc kubenswrapper[4631]: I1128 13:21:18.918921 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Nov 28 13:21:18 crc kubenswrapper[4631]: I1128 13:21:18.919019 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Nov 28 13:21:18 crc kubenswrapper[4631]: I1128 13:21:18.919075 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Nov 28 13:21:18 crc kubenswrapper[4631]: I1128 13:21:18.919472 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-c2rks"] Nov 28 13:21:18 crc kubenswrapper[4631]: I1128 13:21:18.920081 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-r5b6b"] Nov 28 13:21:18 crc kubenswrapper[4631]: I1128 13:21:18.920114 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-c2rks" Nov 28 13:21:18 crc kubenswrapper[4631]: I1128 13:21:18.920912 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-r5b6b" Nov 28 13:21:18 crc kubenswrapper[4631]: I1128 13:21:18.920921 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-2ch2b"] Nov 28 13:21:18 crc kubenswrapper[4631]: I1128 13:21:18.921919 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Nov 28 13:21:18 crc kubenswrapper[4631]: I1128 13:21:18.921941 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-bf7bx"] Nov 28 13:21:18 crc kubenswrapper[4631]: I1128 13:21:18.922061 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-2ch2b" Nov 28 13:21:18 crc kubenswrapper[4631]: I1128 13:21:18.922500 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-bf7bx" Nov 28 13:21:18 crc kubenswrapper[4631]: I1128 13:21:18.926090 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-7gmxj"] Nov 28 13:21:18 crc kubenswrapper[4631]: I1128 13:21:18.926587 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-ddx6v"] Nov 28 13:21:18 crc kubenswrapper[4631]: I1128 13:21:18.926870 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-rwpv9"] Nov 28 13:21:18 crc kubenswrapper[4631]: I1128 13:21:18.927248 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-rwpv9" Nov 28 13:21:18 crc kubenswrapper[4631]: I1128 13:21:18.927655 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ddx6v" Nov 28 13:21:18 crc kubenswrapper[4631]: I1128 13:21:18.929020 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-7gmxj" Nov 28 13:21:18 crc kubenswrapper[4631]: I1128 13:21:18.930365 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-qcd8n"] Nov 28 13:21:18 crc kubenswrapper[4631]: I1128 13:21:18.930930 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-qcd8n" Nov 28 13:21:18 crc kubenswrapper[4631]: I1128 13:21:18.931132 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Nov 28 13:21:18 crc kubenswrapper[4631]: I1128 13:21:18.931686 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-hgsfk"] Nov 28 13:21:18 crc kubenswrapper[4631]: I1128 13:21:18.932099 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-hgsfk" Nov 28 13:21:18 crc kubenswrapper[4631]: I1128 13:21:18.932305 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Nov 28 13:21:18 crc kubenswrapper[4631]: I1128 13:21:18.932504 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Nov 28 13:21:18 crc kubenswrapper[4631]: I1128 13:21:18.934065 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-xqbvh"] Nov 28 13:21:18 crc kubenswrapper[4631]: I1128 13:21:18.934938 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-vst8c"] Nov 28 13:21:18 crc kubenswrapper[4631]: I1128 13:21:18.942430 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-xqbvh" Nov 28 13:21:18 crc kubenswrapper[4631]: I1128 13:21:18.943882 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-vst8c" Nov 28 13:21:18 crc kubenswrapper[4631]: I1128 13:21:18.952271 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-659jg"] Nov 28 13:21:18 crc kubenswrapper[4631]: I1128 13:21:18.970011 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-659jg" Nov 28 13:21:18 crc kubenswrapper[4631]: I1128 13:21:18.972202 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-7977d"] Nov 28 13:21:18 crc kubenswrapper[4631]: I1128 13:21:18.973667 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Nov 28 13:21:18 crc kubenswrapper[4631]: I1128 13:21:18.975795 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-vhmzs"] Nov 28 13:21:18 crc kubenswrapper[4631]: I1128 13:21:18.976589 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-vhmzs" Nov 28 13:21:18 crc kubenswrapper[4631]: I1128 13:21:18.976676 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-7977d" Nov 28 13:21:18 crc kubenswrapper[4631]: I1128 13:21:18.980518 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Nov 28 13:21:18 crc kubenswrapper[4631]: I1128 13:21:18.980798 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Nov 28 13:21:18 crc kubenswrapper[4631]: I1128 13:21:18.980905 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Nov 28 13:21:18 crc kubenswrapper[4631]: I1128 13:21:18.981048 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Nov 28 13:21:18 crc kubenswrapper[4631]: I1128 13:21:18.981197 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Nov 28 13:21:18 crc kubenswrapper[4631]: I1128 13:21:18.981306 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Nov 28 13:21:18 crc kubenswrapper[4631]: I1128 13:21:18.981404 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Nov 28 13:21:18 crc kubenswrapper[4631]: I1128 13:21:18.981496 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Nov 28 13:21:18 crc kubenswrapper[4631]: I1128 13:21:18.981593 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Nov 28 13:21:18 crc kubenswrapper[4631]: I1128 13:21:18.981679 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Nov 28 13:21:18 crc kubenswrapper[4631]: I1128 13:21:18.981917 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Nov 28 13:21:18 crc kubenswrapper[4631]: I1128 13:21:18.982010 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Nov 28 13:21:18 crc kubenswrapper[4631]: I1128 13:21:18.982408 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Nov 28 13:21:18 crc kubenswrapper[4631]: I1128 13:21:18.982627 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Nov 28 13:21:18 crc kubenswrapper[4631]: I1128 13:21:18.982725 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Nov 28 13:21:18 crc kubenswrapper[4631]: I1128 13:21:18.982841 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Nov 28 13:21:18 crc kubenswrapper[4631]: I1128 13:21:18.982956 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Nov 28 13:21:18 crc kubenswrapper[4631]: I1128 13:21:18.983088 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Nov 28 13:21:18 crc kubenswrapper[4631]: I1128 13:21:18.983138 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Nov 28 13:21:18 crc kubenswrapper[4631]: I1128 13:21:18.983228 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Nov 28 13:21:18 crc kubenswrapper[4631]: I1128 13:21:18.983333 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Nov 28 13:21:18 crc kubenswrapper[4631]: I1128 13:21:18.983434 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Nov 28 13:21:18 crc kubenswrapper[4631]: I1128 13:21:18.983528 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Nov 28 13:21:18 crc kubenswrapper[4631]: I1128 13:21:18.983607 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Nov 28 13:21:18 crc kubenswrapper[4631]: I1128 13:21:18.983717 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Nov 28 13:21:18 crc kubenswrapper[4631]: I1128 13:21:18.983837 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Nov 28 13:21:18 crc kubenswrapper[4631]: I1128 13:21:18.983942 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Nov 28 13:21:18 crc kubenswrapper[4631]: I1128 13:21:18.984047 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Nov 28 13:21:18 crc kubenswrapper[4631]: I1128 13:21:18.984152 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Nov 28 13:21:18 crc kubenswrapper[4631]: I1128 13:21:18.984248 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Nov 28 13:21:18 crc kubenswrapper[4631]: I1128 13:21:18.984512 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Nov 28 13:21:18 crc kubenswrapper[4631]: I1128 13:21:18.984821 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Nov 28 13:21:18 crc kubenswrapper[4631]: I1128 13:21:18.985045 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Nov 28 13:21:18 crc kubenswrapper[4631]: I1128 13:21:18.985043 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Nov 28 13:21:18 crc kubenswrapper[4631]: I1128 13:21:18.985630 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Nov 28 13:21:18 crc kubenswrapper[4631]: I1128 13:21:18.985878 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Nov 28 13:21:18 crc kubenswrapper[4631]: I1128 13:21:18.985974 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Nov 28 13:21:18 crc kubenswrapper[4631]: I1128 13:21:18.986119 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Nov 28 13:21:18 crc kubenswrapper[4631]: I1128 13:21:18.986210 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Nov 28 13:21:18 crc kubenswrapper[4631]: I1128 13:21:18.986330 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Nov 28 13:21:18 crc kubenswrapper[4631]: I1128 13:21:18.988728 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Nov 28 13:21:18 crc kubenswrapper[4631]: I1128 13:21:18.988918 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Nov 28 13:21:18 crc kubenswrapper[4631]: I1128 13:21:18.989049 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Nov 28 13:21:18 crc kubenswrapper[4631]: I1128 13:21:18.989434 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-8tbnv"] Nov 28 13:21:18 crc kubenswrapper[4631]: I1128 13:21:18.989469 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-r5b6b"] Nov 28 13:21:18 crc kubenswrapper[4631]: I1128 13:21:18.989483 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-jgx6f"] Nov 28 13:21:18 crc kubenswrapper[4631]: I1128 13:21:18.993979 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/cbd1163d-84ad-427f-b270-787520942822-available-featuregates\") pod \"openshift-config-operator-7777fb866f-rwpv9\" (UID: \"cbd1163d-84ad-427f-b270-787520942822\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-rwpv9" Nov 28 13:21:18 crc kubenswrapper[4631]: I1128 13:21:18.994010 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ccct6\" (UniqueName: \"kubernetes.io/projected/cbd1163d-84ad-427f-b270-787520942822-kube-api-access-ccct6\") pod \"openshift-config-operator-7777fb866f-rwpv9\" (UID: \"cbd1163d-84ad-427f-b270-787520942822\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-rwpv9" Nov 28 13:21:18 crc kubenswrapper[4631]: I1128 13:21:18.994034 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/aa813d64-b22e-40a0-b72a-b4e220e85f5a-etcd-client\") pod \"apiserver-7bbb656c7d-8tbnv\" (UID: \"aa813d64-b22e-40a0-b72a-b4e220e85f5a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8tbnv" Nov 28 13:21:18 crc kubenswrapper[4631]: I1128 13:21:18.994055 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/5e56718f-0cda-4524-aa3a-068bc2e7d569-image-import-ca\") pod \"apiserver-76f77b778f-jgx6f\" (UID: \"5e56718f-0cda-4524-aa3a-068bc2e7d569\") " pod="openshift-apiserver/apiserver-76f77b778f-jgx6f" Nov 28 13:21:18 crc kubenswrapper[4631]: I1128 13:21:18.994072 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/aa813d64-b22e-40a0-b72a-b4e220e85f5a-audit-policies\") pod \"apiserver-7bbb656c7d-8tbnv\" (UID: \"aa813d64-b22e-40a0-b72a-b4e220e85f5a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8tbnv" Nov 28 13:21:18 crc kubenswrapper[4631]: I1128 13:21:18.994088 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e7c7bd85-4add-4bda-a4fa-71a7b75787e4-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-7gmxj\" (UID: \"e7c7bd85-4add-4bda-a4fa-71a7b75787e4\") " pod="openshift-controller-manager/controller-manager-879f6c89f-7gmxj" Nov 28 13:21:18 crc kubenswrapper[4631]: I1128 13:21:18.994106 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5e56718f-0cda-4524-aa3a-068bc2e7d569-config\") pod \"apiserver-76f77b778f-jgx6f\" (UID: \"5e56718f-0cda-4524-aa3a-068bc2e7d569\") " pod="openshift-apiserver/apiserver-76f77b778f-jgx6f" Nov 28 13:21:18 crc kubenswrapper[4631]: I1128 13:21:18.994120 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r28b8\" (UniqueName: \"kubernetes.io/projected/5e56718f-0cda-4524-aa3a-068bc2e7d569-kube-api-access-r28b8\") pod \"apiserver-76f77b778f-jgx6f\" (UID: \"5e56718f-0cda-4524-aa3a-068bc2e7d569\") " pod="openshift-apiserver/apiserver-76f77b778f-jgx6f" Nov 28 13:21:18 crc kubenswrapper[4631]: I1128 13:21:18.994137 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/594fba15-6b93-4f34-af64-403a1b2d694c-client-ca\") pod \"route-controller-manager-6576b87f9c-ddx6v\" (UID: \"594fba15-6b93-4f34-af64-403a1b2d694c\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ddx6v" Nov 28 13:21:18 crc kubenswrapper[4631]: I1128 13:21:18.994156 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g6rlj\" (UniqueName: \"kubernetes.io/projected/6947cfe2-0015-42bd-b68e-465e50d41f40-kube-api-access-g6rlj\") pod \"cluster-image-registry-operator-dc59b4c8b-r5b6b\" (UID: \"6947cfe2-0015-42bd-b68e-465e50d41f40\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-r5b6b" Nov 28 13:21:18 crc kubenswrapper[4631]: I1128 13:21:18.994173 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/594fba15-6b93-4f34-af64-403a1b2d694c-serving-cert\") pod \"route-controller-manager-6576b87f9c-ddx6v\" (UID: \"594fba15-6b93-4f34-af64-403a1b2d694c\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ddx6v" Nov 28 13:21:18 crc kubenswrapper[4631]: I1128 13:21:18.994193 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/aa813d64-b22e-40a0-b72a-b4e220e85f5a-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-8tbnv\" (UID: \"aa813d64-b22e-40a0-b72a-b4e220e85f5a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8tbnv" Nov 28 13:21:18 crc kubenswrapper[4631]: I1128 13:21:18.994211 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/5e56718f-0cda-4524-aa3a-068bc2e7d569-etcd-serving-ca\") pod \"apiserver-76f77b778f-jgx6f\" (UID: \"5e56718f-0cda-4524-aa3a-068bc2e7d569\") " pod="openshift-apiserver/apiserver-76f77b778f-jgx6f" Nov 28 13:21:18 crc kubenswrapper[4631]: I1128 13:21:18.994227 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/5e56718f-0cda-4524-aa3a-068bc2e7d569-encryption-config\") pod \"apiserver-76f77b778f-jgx6f\" (UID: \"5e56718f-0cda-4524-aa3a-068bc2e7d569\") " pod="openshift-apiserver/apiserver-76f77b778f-jgx6f" Nov 28 13:21:18 crc kubenswrapper[4631]: I1128 13:21:18.994262 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/6947cfe2-0015-42bd-b68e-465e50d41f40-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-r5b6b\" (UID: \"6947cfe2-0015-42bd-b68e-465e50d41f40\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-r5b6b" Nov 28 13:21:18 crc kubenswrapper[4631]: I1128 13:21:18.994279 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d1b540aa-a7ac-4f13-b430-91e1328d6602-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-bf7bx\" (UID: \"d1b540aa-a7ac-4f13-b430-91e1328d6602\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-bf7bx" Nov 28 13:21:18 crc kubenswrapper[4631]: I1128 13:21:18.994314 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7c7bd85-4add-4bda-a4fa-71a7b75787e4-config\") pod \"controller-manager-879f6c89f-7gmxj\" (UID: \"e7c7bd85-4add-4bda-a4fa-71a7b75787e4\") " pod="openshift-controller-manager/controller-manager-879f6c89f-7gmxj" Nov 28 13:21:18 crc kubenswrapper[4631]: I1128 13:21:18.994330 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/5e56718f-0cda-4524-aa3a-068bc2e7d569-audit\") pod \"apiserver-76f77b778f-jgx6f\" (UID: \"5e56718f-0cda-4524-aa3a-068bc2e7d569\") " pod="openshift-apiserver/apiserver-76f77b778f-jgx6f" Nov 28 13:21:18 crc kubenswrapper[4631]: I1128 13:21:18.994345 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5e56718f-0cda-4524-aa3a-068bc2e7d569-trusted-ca-bundle\") pod \"apiserver-76f77b778f-jgx6f\" (UID: \"5e56718f-0cda-4524-aa3a-068bc2e7d569\") " pod="openshift-apiserver/apiserver-76f77b778f-jgx6f" Nov 28 13:21:18 crc kubenswrapper[4631]: I1128 13:21:18.994361 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/aa813d64-b22e-40a0-b72a-b4e220e85f5a-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-8tbnv\" (UID: \"aa813d64-b22e-40a0-b72a-b4e220e85f5a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8tbnv" Nov 28 13:21:18 crc kubenswrapper[4631]: I1128 13:21:18.994379 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/6947cfe2-0015-42bd-b68e-465e50d41f40-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-r5b6b\" (UID: \"6947cfe2-0015-42bd-b68e-465e50d41f40\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-r5b6b" Nov 28 13:21:18 crc kubenswrapper[4631]: I1128 13:21:18.994395 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d1b540aa-a7ac-4f13-b430-91e1328d6602-serving-cert\") pod \"authentication-operator-69f744f599-bf7bx\" (UID: \"d1b540aa-a7ac-4f13-b430-91e1328d6602\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-bf7bx" Nov 28 13:21:18 crc kubenswrapper[4631]: I1128 13:21:18.994410 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mfgnq\" (UniqueName: \"kubernetes.io/projected/594fba15-6b93-4f34-af64-403a1b2d694c-kube-api-access-mfgnq\") pod \"route-controller-manager-6576b87f9c-ddx6v\" (UID: \"594fba15-6b93-4f34-af64-403a1b2d694c\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ddx6v" Nov 28 13:21:18 crc kubenswrapper[4631]: I1128 13:21:18.994426 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d8jfw\" (UniqueName: \"kubernetes.io/projected/e7c7bd85-4add-4bda-a4fa-71a7b75787e4-kube-api-access-d8jfw\") pod \"controller-manager-879f6c89f-7gmxj\" (UID: \"e7c7bd85-4add-4bda-a4fa-71a7b75787e4\") " pod="openshift-controller-manager/controller-manager-879f6c89f-7gmxj" Nov 28 13:21:18 crc kubenswrapper[4631]: I1128 13:21:18.994451 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cbd1163d-84ad-427f-b270-787520942822-serving-cert\") pod \"openshift-config-operator-7777fb866f-rwpv9\" (UID: \"cbd1163d-84ad-427f-b270-787520942822\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-rwpv9" Nov 28 13:21:18 crc kubenswrapper[4631]: I1128 13:21:18.994468 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e7c7bd85-4add-4bda-a4fa-71a7b75787e4-client-ca\") pod \"controller-manager-879f6c89f-7gmxj\" (UID: \"e7c7bd85-4add-4bda-a4fa-71a7b75787e4\") " pod="openshift-controller-manager/controller-manager-879f6c89f-7gmxj" Nov 28 13:21:18 crc kubenswrapper[4631]: I1128 13:21:18.994484 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/5e56718f-0cda-4524-aa3a-068bc2e7d569-node-pullsecrets\") pod \"apiserver-76f77b778f-jgx6f\" (UID: \"5e56718f-0cda-4524-aa3a-068bc2e7d569\") " pod="openshift-apiserver/apiserver-76f77b778f-jgx6f" Nov 28 13:21:18 crc kubenswrapper[4631]: I1128 13:21:18.994501 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vmtq8\" (UniqueName: \"kubernetes.io/projected/889c117d-ea46-4f02-a2e0-42a47c6e4683-kube-api-access-vmtq8\") pod \"machine-api-operator-5694c8668f-2ch2b\" (UID: \"889c117d-ea46-4f02-a2e0-42a47c6e4683\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-2ch2b" Nov 28 13:21:18 crc kubenswrapper[4631]: I1128 13:21:18.994517 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d1b540aa-a7ac-4f13-b430-91e1328d6602-config\") pod \"authentication-operator-69f744f599-bf7bx\" (UID: \"d1b540aa-a7ac-4f13-b430-91e1328d6602\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-bf7bx" Nov 28 13:21:18 crc kubenswrapper[4631]: I1128 13:21:18.994536 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d1b540aa-a7ac-4f13-b430-91e1328d6602-service-ca-bundle\") pod \"authentication-operator-69f744f599-bf7bx\" (UID: \"d1b540aa-a7ac-4f13-b430-91e1328d6602\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-bf7bx" Nov 28 13:21:18 crc kubenswrapper[4631]: I1128 13:21:18.994552 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/aa813d64-b22e-40a0-b72a-b4e220e85f5a-serving-cert\") pod \"apiserver-7bbb656c7d-8tbnv\" (UID: \"aa813d64-b22e-40a0-b72a-b4e220e85f5a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8tbnv" Nov 28 13:21:18 crc kubenswrapper[4631]: I1128 13:21:18.994570 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/889c117d-ea46-4f02-a2e0-42a47c6e4683-config\") pod \"machine-api-operator-5694c8668f-2ch2b\" (UID: \"889c117d-ea46-4f02-a2e0-42a47c6e4683\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-2ch2b" Nov 28 13:21:18 crc kubenswrapper[4631]: I1128 13:21:18.994586 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8l2cr\" (UniqueName: \"kubernetes.io/projected/aa813d64-b22e-40a0-b72a-b4e220e85f5a-kube-api-access-8l2cr\") pod \"apiserver-7bbb656c7d-8tbnv\" (UID: \"aa813d64-b22e-40a0-b72a-b4e220e85f5a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8tbnv" Nov 28 13:21:18 crc kubenswrapper[4631]: I1128 13:21:18.994612 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/aa813d64-b22e-40a0-b72a-b4e220e85f5a-encryption-config\") pod \"apiserver-7bbb656c7d-8tbnv\" (UID: \"aa813d64-b22e-40a0-b72a-b4e220e85f5a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8tbnv" Nov 28 13:21:18 crc kubenswrapper[4631]: I1128 13:21:18.994627 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/aa813d64-b22e-40a0-b72a-b4e220e85f5a-audit-dir\") pod \"apiserver-7bbb656c7d-8tbnv\" (UID: \"aa813d64-b22e-40a0-b72a-b4e220e85f5a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8tbnv" Nov 28 13:21:18 crc kubenswrapper[4631]: I1128 13:21:18.994652 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5e56718f-0cda-4524-aa3a-068bc2e7d569-serving-cert\") pod \"apiserver-76f77b778f-jgx6f\" (UID: \"5e56718f-0cda-4524-aa3a-068bc2e7d569\") " pod="openshift-apiserver/apiserver-76f77b778f-jgx6f" Nov 28 13:21:18 crc kubenswrapper[4631]: I1128 13:21:18.994668 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7c7bd85-4add-4bda-a4fa-71a7b75787e4-serving-cert\") pod \"controller-manager-879f6c89f-7gmxj\" (UID: \"e7c7bd85-4add-4bda-a4fa-71a7b75787e4\") " pod="openshift-controller-manager/controller-manager-879f6c89f-7gmxj" Nov 28 13:21:18 crc kubenswrapper[4631]: I1128 13:21:18.994685 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/6947cfe2-0015-42bd-b68e-465e50d41f40-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-r5b6b\" (UID: \"6947cfe2-0015-42bd-b68e-465e50d41f40\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-r5b6b" Nov 28 13:21:18 crc kubenswrapper[4631]: I1128 13:21:18.994703 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/889c117d-ea46-4f02-a2e0-42a47c6e4683-images\") pod \"machine-api-operator-5694c8668f-2ch2b\" (UID: \"889c117d-ea46-4f02-a2e0-42a47c6e4683\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-2ch2b" Nov 28 13:21:18 crc kubenswrapper[4631]: I1128 13:21:18.994719 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/889c117d-ea46-4f02-a2e0-42a47c6e4683-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-2ch2b\" (UID: \"889c117d-ea46-4f02-a2e0-42a47c6e4683\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-2ch2b" Nov 28 13:21:18 crc kubenswrapper[4631]: I1128 13:21:18.994743 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mjnvx\" (UniqueName: \"kubernetes.io/projected/d1b540aa-a7ac-4f13-b430-91e1328d6602-kube-api-access-mjnvx\") pod \"authentication-operator-69f744f599-bf7bx\" (UID: \"d1b540aa-a7ac-4f13-b430-91e1328d6602\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-bf7bx" Nov 28 13:21:18 crc kubenswrapper[4631]: I1128 13:21:18.994760 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/5e56718f-0cda-4524-aa3a-068bc2e7d569-etcd-client\") pod \"apiserver-76f77b778f-jgx6f\" (UID: \"5e56718f-0cda-4524-aa3a-068bc2e7d569\") " pod="openshift-apiserver/apiserver-76f77b778f-jgx6f" Nov 28 13:21:18 crc kubenswrapper[4631]: I1128 13:21:18.994777 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tjfml\" (UniqueName: \"kubernetes.io/projected/feeda081-8bb0-4a89-ae6d-2e87f8f73344-kube-api-access-tjfml\") pod \"downloads-7954f5f757-c2rks\" (UID: \"feeda081-8bb0-4a89-ae6d-2e87f8f73344\") " pod="openshift-console/downloads-7954f5f757-c2rks" Nov 28 13:21:18 crc kubenswrapper[4631]: I1128 13:21:18.994792 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/5e56718f-0cda-4524-aa3a-068bc2e7d569-audit-dir\") pod \"apiserver-76f77b778f-jgx6f\" (UID: \"5e56718f-0cda-4524-aa3a-068bc2e7d569\") " pod="openshift-apiserver/apiserver-76f77b778f-jgx6f" Nov 28 13:21:18 crc kubenswrapper[4631]: I1128 13:21:18.994808 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/594fba15-6b93-4f34-af64-403a1b2d694c-config\") pod \"route-controller-manager-6576b87f9c-ddx6v\" (UID: \"594fba15-6b93-4f34-af64-403a1b2d694c\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ddx6v" Nov 28 13:21:18 crc kubenswrapper[4631]: I1128 13:21:18.995660 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Nov 28 13:21:18 crc kubenswrapper[4631]: I1128 13:21:18.995939 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Nov 28 13:21:18 crc kubenswrapper[4631]: I1128 13:21:18.996169 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Nov 28 13:21:18 crc kubenswrapper[4631]: I1128 13:21:18.997086 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Nov 28 13:21:18 crc kubenswrapper[4631]: I1128 13:21:18.998097 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-2ch2b"] Nov 28 13:21:18 crc kubenswrapper[4631]: I1128 13:21:18.998141 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-bf7bx"] Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.000230 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.000540 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.000767 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.001008 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.007871 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.009461 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-c2rks"] Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.010395 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.010697 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.010891 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.013072 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.013380 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.013477 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.013669 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.016162 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.016731 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.017171 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.018073 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.018570 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.018742 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.022771 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.023055 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.024013 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-7gmxj"] Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.028093 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.654676 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/cbd1163d-84ad-427f-b270-787520942822-available-featuregates\") pod \"openshift-config-operator-7777fb866f-rwpv9\" (UID: \"cbd1163d-84ad-427f-b270-787520942822\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-rwpv9" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.654748 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ccct6\" (UniqueName: \"kubernetes.io/projected/cbd1163d-84ad-427f-b270-787520942822-kube-api-access-ccct6\") pod \"openshift-config-operator-7777fb866f-rwpv9\" (UID: \"cbd1163d-84ad-427f-b270-787520942822\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-rwpv9" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.654787 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/aa813d64-b22e-40a0-b72a-b4e220e85f5a-etcd-client\") pod \"apiserver-7bbb656c7d-8tbnv\" (UID: \"aa813d64-b22e-40a0-b72a-b4e220e85f5a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8tbnv" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.654797 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.654830 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/5e56718f-0cda-4524-aa3a-068bc2e7d569-image-import-ca\") pod \"apiserver-76f77b778f-jgx6f\" (UID: \"5e56718f-0cda-4524-aa3a-068bc2e7d569\") " pod="openshift-apiserver/apiserver-76f77b778f-jgx6f" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.654867 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/aa813d64-b22e-40a0-b72a-b4e220e85f5a-audit-policies\") pod \"apiserver-7bbb656c7d-8tbnv\" (UID: \"aa813d64-b22e-40a0-b72a-b4e220e85f5a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8tbnv" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.654902 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e7c7bd85-4add-4bda-a4fa-71a7b75787e4-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-7gmxj\" (UID: \"e7c7bd85-4add-4bda-a4fa-71a7b75787e4\") " pod="openshift-controller-manager/controller-manager-879f6c89f-7gmxj" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.654951 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/145c4da2-bb50-4900-84d7-974b0cfaa4d8-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-hgsfk\" (UID: \"145c4da2-bb50-4900-84d7-974b0cfaa4d8\") " pod="openshift-authentication/oauth-openshift-558db77b4-hgsfk" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.654995 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5e56718f-0cda-4524-aa3a-068bc2e7d569-config\") pod \"apiserver-76f77b778f-jgx6f\" (UID: \"5e56718f-0cda-4524-aa3a-068bc2e7d569\") " pod="openshift-apiserver/apiserver-76f77b778f-jgx6f" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.655030 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r28b8\" (UniqueName: \"kubernetes.io/projected/5e56718f-0cda-4524-aa3a-068bc2e7d569-kube-api-access-r28b8\") pod \"apiserver-76f77b778f-jgx6f\" (UID: \"5e56718f-0cda-4524-aa3a-068bc2e7d569\") " pod="openshift-apiserver/apiserver-76f77b778f-jgx6f" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.655061 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/594fba15-6b93-4f34-af64-403a1b2d694c-client-ca\") pod \"route-controller-manager-6576b87f9c-ddx6v\" (UID: \"594fba15-6b93-4f34-af64-403a1b2d694c\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ddx6v" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.655096 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g6rlj\" (UniqueName: \"kubernetes.io/projected/6947cfe2-0015-42bd-b68e-465e50d41f40-kube-api-access-g6rlj\") pod \"cluster-image-registry-operator-dc59b4c8b-r5b6b\" (UID: \"6947cfe2-0015-42bd-b68e-465e50d41f40\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-r5b6b" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.655130 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/594fba15-6b93-4f34-af64-403a1b2d694c-serving-cert\") pod \"route-controller-manager-6576b87f9c-ddx6v\" (UID: \"594fba15-6b93-4f34-af64-403a1b2d694c\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ddx6v" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.655163 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/145c4da2-bb50-4900-84d7-974b0cfaa4d8-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-hgsfk\" (UID: \"145c4da2-bb50-4900-84d7-974b0cfaa4d8\") " pod="openshift-authentication/oauth-openshift-558db77b4-hgsfk" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.655200 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ad9ea570-0543-4731-bda7-9fdc4f3f0ef9-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-7977d\" (UID: \"ad9ea570-0543-4731-bda7-9fdc4f3f0ef9\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-7977d" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.655243 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/aa813d64-b22e-40a0-b72a-b4e220e85f5a-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-8tbnv\" (UID: \"aa813d64-b22e-40a0-b72a-b4e220e85f5a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8tbnv" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.655280 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/5e56718f-0cda-4524-aa3a-068bc2e7d569-etcd-serving-ca\") pod \"apiserver-76f77b778f-jgx6f\" (UID: \"5e56718f-0cda-4524-aa3a-068bc2e7d569\") " pod="openshift-apiserver/apiserver-76f77b778f-jgx6f" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.655339 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/5e56718f-0cda-4524-aa3a-068bc2e7d569-encryption-config\") pod \"apiserver-76f77b778f-jgx6f\" (UID: \"5e56718f-0cda-4524-aa3a-068bc2e7d569\") " pod="openshift-apiserver/apiserver-76f77b778f-jgx6f" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.655376 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/145c4da2-bb50-4900-84d7-974b0cfaa4d8-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-hgsfk\" (UID: \"145c4da2-bb50-4900-84d7-974b0cfaa4d8\") " pod="openshift-authentication/oauth-openshift-558db77b4-hgsfk" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.655432 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/6947cfe2-0015-42bd-b68e-465e50d41f40-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-r5b6b\" (UID: \"6947cfe2-0015-42bd-b68e-465e50d41f40\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-r5b6b" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.655465 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d1b540aa-a7ac-4f13-b430-91e1328d6602-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-bf7bx\" (UID: \"d1b540aa-a7ac-4f13-b430-91e1328d6602\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-bf7bx" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.655510 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7c7bd85-4add-4bda-a4fa-71a7b75787e4-config\") pod \"controller-manager-879f6c89f-7gmxj\" (UID: \"e7c7bd85-4add-4bda-a4fa-71a7b75787e4\") " pod="openshift-controller-manager/controller-manager-879f6c89f-7gmxj" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.655548 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/5e56718f-0cda-4524-aa3a-068bc2e7d569-audit\") pod \"apiserver-76f77b778f-jgx6f\" (UID: \"5e56718f-0cda-4524-aa3a-068bc2e7d569\") " pod="openshift-apiserver/apiserver-76f77b778f-jgx6f" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.655582 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5e56718f-0cda-4524-aa3a-068bc2e7d569-trusted-ca-bundle\") pod \"apiserver-76f77b778f-jgx6f\" (UID: \"5e56718f-0cda-4524-aa3a-068bc2e7d569\") " pod="openshift-apiserver/apiserver-76f77b778f-jgx6f" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.655619 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/aa813d64-b22e-40a0-b72a-b4e220e85f5a-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-8tbnv\" (UID: \"aa813d64-b22e-40a0-b72a-b4e220e85f5a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8tbnv" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.655655 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/145c4da2-bb50-4900-84d7-974b0cfaa4d8-audit-policies\") pod \"oauth-openshift-558db77b4-hgsfk\" (UID: \"145c4da2-bb50-4900-84d7-974b0cfaa4d8\") " pod="openshift-authentication/oauth-openshift-558db77b4-hgsfk" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.655697 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/6947cfe2-0015-42bd-b68e-465e50d41f40-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-r5b6b\" (UID: \"6947cfe2-0015-42bd-b68e-465e50d41f40\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-r5b6b" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.655728 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d1b540aa-a7ac-4f13-b430-91e1328d6602-serving-cert\") pod \"authentication-operator-69f744f599-bf7bx\" (UID: \"d1b540aa-a7ac-4f13-b430-91e1328d6602\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-bf7bx" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.655762 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mfgnq\" (UniqueName: \"kubernetes.io/projected/594fba15-6b93-4f34-af64-403a1b2d694c-kube-api-access-mfgnq\") pod \"route-controller-manager-6576b87f9c-ddx6v\" (UID: \"594fba15-6b93-4f34-af64-403a1b2d694c\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ddx6v" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.655796 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d8jfw\" (UniqueName: \"kubernetes.io/projected/e7c7bd85-4add-4bda-a4fa-71a7b75787e4-kube-api-access-d8jfw\") pod \"controller-manager-879f6c89f-7gmxj\" (UID: \"e7c7bd85-4add-4bda-a4fa-71a7b75787e4\") " pod="openshift-controller-manager/controller-manager-879f6c89f-7gmxj" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.655828 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/af9566e8-f912-41f1-b694-1d9d6d15e91f-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-vst8c\" (UID: \"af9566e8-f912-41f1-b694-1d9d6d15e91f\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-vst8c" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.655894 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cbd1163d-84ad-427f-b270-787520942822-serving-cert\") pod \"openshift-config-operator-7777fb866f-rwpv9\" (UID: \"cbd1163d-84ad-427f-b270-787520942822\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-rwpv9" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.655926 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e7c7bd85-4add-4bda-a4fa-71a7b75787e4-client-ca\") pod \"controller-manager-879f6c89f-7gmxj\" (UID: \"e7c7bd85-4add-4bda-a4fa-71a7b75787e4\") " pod="openshift-controller-manager/controller-manager-879f6c89f-7gmxj" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.655960 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j25br\" (UniqueName: \"kubernetes.io/projected/af9566e8-f912-41f1-b694-1d9d6d15e91f-kube-api-access-j25br\") pod \"openshift-apiserver-operator-796bbdcf4f-vst8c\" (UID: \"af9566e8-f912-41f1-b694-1d9d6d15e91f\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-vst8c" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.656001 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/5e56718f-0cda-4524-aa3a-068bc2e7d569-node-pullsecrets\") pod \"apiserver-76f77b778f-jgx6f\" (UID: \"5e56718f-0cda-4524-aa3a-068bc2e7d569\") " pod="openshift-apiserver/apiserver-76f77b778f-jgx6f" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.656033 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vmtq8\" (UniqueName: \"kubernetes.io/projected/889c117d-ea46-4f02-a2e0-42a47c6e4683-kube-api-access-vmtq8\") pod \"machine-api-operator-5694c8668f-2ch2b\" (UID: \"889c117d-ea46-4f02-a2e0-42a47c6e4683\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-2ch2b" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.656069 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d1b540aa-a7ac-4f13-b430-91e1328d6602-config\") pod \"authentication-operator-69f744f599-bf7bx\" (UID: \"d1b540aa-a7ac-4f13-b430-91e1328d6602\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-bf7bx" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.656101 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/145c4da2-bb50-4900-84d7-974b0cfaa4d8-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-hgsfk\" (UID: \"145c4da2-bb50-4900-84d7-974b0cfaa4d8\") " pod="openshift-authentication/oauth-openshift-558db77b4-hgsfk" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.656139 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/145c4da2-bb50-4900-84d7-974b0cfaa4d8-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-hgsfk\" (UID: \"145c4da2-bb50-4900-84d7-974b0cfaa4d8\") " pod="openshift-authentication/oauth-openshift-558db77b4-hgsfk" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.656173 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/145c4da2-bb50-4900-84d7-974b0cfaa4d8-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-hgsfk\" (UID: \"145c4da2-bb50-4900-84d7-974b0cfaa4d8\") " pod="openshift-authentication/oauth-openshift-558db77b4-hgsfk" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.656205 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/145c4da2-bb50-4900-84d7-974b0cfaa4d8-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-hgsfk\" (UID: \"145c4da2-bb50-4900-84d7-974b0cfaa4d8\") " pod="openshift-authentication/oauth-openshift-558db77b4-hgsfk" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.656240 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ad9ea570-0543-4731-bda7-9fdc4f3f0ef9-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-7977d\" (UID: \"ad9ea570-0543-4731-bda7-9fdc4f3f0ef9\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-7977d" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.656278 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d1b540aa-a7ac-4f13-b430-91e1328d6602-service-ca-bundle\") pod \"authentication-operator-69f744f599-bf7bx\" (UID: \"d1b540aa-a7ac-4f13-b430-91e1328d6602\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-bf7bx" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.656339 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/aa813d64-b22e-40a0-b72a-b4e220e85f5a-serving-cert\") pod \"apiserver-7bbb656c7d-8tbnv\" (UID: \"aa813d64-b22e-40a0-b72a-b4e220e85f5a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8tbnv" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.656378 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/889c117d-ea46-4f02-a2e0-42a47c6e4683-config\") pod \"machine-api-operator-5694c8668f-2ch2b\" (UID: \"889c117d-ea46-4f02-a2e0-42a47c6e4683\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-2ch2b" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.656410 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/145c4da2-bb50-4900-84d7-974b0cfaa4d8-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-hgsfk\" (UID: \"145c4da2-bb50-4900-84d7-974b0cfaa4d8\") " pod="openshift-authentication/oauth-openshift-558db77b4-hgsfk" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.656446 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8l2cr\" (UniqueName: \"kubernetes.io/projected/aa813d64-b22e-40a0-b72a-b4e220e85f5a-kube-api-access-8l2cr\") pod \"apiserver-7bbb656c7d-8tbnv\" (UID: \"aa813d64-b22e-40a0-b72a-b4e220e85f5a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8tbnv" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.656487 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/145c4da2-bb50-4900-84d7-974b0cfaa4d8-audit-dir\") pod \"oauth-openshift-558db77b4-hgsfk\" (UID: \"145c4da2-bb50-4900-84d7-974b0cfaa4d8\") " pod="openshift-authentication/oauth-openshift-558db77b4-hgsfk" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.656546 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/aa813d64-b22e-40a0-b72a-b4e220e85f5a-encryption-config\") pod \"apiserver-7bbb656c7d-8tbnv\" (UID: \"aa813d64-b22e-40a0-b72a-b4e220e85f5a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8tbnv" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.656582 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/aa813d64-b22e-40a0-b72a-b4e220e85f5a-audit-dir\") pod \"apiserver-7bbb656c7d-8tbnv\" (UID: \"aa813d64-b22e-40a0-b72a-b4e220e85f5a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8tbnv" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.656637 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5e56718f-0cda-4524-aa3a-068bc2e7d569-serving-cert\") pod \"apiserver-76f77b778f-jgx6f\" (UID: \"5e56718f-0cda-4524-aa3a-068bc2e7d569\") " pod="openshift-apiserver/apiserver-76f77b778f-jgx6f" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.656673 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7c7bd85-4add-4bda-a4fa-71a7b75787e4-serving-cert\") pod \"controller-manager-879f6c89f-7gmxj\" (UID: \"e7c7bd85-4add-4bda-a4fa-71a7b75787e4\") " pod="openshift-controller-manager/controller-manager-879f6c89f-7gmxj" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.656707 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/145c4da2-bb50-4900-84d7-974b0cfaa4d8-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-hgsfk\" (UID: \"145c4da2-bb50-4900-84d7-974b0cfaa4d8\") " pod="openshift-authentication/oauth-openshift-558db77b4-hgsfk" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.656732 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/5e56718f-0cda-4524-aa3a-068bc2e7d569-image-import-ca\") pod \"apiserver-76f77b778f-jgx6f\" (UID: \"5e56718f-0cda-4524-aa3a-068bc2e7d569\") " pod="openshift-apiserver/apiserver-76f77b778f-jgx6f" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.656738 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/af9566e8-f912-41f1-b694-1d9d6d15e91f-config\") pod \"openshift-apiserver-operator-796bbdcf4f-vst8c\" (UID: \"af9566e8-f912-41f1-b694-1d9d6d15e91f\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-vst8c" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.656834 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/6947cfe2-0015-42bd-b68e-465e50d41f40-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-r5b6b\" (UID: \"6947cfe2-0015-42bd-b68e-465e50d41f40\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-r5b6b" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.656865 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/889c117d-ea46-4f02-a2e0-42a47c6e4683-images\") pod \"machine-api-operator-5694c8668f-2ch2b\" (UID: \"889c117d-ea46-4f02-a2e0-42a47c6e4683\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-2ch2b" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.656896 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/889c117d-ea46-4f02-a2e0-42a47c6e4683-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-2ch2b\" (UID: \"889c117d-ea46-4f02-a2e0-42a47c6e4683\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-2ch2b" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.656944 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mjnvx\" (UniqueName: \"kubernetes.io/projected/d1b540aa-a7ac-4f13-b430-91e1328d6602-kube-api-access-mjnvx\") pod \"authentication-operator-69f744f599-bf7bx\" (UID: \"d1b540aa-a7ac-4f13-b430-91e1328d6602\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-bf7bx" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.656975 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/5e56718f-0cda-4524-aa3a-068bc2e7d569-etcd-client\") pod \"apiserver-76f77b778f-jgx6f\" (UID: \"5e56718f-0cda-4524-aa3a-068bc2e7d569\") " pod="openshift-apiserver/apiserver-76f77b778f-jgx6f" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.657003 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tjfml\" (UniqueName: \"kubernetes.io/projected/feeda081-8bb0-4a89-ae6d-2e87f8f73344-kube-api-access-tjfml\") pod \"downloads-7954f5f757-c2rks\" (UID: \"feeda081-8bb0-4a89-ae6d-2e87f8f73344\") " pod="openshift-console/downloads-7954f5f757-c2rks" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.657029 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/5e56718f-0cda-4524-aa3a-068bc2e7d569-audit-dir\") pod \"apiserver-76f77b778f-jgx6f\" (UID: \"5e56718f-0cda-4524-aa3a-068bc2e7d569\") " pod="openshift-apiserver/apiserver-76f77b778f-jgx6f" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.657055 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/594fba15-6b93-4f34-af64-403a1b2d694c-config\") pod \"route-controller-manager-6576b87f9c-ddx6v\" (UID: \"594fba15-6b93-4f34-af64-403a1b2d694c\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ddx6v" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.657083 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/145c4da2-bb50-4900-84d7-974b0cfaa4d8-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-hgsfk\" (UID: \"145c4da2-bb50-4900-84d7-974b0cfaa4d8\") " pod="openshift-authentication/oauth-openshift-558db77b4-hgsfk" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.657111 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9k5pd\" (UniqueName: \"kubernetes.io/projected/ad9ea570-0543-4731-bda7-9fdc4f3f0ef9-kube-api-access-9k5pd\") pod \"openshift-controller-manager-operator-756b6f6bc6-7977d\" (UID: \"ad9ea570-0543-4731-bda7-9fdc4f3f0ef9\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-7977d" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.657140 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/145c4da2-bb50-4900-84d7-974b0cfaa4d8-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-hgsfk\" (UID: \"145c4da2-bb50-4900-84d7-974b0cfaa4d8\") " pod="openshift-authentication/oauth-openshift-558db77b4-hgsfk" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.657164 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2l8sz\" (UniqueName: \"kubernetes.io/projected/145c4da2-bb50-4900-84d7-974b0cfaa4d8-kube-api-access-2l8sz\") pod \"oauth-openshift-558db77b4-hgsfk\" (UID: \"145c4da2-bb50-4900-84d7-974b0cfaa4d8\") " pod="openshift-authentication/oauth-openshift-558db77b4-hgsfk" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.657914 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/aa813d64-b22e-40a0-b72a-b4e220e85f5a-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-8tbnv\" (UID: \"aa813d64-b22e-40a0-b72a-b4e220e85f5a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8tbnv" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.658815 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/5e56718f-0cda-4524-aa3a-068bc2e7d569-etcd-serving-ca\") pod \"apiserver-76f77b778f-jgx6f\" (UID: \"5e56718f-0cda-4524-aa3a-068bc2e7d569\") " pod="openshift-apiserver/apiserver-76f77b778f-jgx6f" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.659015 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/aa813d64-b22e-40a0-b72a-b4e220e85f5a-audit-policies\") pod \"apiserver-7bbb656c7d-8tbnv\" (UID: \"aa813d64-b22e-40a0-b72a-b4e220e85f5a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8tbnv" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.661000 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d1b540aa-a7ac-4f13-b430-91e1328d6602-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-bf7bx\" (UID: \"d1b540aa-a7ac-4f13-b430-91e1328d6602\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-bf7bx" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.683602 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e7c7bd85-4add-4bda-a4fa-71a7b75787e4-client-ca\") pod \"controller-manager-879f6c89f-7gmxj\" (UID: \"e7c7bd85-4add-4bda-a4fa-71a7b75787e4\") " pod="openshift-controller-manager/controller-manager-879f6c89f-7gmxj" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.661752 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/889c117d-ea46-4f02-a2e0-42a47c6e4683-images\") pod \"machine-api-operator-5694c8668f-2ch2b\" (UID: \"889c117d-ea46-4f02-a2e0-42a47c6e4683\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-2ch2b" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.662025 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/889c117d-ea46-4f02-a2e0-42a47c6e4683-config\") pod \"machine-api-operator-5694c8668f-2ch2b\" (UID: \"889c117d-ea46-4f02-a2e0-42a47c6e4683\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-2ch2b" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.662694 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d1b540aa-a7ac-4f13-b430-91e1328d6602-config\") pod \"authentication-operator-69f744f599-bf7bx\" (UID: \"d1b540aa-a7ac-4f13-b430-91e1328d6602\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-bf7bx" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.663306 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d1b540aa-a7ac-4f13-b430-91e1328d6602-service-ca-bundle\") pod \"authentication-operator-69f744f599-bf7bx\" (UID: \"d1b540aa-a7ac-4f13-b430-91e1328d6602\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-bf7bx" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.664134 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/cbd1163d-84ad-427f-b270-787520942822-available-featuregates\") pod \"openshift-config-operator-7777fb866f-rwpv9\" (UID: \"cbd1163d-84ad-427f-b270-787520942822\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-rwpv9" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.669772 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/aa813d64-b22e-40a0-b72a-b4e220e85f5a-serving-cert\") pod \"apiserver-7bbb656c7d-8tbnv\" (UID: \"aa813d64-b22e-40a0-b72a-b4e220e85f5a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8tbnv" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.669755 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5e56718f-0cda-4524-aa3a-068bc2e7d569-config\") pod \"apiserver-76f77b778f-jgx6f\" (UID: \"5e56718f-0cda-4524-aa3a-068bc2e7d569\") " pod="openshift-apiserver/apiserver-76f77b778f-jgx6f" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.671330 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/594fba15-6b93-4f34-af64-403a1b2d694c-client-ca\") pod \"route-controller-manager-6576b87f9c-ddx6v\" (UID: \"594fba15-6b93-4f34-af64-403a1b2d694c\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ddx6v" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.673968 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/5e56718f-0cda-4524-aa3a-068bc2e7d569-encryption-config\") pod \"apiserver-76f77b778f-jgx6f\" (UID: \"5e56718f-0cda-4524-aa3a-068bc2e7d569\") " pod="openshift-apiserver/apiserver-76f77b778f-jgx6f" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.676074 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5e56718f-0cda-4524-aa3a-068bc2e7d569-serving-cert\") pod \"apiserver-76f77b778f-jgx6f\" (UID: \"5e56718f-0cda-4524-aa3a-068bc2e7d569\") " pod="openshift-apiserver/apiserver-76f77b778f-jgx6f" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.676124 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/aa813d64-b22e-40a0-b72a-b4e220e85f5a-audit-dir\") pod \"apiserver-7bbb656c7d-8tbnv\" (UID: \"aa813d64-b22e-40a0-b72a-b4e220e85f5a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8tbnv" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.676505 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/aa813d64-b22e-40a0-b72a-b4e220e85f5a-encryption-config\") pod \"apiserver-7bbb656c7d-8tbnv\" (UID: \"aa813d64-b22e-40a0-b72a-b4e220e85f5a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8tbnv" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.679516 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7c7bd85-4add-4bda-a4fa-71a7b75787e4-serving-cert\") pod \"controller-manager-879f6c89f-7gmxj\" (UID: \"e7c7bd85-4add-4bda-a4fa-71a7b75787e4\") " pod="openshift-controller-manager/controller-manager-879f6c89f-7gmxj" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.685344 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/5e56718f-0cda-4524-aa3a-068bc2e7d569-node-pullsecrets\") pod \"apiserver-76f77b778f-jgx6f\" (UID: \"5e56718f-0cda-4524-aa3a-068bc2e7d569\") " pod="openshift-apiserver/apiserver-76f77b778f-jgx6f" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.661092 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/aa813d64-b22e-40a0-b72a-b4e220e85f5a-etcd-client\") pod \"apiserver-7bbb656c7d-8tbnv\" (UID: \"aa813d64-b22e-40a0-b72a-b4e220e85f5a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8tbnv" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.686850 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/5e56718f-0cda-4524-aa3a-068bc2e7d569-audit-dir\") pod \"apiserver-76f77b778f-jgx6f\" (UID: \"5e56718f-0cda-4524-aa3a-068bc2e7d569\") " pod="openshift-apiserver/apiserver-76f77b778f-jgx6f" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.690442 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/5e56718f-0cda-4524-aa3a-068bc2e7d569-etcd-client\") pod \"apiserver-76f77b778f-jgx6f\" (UID: \"5e56718f-0cda-4524-aa3a-068bc2e7d569\") " pod="openshift-apiserver/apiserver-76f77b778f-jgx6f" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.692174 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5e56718f-0cda-4524-aa3a-068bc2e7d569-trusted-ca-bundle\") pod \"apiserver-76f77b778f-jgx6f\" (UID: \"5e56718f-0cda-4524-aa3a-068bc2e7d569\") " pod="openshift-apiserver/apiserver-76f77b778f-jgx6f" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.693549 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7c7bd85-4add-4bda-a4fa-71a7b75787e4-config\") pod \"controller-manager-879f6c89f-7gmxj\" (UID: \"e7c7bd85-4add-4bda-a4fa-71a7b75787e4\") " pod="openshift-controller-manager/controller-manager-879f6c89f-7gmxj" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.694097 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/5e56718f-0cda-4524-aa3a-068bc2e7d569-audit\") pod \"apiserver-76f77b778f-jgx6f\" (UID: \"5e56718f-0cda-4524-aa3a-068bc2e7d569\") " pod="openshift-apiserver/apiserver-76f77b778f-jgx6f" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.695206 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/594fba15-6b93-4f34-af64-403a1b2d694c-config\") pod \"route-controller-manager-6576b87f9c-ddx6v\" (UID: \"594fba15-6b93-4f34-af64-403a1b2d694c\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ddx6v" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.696055 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/aa813d64-b22e-40a0-b72a-b4e220e85f5a-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-8tbnv\" (UID: \"aa813d64-b22e-40a0-b72a-b4e220e85f5a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8tbnv" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.696344 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-qcd8n"] Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.696755 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.697030 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.697194 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.697376 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.701891 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.702140 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.702448 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.703990 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cbd1163d-84ad-427f-b270-787520942822-serving-cert\") pod \"openshift-config-operator-7777fb866f-rwpv9\" (UID: \"cbd1163d-84ad-427f-b270-787520942822\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-rwpv9" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.704110 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-ddx6v"] Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.705720 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-hgsfk"] Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.707191 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-659jg"] Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.708222 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/6947cfe2-0015-42bd-b68e-465e50d41f40-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-r5b6b\" (UID: \"6947cfe2-0015-42bd-b68e-465e50d41f40\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-r5b6b" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.708304 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/594fba15-6b93-4f34-af64-403a1b2d694c-serving-cert\") pod \"route-controller-manager-6576b87f9c-ddx6v\" (UID: \"594fba15-6b93-4f34-af64-403a1b2d694c\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ddx6v" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.709087 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.709173 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.709229 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-rwpv9"] Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.710505 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/6947cfe2-0015-42bd-b68e-465e50d41f40-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-r5b6b\" (UID: \"6947cfe2-0015-42bd-b68e-465e50d41f40\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-r5b6b" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.710750 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d1b540aa-a7ac-4f13-b430-91e1328d6602-serving-cert\") pod \"authentication-operator-69f744f599-bf7bx\" (UID: \"d1b540aa-a7ac-4f13-b430-91e1328d6602\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-bf7bx" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.710846 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.713107 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.714234 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e7c7bd85-4add-4bda-a4fa-71a7b75787e4-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-7gmxj\" (UID: \"e7c7bd85-4add-4bda-a4fa-71a7b75787e4\") " pod="openshift-controller-manager/controller-manager-879f6c89f-7gmxj" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.715894 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/889c117d-ea46-4f02-a2e0-42a47c6e4683-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-2ch2b\" (UID: \"889c117d-ea46-4f02-a2e0-42a47c6e4683\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-2ch2b" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.729187 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.730191 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-vst8c"] Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.730252 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-7977d"] Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.737167 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mjnvx\" (UniqueName: \"kubernetes.io/projected/d1b540aa-a7ac-4f13-b430-91e1328d6602-kube-api-access-mjnvx\") pod \"authentication-operator-69f744f599-bf7bx\" (UID: \"d1b540aa-a7ac-4f13-b430-91e1328d6602\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-bf7bx" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.737260 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-vhmzs"] Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.737278 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g6rlj\" (UniqueName: \"kubernetes.io/projected/6947cfe2-0015-42bd-b68e-465e50d41f40-kube-api-access-g6rlj\") pod \"cluster-image-registry-operator-dc59b4c8b-r5b6b\" (UID: \"6947cfe2-0015-42bd-b68e-465e50d41f40\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-r5b6b" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.737935 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vmtq8\" (UniqueName: \"kubernetes.io/projected/889c117d-ea46-4f02-a2e0-42a47c6e4683-kube-api-access-vmtq8\") pod \"machine-api-operator-5694c8668f-2ch2b\" (UID: \"889c117d-ea46-4f02-a2e0-42a47c6e4683\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-2ch2b" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.737979 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-52qtd"] Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.738162 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8l2cr\" (UniqueName: \"kubernetes.io/projected/aa813d64-b22e-40a0-b72a-b4e220e85f5a-kube-api-access-8l2cr\") pod \"apiserver-7bbb656c7d-8tbnv\" (UID: \"aa813d64-b22e-40a0-b72a-b4e220e85f5a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8tbnv" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.738547 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-52qtd" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.738787 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-xr5v6"] Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.739793 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-xr5v6" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.744357 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.744620 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.744641 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.744876 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.754049 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mfgnq\" (UniqueName: \"kubernetes.io/projected/594fba15-6b93-4f34-af64-403a1b2d694c-kube-api-access-mfgnq\") pod \"route-controller-manager-6576b87f9c-ddx6v\" (UID: \"594fba15-6b93-4f34-af64-403a1b2d694c\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ddx6v" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.755045 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.756300 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d8jfw\" (UniqueName: \"kubernetes.io/projected/e7c7bd85-4add-4bda-a4fa-71a7b75787e4-kube-api-access-d8jfw\") pod \"controller-manager-879f6c89f-7gmxj\" (UID: \"e7c7bd85-4add-4bda-a4fa-71a7b75787e4\") " pod="openshift-controller-manager/controller-manager-879f6c89f-7gmxj" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.757819 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.760879 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/145c4da2-bb50-4900-84d7-974b0cfaa4d8-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-hgsfk\" (UID: \"145c4da2-bb50-4900-84d7-974b0cfaa4d8\") " pod="openshift-authentication/oauth-openshift-558db77b4-hgsfk" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.761086 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/145c4da2-bb50-4900-84d7-974b0cfaa4d8-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-hgsfk\" (UID: \"145c4da2-bb50-4900-84d7-974b0cfaa4d8\") " pod="openshift-authentication/oauth-openshift-558db77b4-hgsfk" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.768540 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-r6sdp"] Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.770323 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ad9ea570-0543-4731-bda7-9fdc4f3f0ef9-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-7977d\" (UID: \"ad9ea570-0543-4731-bda7-9fdc4f3f0ef9\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-7977d" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.783107 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/6947cfe2-0015-42bd-b68e-465e50d41f40-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-r5b6b\" (UID: \"6947cfe2-0015-42bd-b68e-465e50d41f40\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-r5b6b" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.784939 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ad9ea570-0543-4731-bda7-9fdc4f3f0ef9-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-7977d\" (UID: \"ad9ea570-0543-4731-bda7-9fdc4f3f0ef9\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-7977d" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.786036 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/8e852fdc-5b82-4b51-b0e5-804a1b3be0ec-console-config\") pod \"console-f9d7485db-659jg\" (UID: \"8e852fdc-5b82-4b51-b0e5-804a1b3be0ec\") " pod="openshift-console/console-f9d7485db-659jg" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.786077 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b0690955-1897-45fe-94d0-f756117c7cb5-trusted-ca\") pod \"console-operator-58897d9998-qcd8n\" (UID: \"b0690955-1897-45fe-94d0-f756117c7cb5\") " pod="openshift-console-operator/console-operator-58897d9998-qcd8n" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.786099 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wrx26\" (UniqueName: \"kubernetes.io/projected/8e852fdc-5b82-4b51-b0e5-804a1b3be0ec-kube-api-access-wrx26\") pod \"console-f9d7485db-659jg\" (UID: \"8e852fdc-5b82-4b51-b0e5-804a1b3be0ec\") " pod="openshift-console/console-f9d7485db-659jg" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.786123 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/adf078af-a8b4-4988-89e9-7daa0a80a4bb-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-vhmzs\" (UID: \"adf078af-a8b4-4988-89e9-7daa0a80a4bb\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-vhmzs" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.786156 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/145c4da2-bb50-4900-84d7-974b0cfaa4d8-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-hgsfk\" (UID: \"145c4da2-bb50-4900-84d7-974b0cfaa4d8\") " pod="openshift-authentication/oauth-openshift-558db77b4-hgsfk" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.786203 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/145c4da2-bb50-4900-84d7-974b0cfaa4d8-audit-policies\") pod \"oauth-openshift-558db77b4-hgsfk\" (UID: \"145c4da2-bb50-4900-84d7-974b0cfaa4d8\") " pod="openshift-authentication/oauth-openshift-558db77b4-hgsfk" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.786228 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/84ff0842-5520-478f-ba3d-bc8f118ab6ef-auth-proxy-config\") pod \"machine-approver-56656f9798-xqbvh\" (UID: \"84ff0842-5520-478f-ba3d-bc8f118ab6ef\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-xqbvh" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.786260 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/af9566e8-f912-41f1-b694-1d9d6d15e91f-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-vst8c\" (UID: \"af9566e8-f912-41f1-b694-1d9d6d15e91f\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-vst8c" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.786279 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8e852fdc-5b82-4b51-b0e5-804a1b3be0ec-trusted-ca-bundle\") pod \"console-f9d7485db-659jg\" (UID: \"8e852fdc-5b82-4b51-b0e5-804a1b3be0ec\") " pod="openshift-console/console-f9d7485db-659jg" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.786332 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j25br\" (UniqueName: \"kubernetes.io/projected/af9566e8-f912-41f1-b694-1d9d6d15e91f-kube-api-access-j25br\") pod \"openshift-apiserver-operator-796bbdcf4f-vst8c\" (UID: \"af9566e8-f912-41f1-b694-1d9d6d15e91f\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-vst8c" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.786352 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/84ff0842-5520-478f-ba3d-bc8f118ab6ef-machine-approver-tls\") pod \"machine-approver-56656f9798-xqbvh\" (UID: \"84ff0842-5520-478f-ba3d-bc8f118ab6ef\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-xqbvh" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.786385 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/145c4da2-bb50-4900-84d7-974b0cfaa4d8-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-hgsfk\" (UID: \"145c4da2-bb50-4900-84d7-974b0cfaa4d8\") " pod="openshift-authentication/oauth-openshift-558db77b4-hgsfk" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.786403 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5v8wq\" (UniqueName: \"kubernetes.io/projected/adf078af-a8b4-4988-89e9-7daa0a80a4bb-kube-api-access-5v8wq\") pod \"cluster-samples-operator-665b6dd947-vhmzs\" (UID: \"adf078af-a8b4-4988-89e9-7daa0a80a4bb\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-vhmzs" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.786427 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/145c4da2-bb50-4900-84d7-974b0cfaa4d8-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-hgsfk\" (UID: \"145c4da2-bb50-4900-84d7-974b0cfaa4d8\") " pod="openshift-authentication/oauth-openshift-558db77b4-hgsfk" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.786480 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/145c4da2-bb50-4900-84d7-974b0cfaa4d8-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-hgsfk\" (UID: \"145c4da2-bb50-4900-84d7-974b0cfaa4d8\") " pod="openshift-authentication/oauth-openshift-558db77b4-hgsfk" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.786505 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/145c4da2-bb50-4900-84d7-974b0cfaa4d8-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-hgsfk\" (UID: \"145c4da2-bb50-4900-84d7-974b0cfaa4d8\") " pod="openshift-authentication/oauth-openshift-558db77b4-hgsfk" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.786547 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ad9ea570-0543-4731-bda7-9fdc4f3f0ef9-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-7977d\" (UID: \"ad9ea570-0543-4731-bda7-9fdc4f3f0ef9\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-7977d" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.786625 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8s5vw\" (UniqueName: \"kubernetes.io/projected/84ff0842-5520-478f-ba3d-bc8f118ab6ef-kube-api-access-8s5vw\") pod \"machine-approver-56656f9798-xqbvh\" (UID: \"84ff0842-5520-478f-ba3d-bc8f118ab6ef\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-xqbvh" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.786646 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b0690955-1897-45fe-94d0-f756117c7cb5-serving-cert\") pod \"console-operator-58897d9998-qcd8n\" (UID: \"b0690955-1897-45fe-94d0-f756117c7cb5\") " pod="openshift-console-operator/console-operator-58897d9998-qcd8n" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.786670 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/8e852fdc-5b82-4b51-b0e5-804a1b3be0ec-console-serving-cert\") pod \"console-f9d7485db-659jg\" (UID: \"8e852fdc-5b82-4b51-b0e5-804a1b3be0ec\") " pod="openshift-console/console-f9d7485db-659jg" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.786694 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/145c4da2-bb50-4900-84d7-974b0cfaa4d8-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-hgsfk\" (UID: \"145c4da2-bb50-4900-84d7-974b0cfaa4d8\") " pod="openshift-authentication/oauth-openshift-558db77b4-hgsfk" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.786718 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/145c4da2-bb50-4900-84d7-974b0cfaa4d8-audit-dir\") pod \"oauth-openshift-558db77b4-hgsfk\" (UID: \"145c4da2-bb50-4900-84d7-974b0cfaa4d8\") " pod="openshift-authentication/oauth-openshift-558db77b4-hgsfk" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.786732 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r28b8\" (UniqueName: \"kubernetes.io/projected/5e56718f-0cda-4524-aa3a-068bc2e7d569-kube-api-access-r28b8\") pod \"apiserver-76f77b778f-jgx6f\" (UID: \"5e56718f-0cda-4524-aa3a-068bc2e7d569\") " pod="openshift-apiserver/apiserver-76f77b778f-jgx6f" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.786770 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hpfgh\" (UniqueName: \"kubernetes.io/projected/b0690955-1897-45fe-94d0-f756117c7cb5-kube-api-access-hpfgh\") pod \"console-operator-58897d9998-qcd8n\" (UID: \"b0690955-1897-45fe-94d0-f756117c7cb5\") " pod="openshift-console-operator/console-operator-58897d9998-qcd8n" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.786836 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/145c4da2-bb50-4900-84d7-974b0cfaa4d8-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-hgsfk\" (UID: \"145c4da2-bb50-4900-84d7-974b0cfaa4d8\") " pod="openshift-authentication/oauth-openshift-558db77b4-hgsfk" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.786869 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/af9566e8-f912-41f1-b694-1d9d6d15e91f-config\") pod \"openshift-apiserver-operator-796bbdcf4f-vst8c\" (UID: \"af9566e8-f912-41f1-b694-1d9d6d15e91f\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-vst8c" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.786902 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/8e852fdc-5b82-4b51-b0e5-804a1b3be0ec-console-oauth-config\") pod \"console-f9d7485db-659jg\" (UID: \"8e852fdc-5b82-4b51-b0e5-804a1b3be0ec\") " pod="openshift-console/console-f9d7485db-659jg" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.786929 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/8e852fdc-5b82-4b51-b0e5-804a1b3be0ec-service-ca\") pod \"console-f9d7485db-659jg\" (UID: \"8e852fdc-5b82-4b51-b0e5-804a1b3be0ec\") " pod="openshift-console/console-f9d7485db-659jg" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.786949 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/84ff0842-5520-478f-ba3d-bc8f118ab6ef-config\") pod \"machine-approver-56656f9798-xqbvh\" (UID: \"84ff0842-5520-478f-ba3d-bc8f118ab6ef\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-xqbvh" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.786997 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/145c4da2-bb50-4900-84d7-974b0cfaa4d8-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-hgsfk\" (UID: \"145c4da2-bb50-4900-84d7-974b0cfaa4d8\") " pod="openshift-authentication/oauth-openshift-558db77b4-hgsfk" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.787018 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9k5pd\" (UniqueName: \"kubernetes.io/projected/ad9ea570-0543-4731-bda7-9fdc4f3f0ef9-kube-api-access-9k5pd\") pod \"openshift-controller-manager-operator-756b6f6bc6-7977d\" (UID: \"ad9ea570-0543-4731-bda7-9fdc4f3f0ef9\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-7977d" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.787038 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b0690955-1897-45fe-94d0-f756117c7cb5-config\") pod \"console-operator-58897d9998-qcd8n\" (UID: \"b0690955-1897-45fe-94d0-f756117c7cb5\") " pod="openshift-console-operator/console-operator-58897d9998-qcd8n" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.787079 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/145c4da2-bb50-4900-84d7-974b0cfaa4d8-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-hgsfk\" (UID: \"145c4da2-bb50-4900-84d7-974b0cfaa4d8\") " pod="openshift-authentication/oauth-openshift-558db77b4-hgsfk" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.787108 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2l8sz\" (UniqueName: \"kubernetes.io/projected/145c4da2-bb50-4900-84d7-974b0cfaa4d8-kube-api-access-2l8sz\") pod \"oauth-openshift-558db77b4-hgsfk\" (UID: \"145c4da2-bb50-4900-84d7-974b0cfaa4d8\") " pod="openshift-authentication/oauth-openshift-558db77b4-hgsfk" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.787126 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/8e852fdc-5b82-4b51-b0e5-804a1b3be0ec-oauth-serving-cert\") pod \"console-f9d7485db-659jg\" (UID: \"8e852fdc-5b82-4b51-b0e5-804a1b3be0ec\") " pod="openshift-console/console-f9d7485db-659jg" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.788199 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/af9566e8-f912-41f1-b694-1d9d6d15e91f-config\") pod \"openshift-apiserver-operator-796bbdcf4f-vst8c\" (UID: \"af9566e8-f912-41f1-b694-1d9d6d15e91f\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-vst8c" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.797519 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/145c4da2-bb50-4900-84d7-974b0cfaa4d8-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-hgsfk\" (UID: \"145c4da2-bb50-4900-84d7-974b0cfaa4d8\") " pod="openshift-authentication/oauth-openshift-558db77b4-hgsfk" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.799537 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/145c4da2-bb50-4900-84d7-974b0cfaa4d8-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-hgsfk\" (UID: \"145c4da2-bb50-4900-84d7-974b0cfaa4d8\") " pod="openshift-authentication/oauth-openshift-558db77b4-hgsfk" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.799564 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/145c4da2-bb50-4900-84d7-974b0cfaa4d8-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-hgsfk\" (UID: \"145c4da2-bb50-4900-84d7-974b0cfaa4d8\") " pod="openshift-authentication/oauth-openshift-558db77b4-hgsfk" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.800119 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/145c4da2-bb50-4900-84d7-974b0cfaa4d8-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-hgsfk\" (UID: \"145c4da2-bb50-4900-84d7-974b0cfaa4d8\") " pod="openshift-authentication/oauth-openshift-558db77b4-hgsfk" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.803088 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ad9ea570-0543-4731-bda7-9fdc4f3f0ef9-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-7977d\" (UID: \"ad9ea570-0543-4731-bda7-9fdc4f3f0ef9\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-7977d" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.803757 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/145c4da2-bb50-4900-84d7-974b0cfaa4d8-audit-dir\") pod \"oauth-openshift-558db77b4-hgsfk\" (UID: \"145c4da2-bb50-4900-84d7-974b0cfaa4d8\") " pod="openshift-authentication/oauth-openshift-558db77b4-hgsfk" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.810212 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8tbnv" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.810691 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/145c4da2-bb50-4900-84d7-974b0cfaa4d8-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-hgsfk\" (UID: \"145c4da2-bb50-4900-84d7-974b0cfaa4d8\") " pod="openshift-authentication/oauth-openshift-558db77b4-hgsfk" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.810882 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/145c4da2-bb50-4900-84d7-974b0cfaa4d8-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-hgsfk\" (UID: \"145c4da2-bb50-4900-84d7-974b0cfaa4d8\") " pod="openshift-authentication/oauth-openshift-558db77b4-hgsfk" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.810998 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/af9566e8-f912-41f1-b694-1d9d6d15e91f-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-vst8c\" (UID: \"af9566e8-f912-41f1-b694-1d9d6d15e91f\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-vst8c" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.811347 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/145c4da2-bb50-4900-84d7-974b0cfaa4d8-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-hgsfk\" (UID: \"145c4da2-bb50-4900-84d7-974b0cfaa4d8\") " pod="openshift-authentication/oauth-openshift-558db77b4-hgsfk" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.811508 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/145c4da2-bb50-4900-84d7-974b0cfaa4d8-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-hgsfk\" (UID: \"145c4da2-bb50-4900-84d7-974b0cfaa4d8\") " pod="openshift-authentication/oauth-openshift-558db77b4-hgsfk" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.812128 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-xltdb"] Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.814133 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-896hq"] Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.814431 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-mdws7"] Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.814836 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/145c4da2-bb50-4900-84d7-974b0cfaa4d8-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-hgsfk\" (UID: \"145c4da2-bb50-4900-84d7-974b0cfaa4d8\") " pod="openshift-authentication/oauth-openshift-558db77b4-hgsfk" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.814913 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-xltdb" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.815201 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9k5pd\" (UniqueName: \"kubernetes.io/projected/ad9ea570-0543-4731-bda7-9fdc4f3f0ef9-kube-api-access-9k5pd\") pod \"openshift-controller-manager-operator-756b6f6bc6-7977d\" (UID: \"ad9ea570-0543-4731-bda7-9fdc4f3f0ef9\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-7977d" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.815657 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-fx9mb"] Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.812553 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tjfml\" (UniqueName: \"kubernetes.io/projected/feeda081-8bb0-4a89-ae6d-2e87f8f73344-kube-api-access-tjfml\") pod \"downloads-7954f5f757-c2rks\" (UID: \"feeda081-8bb0-4a89-ae6d-2e87f8f73344\") " pod="openshift-console/downloads-7954f5f757-c2rks" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.812205 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-r6sdp" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.816030 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-896hq" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.816004 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-5v5xp"] Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.816408 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-mdws7" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.816477 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-fx9mb" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.816529 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-5v5xp" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.819272 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ddx6v" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.819363 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/145c4da2-bb50-4900-84d7-974b0cfaa4d8-audit-policies\") pod \"oauth-openshift-558db77b4-hgsfk\" (UID: \"145c4da2-bb50-4900-84d7-974b0cfaa4d8\") " pod="openshift-authentication/oauth-openshift-558db77b4-hgsfk" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.820097 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ccct6\" (UniqueName: \"kubernetes.io/projected/cbd1163d-84ad-427f-b270-787520942822-kube-api-access-ccct6\") pod \"openshift-config-operator-7777fb866f-rwpv9\" (UID: \"cbd1163d-84ad-427f-b270-787520942822\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-rwpv9" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.820181 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.821762 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-bf7bx" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.822231 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/145c4da2-bb50-4900-84d7-974b0cfaa4d8-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-hgsfk\" (UID: \"145c4da2-bb50-4900-84d7-974b0cfaa4d8\") " pod="openshift-authentication/oauth-openshift-558db77b4-hgsfk" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.822550 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.822598 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-rwpv9" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.822715 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.822897 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.823021 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.823083 4631 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.823654 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-jgx6f" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.823648 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2l8sz\" (UniqueName: \"kubernetes.io/projected/145c4da2-bb50-4900-84d7-974b0cfaa4d8-kube-api-access-2l8sz\") pod \"oauth-openshift-558db77b4-hgsfk\" (UID: \"145c4da2-bb50-4900-84d7-974b0cfaa4d8\") " pod="openshift-authentication/oauth-openshift-558db77b4-hgsfk" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.825709 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.827802 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.828104 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.829601 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.829817 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.829997 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.830107 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.830204 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.832079 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j25br\" (UniqueName: \"kubernetes.io/projected/af9566e8-f912-41f1-b694-1d9d6d15e91f-kube-api-access-j25br\") pod \"openshift-apiserver-operator-796bbdcf4f-vst8c\" (UID: \"af9566e8-f912-41f1-b694-1d9d6d15e91f\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-vst8c" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.833491 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/145c4da2-bb50-4900-84d7-974b0cfaa4d8-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-hgsfk\" (UID: \"145c4da2-bb50-4900-84d7-974b0cfaa4d8\") " pod="openshift-authentication/oauth-openshift-558db77b4-hgsfk" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.833792 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.835015 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-qdxf6"] Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.835921 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-qdxf6" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.836215 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-7gmxj" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.836023 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-jcnpq"] Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.837503 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29405595-2v6fk"] Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.838870 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bzj6n"] Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.837728 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-jcnpq" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.839058 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29405595-2v6fk" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.840042 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-rsxk2"] Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.840434 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-cg4zv"] Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.840364 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bzj6n" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.841040 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-82qbh"] Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.841128 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-cg4zv" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.841626 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-82qbh" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.842032 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-dxzdh"] Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.842586 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-dxzdh" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.843593 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.843879 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-rsxk2" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.845006 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-658pf"] Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.845808 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-658pf" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.847019 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-2v5d6"] Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.847721 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-2v5d6" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.854091 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-977k9"] Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.854924 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.854976 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/cni-sysctl-allowlist-ds-44pwx"] Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.855618 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/cni-sysctl-allowlist-ds-44pwx" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.855032 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-977k9" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.856968 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-5h9cj"] Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.857592 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-5h9cj" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.859256 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-g8n9t"] Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.859946 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-g8n9t" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.861487 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-2bc4b"] Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.861941 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-2bc4b" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.872807 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-hgsfk" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.873681 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-hzvzz"] Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.877925 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-s7zwj"] Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.879674 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-hzvzz" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.883167 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.883424 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-jqnpq"] Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.884268 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-s7zwj" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.884631 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-gk97s"] Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.886132 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-xr5v6"] Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.886357 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-c2rks" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.888323 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ksp4p\" (UniqueName: \"kubernetes.io/projected/062b3680-8e44-4ab0-8f9b-a319169b53f1-kube-api-access-ksp4p\") pod \"ingress-canary-xr5v6\" (UID: \"062b3680-8e44-4ab0-8f9b-a319169b53f1\") " pod="openshift-ingress-canary/ingress-canary-xr5v6" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.888384 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hpfgh\" (UniqueName: \"kubernetes.io/projected/b0690955-1897-45fe-94d0-f756117c7cb5-kube-api-access-hpfgh\") pod \"console-operator-58897d9998-qcd8n\" (UID: \"b0690955-1897-45fe-94d0-f756117c7cb5\") " pod="openshift-console-operator/console-operator-58897d9998-qcd8n" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.888431 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/8e852fdc-5b82-4b51-b0e5-804a1b3be0ec-console-oauth-config\") pod \"console-f9d7485db-659jg\" (UID: \"8e852fdc-5b82-4b51-b0e5-804a1b3be0ec\") " pod="openshift-console/console-f9d7485db-659jg" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.888479 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/8e852fdc-5b82-4b51-b0e5-804a1b3be0ec-service-ca\") pod \"console-f9d7485db-659jg\" (UID: \"8e852fdc-5b82-4b51-b0e5-804a1b3be0ec\") " pod="openshift-console/console-f9d7485db-659jg" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.888514 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/84ff0842-5520-478f-ba3d-bc8f118ab6ef-config\") pod \"machine-approver-56656f9798-xqbvh\" (UID: \"84ff0842-5520-478f-ba3d-bc8f118ab6ef\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-xqbvh" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.888556 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b0690955-1897-45fe-94d0-f756117c7cb5-config\") pod \"console-operator-58897d9998-qcd8n\" (UID: \"b0690955-1897-45fe-94d0-f756117c7cb5\") " pod="openshift-console-operator/console-operator-58897d9998-qcd8n" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.888597 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/8e852fdc-5b82-4b51-b0e5-804a1b3be0ec-oauth-serving-cert\") pod \"console-f9d7485db-659jg\" (UID: \"8e852fdc-5b82-4b51-b0e5-804a1b3be0ec\") " pod="openshift-console/console-f9d7485db-659jg" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.888646 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/8e852fdc-5b82-4b51-b0e5-804a1b3be0ec-console-config\") pod \"console-f9d7485db-659jg\" (UID: \"8e852fdc-5b82-4b51-b0e5-804a1b3be0ec\") " pod="openshift-console/console-f9d7485db-659jg" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.888688 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b0690955-1897-45fe-94d0-f756117c7cb5-trusted-ca\") pod \"console-operator-58897d9998-qcd8n\" (UID: \"b0690955-1897-45fe-94d0-f756117c7cb5\") " pod="openshift-console-operator/console-operator-58897d9998-qcd8n" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.888750 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wrx26\" (UniqueName: \"kubernetes.io/projected/8e852fdc-5b82-4b51-b0e5-804a1b3be0ec-kube-api-access-wrx26\") pod \"console-f9d7485db-659jg\" (UID: \"8e852fdc-5b82-4b51-b0e5-804a1b3be0ec\") " pod="openshift-console/console-f9d7485db-659jg" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.888813 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/adf078af-a8b4-4988-89e9-7daa0a80a4bb-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-vhmzs\" (UID: \"adf078af-a8b4-4988-89e9-7daa0a80a4bb\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-vhmzs" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.888881 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/062b3680-8e44-4ab0-8f9b-a319169b53f1-cert\") pod \"ingress-canary-xr5v6\" (UID: \"062b3680-8e44-4ab0-8f9b-a319169b53f1\") " pod="openshift-ingress-canary/ingress-canary-xr5v6" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.888917 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/f8a4f24e-7ccd-4527-9020-2f101c1d4cb5-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-52qtd\" (UID: \"f8a4f24e-7ccd-4527-9020-2f101c1d4cb5\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-52qtd" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.888996 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/84ff0842-5520-478f-ba3d-bc8f118ab6ef-auth-proxy-config\") pod \"machine-approver-56656f9798-xqbvh\" (UID: \"84ff0842-5520-478f-ba3d-bc8f118ab6ef\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-xqbvh" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.889072 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8e852fdc-5b82-4b51-b0e5-804a1b3be0ec-trusted-ca-bundle\") pod \"console-f9d7485db-659jg\" (UID: \"8e852fdc-5b82-4b51-b0e5-804a1b3be0ec\") " pod="openshift-console/console-f9d7485db-659jg" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.889120 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tgg82\" (UniqueName: \"kubernetes.io/projected/f8a4f24e-7ccd-4527-9020-2f101c1d4cb5-kube-api-access-tgg82\") pod \"multus-admission-controller-857f4d67dd-52qtd\" (UID: \"f8a4f24e-7ccd-4527-9020-2f101c1d4cb5\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-52qtd" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.889153 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/84ff0842-5520-478f-ba3d-bc8f118ab6ef-machine-approver-tls\") pod \"machine-approver-56656f9798-xqbvh\" (UID: \"84ff0842-5520-478f-ba3d-bc8f118ab6ef\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-xqbvh" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.889198 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5v8wq\" (UniqueName: \"kubernetes.io/projected/adf078af-a8b4-4988-89e9-7daa0a80a4bb-kube-api-access-5v8wq\") pod \"cluster-samples-operator-665b6dd947-vhmzs\" (UID: \"adf078af-a8b4-4988-89e9-7daa0a80a4bb\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-vhmzs" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.889225 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b0690955-1897-45fe-94d0-f756117c7cb5-serving-cert\") pod \"console-operator-58897d9998-qcd8n\" (UID: \"b0690955-1897-45fe-94d0-f756117c7cb5\") " pod="openshift-console-operator/console-operator-58897d9998-qcd8n" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.889249 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8s5vw\" (UniqueName: \"kubernetes.io/projected/84ff0842-5520-478f-ba3d-bc8f118ab6ef-kube-api-access-8s5vw\") pod \"machine-approver-56656f9798-xqbvh\" (UID: \"84ff0842-5520-478f-ba3d-bc8f118ab6ef\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-xqbvh" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.889304 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/8e852fdc-5b82-4b51-b0e5-804a1b3be0ec-console-serving-cert\") pod \"console-f9d7485db-659jg\" (UID: \"8e852fdc-5b82-4b51-b0e5-804a1b3be0ec\") " pod="openshift-console/console-f9d7485db-659jg" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.889815 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-gk97s" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.891268 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-jqnpq" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.892450 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/8e852fdc-5b82-4b51-b0e5-804a1b3be0ec-oauth-serving-cert\") pod \"console-f9d7485db-659jg\" (UID: \"8e852fdc-5b82-4b51-b0e5-804a1b3be0ec\") " pod="openshift-console/console-f9d7485db-659jg" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.892715 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-xltdb"] Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.895941 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/8e852fdc-5b82-4b51-b0e5-804a1b3be0ec-console-config\") pod \"console-f9d7485db-659jg\" (UID: \"8e852fdc-5b82-4b51-b0e5-804a1b3be0ec\") " pod="openshift-console/console-f9d7485db-659jg" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.897573 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b0690955-1897-45fe-94d0-f756117c7cb5-config\") pod \"console-operator-58897d9998-qcd8n\" (UID: \"b0690955-1897-45fe-94d0-f756117c7cb5\") " pod="openshift-console-operator/console-operator-58897d9998-qcd8n" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.898405 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/84ff0842-5520-478f-ba3d-bc8f118ab6ef-auth-proxy-config\") pod \"machine-approver-56656f9798-xqbvh\" (UID: \"84ff0842-5520-478f-ba3d-bc8f118ab6ef\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-xqbvh" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.902343 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8e852fdc-5b82-4b51-b0e5-804a1b3be0ec-trusted-ca-bundle\") pod \"console-f9d7485db-659jg\" (UID: \"8e852fdc-5b82-4b51-b0e5-804a1b3be0ec\") " pod="openshift-console/console-f9d7485db-659jg" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.902424 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-52qtd"] Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.902653 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/8e852fdc-5b82-4b51-b0e5-804a1b3be0ec-service-ca\") pod \"console-f9d7485db-659jg\" (UID: \"8e852fdc-5b82-4b51-b0e5-804a1b3be0ec\") " pod="openshift-console/console-f9d7485db-659jg" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.902745 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.902994 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-vst8c" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.903848 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/84ff0842-5520-478f-ba3d-bc8f118ab6ef-config\") pod \"machine-approver-56656f9798-xqbvh\" (UID: \"84ff0842-5520-478f-ba3d-bc8f118ab6ef\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-xqbvh" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.904009 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/8e852fdc-5b82-4b51-b0e5-804a1b3be0ec-console-serving-cert\") pod \"console-f9d7485db-659jg\" (UID: \"8e852fdc-5b82-4b51-b0e5-804a1b3be0ec\") " pod="openshift-console/console-f9d7485db-659jg" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.905846 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b0690955-1897-45fe-94d0-f756117c7cb5-trusted-ca\") pod \"console-operator-58897d9998-qcd8n\" (UID: \"b0690955-1897-45fe-94d0-f756117c7cb5\") " pod="openshift-console-operator/console-operator-58897d9998-qcd8n" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.906621 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-r5b6b" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.907133 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/8e852fdc-5b82-4b51-b0e5-804a1b3be0ec-console-oauth-config\") pod \"console-f9d7485db-659jg\" (UID: \"8e852fdc-5b82-4b51-b0e5-804a1b3be0ec\") " pod="openshift-console/console-f9d7485db-659jg" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.907447 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-r6sdp"] Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.909340 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-896hq"] Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.910904 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b0690955-1897-45fe-94d0-f756117c7cb5-serving-cert\") pod \"console-operator-58897d9998-qcd8n\" (UID: \"b0690955-1897-45fe-94d0-f756117c7cb5\") " pod="openshift-console-operator/console-operator-58897d9998-qcd8n" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.911492 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.912358 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-fx9mb"] Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.913746 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/84ff0842-5520-478f-ba3d-bc8f118ab6ef-machine-approver-tls\") pod \"machine-approver-56656f9798-xqbvh\" (UID: \"84ff0842-5520-478f-ba3d-bc8f118ab6ef\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-xqbvh" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.920677 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-mdws7"] Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.923171 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-qdxf6"] Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.923232 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-2v5d6"] Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.924318 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-cg4zv"] Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.928465 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bzj6n"] Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.929867 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-82qbh"] Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.930558 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-977k9"] Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.931657 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-jqnpq"] Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.932075 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/adf078af-a8b4-4988-89e9-7daa0a80a4bb-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-vhmzs\" (UID: \"adf078af-a8b4-4988-89e9-7daa0a80a4bb\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-vhmzs" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.939320 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-2ch2b" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.947677 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-658pf"] Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.948119 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-dxzdh"] Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.949258 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-hzvzz"] Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.952605 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.953571 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29405595-2v6fk"] Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.955150 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-g8n9t"] Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.957245 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-5h9cj"] Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.998988 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-5v5xp"] Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.995794 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-7977d" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.997586 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.999231 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-gk97s"] Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.995826 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/2eddcce3-58a6-45f9-b924-30d95fa897ae-proxy-tls\") pod \"machine-config-operator-74547568cd-5h9cj\" (UID: \"2eddcce3-58a6-45f9-b924-30d95fa897ae\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-5h9cj" Nov 28 13:21:19 crc kubenswrapper[4631]: I1128 13:21:19.999936 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6fccb501-125b-4bc4-8cad-3349f332c2de-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-gk97s\" (UID: \"6fccb501-125b-4bc4-8cad-3349f332c2de\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-gk97s" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:19.999971 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/2a2f7727-f05e-4deb-8eba-cb3ec3561621-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-2v5d6\" (UID: \"2a2f7727-f05e-4deb-8eba-cb3ec3561621\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-2v5d6" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.000007 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5stxb\" (UniqueName: \"kubernetes.io/projected/73b80803-714a-446e-9df2-6e3ccf5430ee-kube-api-access-5stxb\") pod \"machine-config-controller-84d6567774-dxzdh\" (UID: \"73b80803-714a-446e-9df2-6e3ccf5430ee\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-dxzdh" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.000061 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qfstd\" (UniqueName: \"kubernetes.io/projected/0e2f0bc1-2261-484e-8f4b-34bddacda1bc-kube-api-access-qfstd\") pod \"catalog-operator-68c6474976-hzvzz\" (UID: \"0e2f0bc1-2261-484e-8f4b-34bddacda1bc\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-hzvzz" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.000083 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/4415ae1d-dbf4-450b-a329-6dc64b0a019f-signing-cabundle\") pod \"service-ca-9c57cc56f-xltdb\" (UID: \"4415ae1d-dbf4-450b-a329-6dc64b0a019f\") " pod="openshift-service-ca/service-ca-9c57cc56f-xltdb" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.000113 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/0e2f0bc1-2261-484e-8f4b-34bddacda1bc-srv-cert\") pod \"catalog-operator-68c6474976-hzvzz\" (UID: \"0e2f0bc1-2261-484e-8f4b-34bddacda1bc\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-hzvzz" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.000167 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/0e2f0bc1-2261-484e-8f4b-34bddacda1bc-profile-collector-cert\") pod \"catalog-operator-68c6474976-hzvzz\" (UID: \"0e2f0bc1-2261-484e-8f4b-34bddacda1bc\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-hzvzz" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.000202 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/1039cddf-be22-4ca1-a622-d21967f9d859-profile-collector-cert\") pod \"olm-operator-6b444d44fb-bzj6n\" (UID: \"1039cddf-be22-4ca1-a622-d21967f9d859\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bzj6n" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.000226 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d5d77be3-1836-4c8b-9fc4-ab34196461c5-serving-cert\") pod \"service-ca-operator-777779d784-qdxf6\" (UID: \"d5d77be3-1836-4c8b-9fc4-ab34196461c5\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-qdxf6" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.000484 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/6fe798f2-84e0-4019-a75c-0ee2f2e672f5-metrics-tls\") pod \"dns-operator-744455d44c-82qbh\" (UID: \"6fe798f2-84e0-4019-a75c-0ee2f2e672f5\") " pod="openshift-dns-operator/dns-operator-744455d44c-82qbh" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.000524 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/760ce379-2153-4d72-acb7-f3f3f08a9028-apiservice-cert\") pod \"packageserver-d55dfcdfc-896hq\" (UID: \"760ce379-2153-4d72-acb7-f3f3f08a9028\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-896hq" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.000571 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/f8a4f24e-7ccd-4527-9020-2f101c1d4cb5-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-52qtd\" (UID: \"f8a4f24e-7ccd-4527-9020-2f101c1d4cb5\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-52qtd" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.000591 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/0709d8d6-5c00-4ec2-8651-df2cd64fe419-certs\") pod \"machine-config-server-rsxk2\" (UID: \"0709d8d6-5c00-4ec2-8651-df2cd64fe419\") " pod="openshift-machine-config-operator/machine-config-server-rsxk2" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.000634 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/9bc5a0ed-1419-4ba0-8478-9bcad6a19954-registration-dir\") pod \"csi-hostpathplugin-r6sdp\" (UID: \"9bc5a0ed-1419-4ba0-8478-9bcad6a19954\") " pod="hostpath-provisioner/csi-hostpathplugin-r6sdp" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.000723 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sssnb\" (UniqueName: \"kubernetes.io/projected/1039cddf-be22-4ca1-a622-d21967f9d859-kube-api-access-sssnb\") pod \"olm-operator-6b444d44fb-bzj6n\" (UID: \"1039cddf-be22-4ca1-a622-d21967f9d859\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bzj6n" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.000767 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/9eb418ce-65b6-4da3-b282-831dd62916cf-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-2bc4b\" (UID: \"9eb418ce-65b6-4da3-b282-831dd62916cf\") " pod="openshift-marketplace/marketplace-operator-79b997595-2bc4b" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.000789 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ready\" (UniqueName: \"kubernetes.io/empty-dir/9504d491-6c27-4387-b6da-fa0192cc7a05-ready\") pod \"cni-sysctl-allowlist-ds-44pwx\" (UID: \"9504d491-6c27-4387-b6da-fa0192cc7a05\") " pod="openshift-multus/cni-sysctl-allowlist-ds-44pwx" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.000815 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/5db3c754-f675-4ec8-b3aa-65d0528bb9ca-bound-sa-token\") pod \"ingress-operator-5b745b69d9-g8n9t\" (UID: \"5db3c754-f675-4ec8-b3aa-65d0528bb9ca\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-g8n9t" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.000837 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bgtn5\" (UniqueName: \"kubernetes.io/projected/9eb418ce-65b6-4da3-b282-831dd62916cf-kube-api-access-bgtn5\") pod \"marketplace-operator-79b997595-2bc4b\" (UID: \"9eb418ce-65b6-4da3-b282-831dd62916cf\") " pod="openshift-marketplace/marketplace-operator-79b997595-2bc4b" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.000855 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/9504d491-6c27-4387-b6da-fa0192cc7a05-cni-sysctl-allowlist\") pod \"cni-sysctl-allowlist-ds-44pwx\" (UID: \"9504d491-6c27-4387-b6da-fa0192cc7a05\") " pod="openshift-multus/cni-sysctl-allowlist-ds-44pwx" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.000896 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/089e1482-6475-40b3-a005-c151f4bb0a87-config-volume\") pod \"dns-default-cg4zv\" (UID: \"089e1482-6475-40b3-a005-c151f4bb0a87\") " pod="openshift-dns/dns-default-cg4zv" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.000912 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/089e1482-6475-40b3-a005-c151f4bb0a87-metrics-tls\") pod \"dns-default-cg4zv\" (UID: \"089e1482-6475-40b3-a005-c151f4bb0a87\") " pod="openshift-dns/dns-default-cg4zv" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.000934 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6b363da7-b1c6-495d-a05b-7ff4d7951b3e-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-jqnpq\" (UID: \"6b363da7-b1c6-495d-a05b-7ff4d7951b3e\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-jqnpq" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.000975 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c7sdh\" (UniqueName: \"kubernetes.io/projected/9bc5a0ed-1419-4ba0-8478-9bcad6a19954-kube-api-access-c7sdh\") pod \"csi-hostpathplugin-r6sdp\" (UID: \"9bc5a0ed-1419-4ba0-8478-9bcad6a19954\") " pod="hostpath-provisioner/csi-hostpathplugin-r6sdp" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.001003 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d5d77be3-1836-4c8b-9fc4-ab34196461c5-config\") pod \"service-ca-operator-777779d784-qdxf6\" (UID: \"d5d77be3-1836-4c8b-9fc4-ab34196461c5\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-qdxf6" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.001023 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-flzfb\" (UniqueName: \"kubernetes.io/projected/2eddcce3-58a6-45f9-b924-30d95fa897ae-kube-api-access-flzfb\") pod \"machine-config-operator-74547568cd-5h9cj\" (UID: \"2eddcce3-58a6-45f9-b924-30d95fa897ae\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-5h9cj" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:19.997642 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.001045 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5db3c754-f675-4ec8-b3aa-65d0528bb9ca-trusted-ca\") pod \"ingress-operator-5b745b69d9-g8n9t\" (UID: \"5db3c754-f675-4ec8-b3aa-65d0528bb9ca\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-g8n9t" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.001069 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l8rxn\" (UniqueName: \"kubernetes.io/projected/5db3c754-f675-4ec8-b3aa-65d0528bb9ca-kube-api-access-l8rxn\") pod \"ingress-operator-5b745b69d9-g8n9t\" (UID: \"5db3c754-f675-4ec8-b3aa-65d0528bb9ca\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-g8n9t" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.001098 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ksp4p\" (UniqueName: \"kubernetes.io/projected/062b3680-8e44-4ab0-8f9b-a319169b53f1-kube-api-access-ksp4p\") pod \"ingress-canary-xr5v6\" (UID: \"062b3680-8e44-4ab0-8f9b-a319169b53f1\") " pod="openshift-ingress-canary/ingress-canary-xr5v6" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.001126 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b4cst\" (UniqueName: \"kubernetes.io/projected/6fe798f2-84e0-4019-a75c-0ee2f2e672f5-kube-api-access-b4cst\") pod \"dns-operator-744455d44c-82qbh\" (UID: \"6fe798f2-84e0-4019-a75c-0ee2f2e672f5\") " pod="openshift-dns-operator/dns-operator-744455d44c-82qbh" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.001180 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/9bc5a0ed-1419-4ba0-8478-9bcad6a19954-plugins-dir\") pod \"csi-hostpathplugin-r6sdp\" (UID: \"9bc5a0ed-1419-4ba0-8478-9bcad6a19954\") " pod="hostpath-provisioner/csi-hostpathplugin-r6sdp" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.001201 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5pq68\" (UniqueName: \"kubernetes.io/projected/f2139e10-e6bf-4193-b242-7e061396f034-kube-api-access-5pq68\") pod \"migrator-59844c95c7-977k9\" (UID: \"f2139e10-e6bf-4193-b242-7e061396f034\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-977k9" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.001217 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1a4f8da7-c099-4a8e-bffc-ca17ca74eb42-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-658pf\" (UID: \"1a4f8da7-c099-4a8e-bffc-ca17ca74eb42\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-658pf" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.001233 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/9bc5a0ed-1419-4ba0-8478-9bcad6a19954-mountpoint-dir\") pod \"csi-hostpathplugin-r6sdp\" (UID: \"9bc5a0ed-1419-4ba0-8478-9bcad6a19954\") " pod="hostpath-provisioner/csi-hostpathplugin-r6sdp" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.001384 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/22010e1d-3ec1-4564-a362-bc963f691f1f-stats-auth\") pod \"router-default-5444994796-jcnpq\" (UID: \"22010e1d-3ec1-4564-a362-bc963f691f1f\") " pod="openshift-ingress/router-default-5444994796-jcnpq" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.001438 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/2eddcce3-58a6-45f9-b924-30d95fa897ae-images\") pod \"machine-config-operator-74547568cd-5h9cj\" (UID: \"2eddcce3-58a6-45f9-b924-30d95fa897ae\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-5h9cj" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.001476 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/76a05ca6-a98c-4313-855c-fd6ffaceee4f-config-volume\") pod \"collect-profiles-29405595-2v6fk\" (UID: \"76a05ca6-a98c-4313-855c-fd6ffaceee4f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405595-2v6fk" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.001518 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4lkgw\" (UniqueName: \"kubernetes.io/projected/089e1482-6475-40b3-a005-c151f4bb0a87-kube-api-access-4lkgw\") pod \"dns-default-cg4zv\" (UID: \"089e1482-6475-40b3-a005-c151f4bb0a87\") " pod="openshift-dns/dns-default-cg4zv" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.001539 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/9bc5a0ed-1419-4ba0-8478-9bcad6a19954-csi-data-dir\") pod \"csi-hostpathplugin-r6sdp\" (UID: \"9bc5a0ed-1419-4ba0-8478-9bcad6a19954\") " pod="hostpath-provisioner/csi-hostpathplugin-r6sdp" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.001554 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6b363da7-b1c6-495d-a05b-7ff4d7951b3e-config\") pod \"kube-apiserver-operator-766d6c64bb-jqnpq\" (UID: \"6b363da7-b1c6-495d-a05b-7ff4d7951b3e\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-jqnpq" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.001573 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/73b80803-714a-446e-9df2-6e3ccf5430ee-proxy-tls\") pod \"machine-config-controller-84d6567774-dxzdh\" (UID: \"73b80803-714a-446e-9df2-6e3ccf5430ee\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-dxzdh" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.001599 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/51595599-be30-46b6-bab6-ee647410efbc-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-fx9mb\" (UID: \"51595599-be30-46b6-bab6-ee647410efbc\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-fx9mb" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.001634 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b54w6\" (UniqueName: \"kubernetes.io/projected/76a05ca6-a98c-4313-855c-fd6ffaceee4f-kube-api-access-b54w6\") pod \"collect-profiles-29405595-2v6fk\" (UID: \"76a05ca6-a98c-4313-855c-fd6ffaceee4f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405595-2v6fk" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.001703 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gw52s\" (UniqueName: \"kubernetes.io/projected/22010e1d-3ec1-4564-a362-bc963f691f1f-kube-api-access-gw52s\") pod \"router-default-5444994796-jcnpq\" (UID: \"22010e1d-3ec1-4564-a362-bc963f691f1f\") " pod="openshift-ingress/router-default-5444994796-jcnpq" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.001722 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/760ce379-2153-4d72-acb7-f3f3f08a9028-tmpfs\") pod \"packageserver-d55dfcdfc-896hq\" (UID: \"760ce379-2153-4d72-acb7-f3f3f08a9028\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-896hq" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.001779 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/6b363da7-b1c6-495d-a05b-7ff4d7951b3e-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-jqnpq\" (UID: \"6b363da7-b1c6-495d-a05b-7ff4d7951b3e\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-jqnpq" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.001797 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j9wzh\" (UniqueName: \"kubernetes.io/projected/9504d491-6c27-4387-b6da-fa0192cc7a05-kube-api-access-j9wzh\") pod \"cni-sysctl-allowlist-ds-44pwx\" (UID: \"9504d491-6c27-4387-b6da-fa0192cc7a05\") " pod="openshift-multus/cni-sysctl-allowlist-ds-44pwx" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.001815 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/1039cddf-be22-4ca1-a622-d21967f9d859-srv-cert\") pod \"olm-operator-6b444d44fb-bzj6n\" (UID: \"1039cddf-be22-4ca1-a622-d21967f9d859\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bzj6n" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.001832 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9eb418ce-65b6-4da3-b282-831dd62916cf-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-2bc4b\" (UID: \"9eb418ce-65b6-4da3-b282-831dd62916cf\") " pod="openshift-marketplace/marketplace-operator-79b997595-2bc4b" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.001850 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/062b3680-8e44-4ab0-8f9b-a319169b53f1-cert\") pod \"ingress-canary-xr5v6\" (UID: \"062b3680-8e44-4ab0-8f9b-a319169b53f1\") " pod="openshift-ingress-canary/ingress-canary-xr5v6" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.001867 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/22010e1d-3ec1-4564-a362-bc963f691f1f-service-ca-bundle\") pod \"router-default-5444994796-jcnpq\" (UID: \"22010e1d-3ec1-4564-a362-bc963f691f1f\") " pod="openshift-ingress/router-default-5444994796-jcnpq" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.001885 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/760ce379-2153-4d72-acb7-f3f3f08a9028-webhook-cert\") pod \"packageserver-d55dfcdfc-896hq\" (UID: \"760ce379-2153-4d72-acb7-f3f3f08a9028\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-896hq" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.001924 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hfxrc\" (UniqueName: \"kubernetes.io/projected/3134d1ce-629c-4e8b-86c7-b69a24578952-kube-api-access-hfxrc\") pod \"package-server-manager-789f6589d5-mdws7\" (UID: \"3134d1ce-629c-4e8b-86c7-b69a24578952\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-mdws7" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.001941 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/22010e1d-3ec1-4564-a362-bc963f691f1f-metrics-certs\") pod \"router-default-5444994796-jcnpq\" (UID: \"22010e1d-3ec1-4564-a362-bc963f691f1f\") " pod="openshift-ingress/router-default-5444994796-jcnpq" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.001957 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/73b80803-714a-446e-9df2-6e3ccf5430ee-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-dxzdh\" (UID: \"73b80803-714a-446e-9df2-6e3ccf5430ee\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-dxzdh" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.001996 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/9bc5a0ed-1419-4ba0-8478-9bcad6a19954-socket-dir\") pod \"csi-hostpathplugin-r6sdp\" (UID: \"9bc5a0ed-1419-4ba0-8478-9bcad6a19954\") " pod="hostpath-provisioner/csi-hostpathplugin-r6sdp" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.002019 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tgg82\" (UniqueName: \"kubernetes.io/projected/f8a4f24e-7ccd-4527-9020-2f101c1d4cb5-kube-api-access-tgg82\") pod \"multus-admission-controller-857f4d67dd-52qtd\" (UID: \"f8a4f24e-7ccd-4527-9020-2f101c1d4cb5\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-52qtd" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.002056 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1a4f8da7-c099-4a8e-bffc-ca17ca74eb42-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-658pf\" (UID: \"1a4f8da7-c099-4a8e-bffc-ca17ca74eb42\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-658pf" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.002072 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1a4f8da7-c099-4a8e-bffc-ca17ca74eb42-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-658pf\" (UID: \"1a4f8da7-c099-4a8e-bffc-ca17ca74eb42\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-658pf" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.002101 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-grc4b\" (UniqueName: \"kubernetes.io/projected/4415ae1d-dbf4-450b-a329-6dc64b0a019f-kube-api-access-grc4b\") pod \"service-ca-9c57cc56f-xltdb\" (UID: \"4415ae1d-dbf4-450b-a329-6dc64b0a019f\") " pod="openshift-service-ca/service-ca-9c57cc56f-xltdb" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.002118 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ghbsr\" (UniqueName: \"kubernetes.io/projected/760ce379-2153-4d72-acb7-f3f3f08a9028-kube-api-access-ghbsr\") pod \"packageserver-d55dfcdfc-896hq\" (UID: \"760ce379-2153-4d72-acb7-f3f3f08a9028\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-896hq" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.002148 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/51595599-be30-46b6-bab6-ee647410efbc-config\") pod \"kube-controller-manager-operator-78b949d7b-fx9mb\" (UID: \"51595599-be30-46b6-bab6-ee647410efbc\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-fx9mb" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.002220 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/2eddcce3-58a6-45f9-b924-30d95fa897ae-auth-proxy-config\") pod \"machine-config-operator-74547568cd-5h9cj\" (UID: \"2eddcce3-58a6-45f9-b924-30d95fa897ae\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-5h9cj" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.002253 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6fccb501-125b-4bc4-8cad-3349f332c2de-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-gk97s\" (UID: \"6fccb501-125b-4bc4-8cad-3349f332c2de\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-gk97s" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.002271 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/51595599-be30-46b6-bab6-ee647410efbc-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-fx9mb\" (UID: \"51595599-be30-46b6-bab6-ee647410efbc\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-fx9mb" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.002303 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pblk2\" (UniqueName: \"kubernetes.io/projected/d5d77be3-1836-4c8b-9fc4-ab34196461c5-kube-api-access-pblk2\") pod \"service-ca-operator-777779d784-qdxf6\" (UID: \"d5d77be3-1836-4c8b-9fc4-ab34196461c5\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-qdxf6" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.002324 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zlnfn\" (UniqueName: \"kubernetes.io/projected/0709d8d6-5c00-4ec2-8651-df2cd64fe419-kube-api-access-zlnfn\") pod \"machine-config-server-rsxk2\" (UID: \"0709d8d6-5c00-4ec2-8651-df2cd64fe419\") " pod="openshift-machine-config-operator/machine-config-server-rsxk2" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.002341 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/0709d8d6-5c00-4ec2-8651-df2cd64fe419-node-bootstrap-token\") pod \"machine-config-server-rsxk2\" (UID: \"0709d8d6-5c00-4ec2-8651-df2cd64fe419\") " pod="openshift-machine-config-operator/machine-config-server-rsxk2" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.002358 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/5db3c754-f675-4ec8-b3aa-65d0528bb9ca-metrics-tls\") pod \"ingress-operator-5b745b69d9-g8n9t\" (UID: \"5db3c754-f675-4ec8-b3aa-65d0528bb9ca\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-g8n9t" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.002389 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/22010e1d-3ec1-4564-a362-bc963f691f1f-default-certificate\") pod \"router-default-5444994796-jcnpq\" (UID: \"22010e1d-3ec1-4564-a362-bc963f691f1f\") " pod="openshift-ingress/router-default-5444994796-jcnpq" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.002420 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gwqqt\" (UniqueName: \"kubernetes.io/projected/2a2f7727-f05e-4deb-8eba-cb3ec3561621-kube-api-access-gwqqt\") pod \"control-plane-machine-set-operator-78cbb6b69f-2v5d6\" (UID: \"2a2f7727-f05e-4deb-8eba-cb3ec3561621\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-2v5d6" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.002448 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5tdhq\" (UniqueName: \"kubernetes.io/projected/6fccb501-125b-4bc4-8cad-3349f332c2de-kube-api-access-5tdhq\") pod \"kube-storage-version-migrator-operator-b67b599dd-gk97s\" (UID: \"6fccb501-125b-4bc4-8cad-3349f332c2de\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-gk97s" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.002469 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3134d1ce-629c-4e8b-86c7-b69a24578952-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-mdws7\" (UID: \"3134d1ce-629c-4e8b-86c7-b69a24578952\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-mdws7" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.002493 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/76a05ca6-a98c-4313-855c-fd6ffaceee4f-secret-volume\") pod \"collect-profiles-29405595-2v6fk\" (UID: \"76a05ca6-a98c-4313-855c-fd6ffaceee4f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405595-2v6fk" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.002513 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/9504d491-6c27-4387-b6da-fa0192cc7a05-tuning-conf-dir\") pod \"cni-sysctl-allowlist-ds-44pwx\" (UID: \"9504d491-6c27-4387-b6da-fa0192cc7a05\") " pod="openshift-multus/cni-sysctl-allowlist-ds-44pwx" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.002552 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/4415ae1d-dbf4-450b-a329-6dc64b0a019f-signing-key\") pod \"service-ca-9c57cc56f-xltdb\" (UID: \"4415ae1d-dbf4-450b-a329-6dc64b0a019f\") " pod="openshift-service-ca/service-ca-9c57cc56f-xltdb" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.012237 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.013297 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-s7zwj"] Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.027423 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-2bc4b"] Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.031522 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/062b3680-8e44-4ab0-8f9b-a319169b53f1-cert\") pod \"ingress-canary-xr5v6\" (UID: \"062b3680-8e44-4ab0-8f9b-a319169b53f1\") " pod="openshift-ingress-canary/ingress-canary-xr5v6" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.036084 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.038102 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/f8a4f24e-7ccd-4527-9020-2f101c1d4cb5-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-52qtd\" (UID: \"f8a4f24e-7ccd-4527-9020-2f101c1d4cb5\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-52qtd" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.055402 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.079070 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.099780 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.103430 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sssnb\" (UniqueName: \"kubernetes.io/projected/1039cddf-be22-4ca1-a622-d21967f9d859-kube-api-access-sssnb\") pod \"olm-operator-6b444d44fb-bzj6n\" (UID: \"1039cddf-be22-4ca1-a622-d21967f9d859\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bzj6n" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.103466 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/9eb418ce-65b6-4da3-b282-831dd62916cf-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-2bc4b\" (UID: \"9eb418ce-65b6-4da3-b282-831dd62916cf\") " pod="openshift-marketplace/marketplace-operator-79b997595-2bc4b" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.103487 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ready\" (UniqueName: \"kubernetes.io/empty-dir/9504d491-6c27-4387-b6da-fa0192cc7a05-ready\") pod \"cni-sysctl-allowlist-ds-44pwx\" (UID: \"9504d491-6c27-4387-b6da-fa0192cc7a05\") " pod="openshift-multus/cni-sysctl-allowlist-ds-44pwx" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.103504 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/5db3c754-f675-4ec8-b3aa-65d0528bb9ca-bound-sa-token\") pod \"ingress-operator-5b745b69d9-g8n9t\" (UID: \"5db3c754-f675-4ec8-b3aa-65d0528bb9ca\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-g8n9t" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.103527 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/089e1482-6475-40b3-a005-c151f4bb0a87-config-volume\") pod \"dns-default-cg4zv\" (UID: \"089e1482-6475-40b3-a005-c151f4bb0a87\") " pod="openshift-dns/dns-default-cg4zv" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.103542 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bgtn5\" (UniqueName: \"kubernetes.io/projected/9eb418ce-65b6-4da3-b282-831dd62916cf-kube-api-access-bgtn5\") pod \"marketplace-operator-79b997595-2bc4b\" (UID: \"9eb418ce-65b6-4da3-b282-831dd62916cf\") " pod="openshift-marketplace/marketplace-operator-79b997595-2bc4b" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.103560 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/9504d491-6c27-4387-b6da-fa0192cc7a05-cni-sysctl-allowlist\") pod \"cni-sysctl-allowlist-ds-44pwx\" (UID: \"9504d491-6c27-4387-b6da-fa0192cc7a05\") " pod="openshift-multus/cni-sysctl-allowlist-ds-44pwx" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.103575 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/089e1482-6475-40b3-a005-c151f4bb0a87-metrics-tls\") pod \"dns-default-cg4zv\" (UID: \"089e1482-6475-40b3-a005-c151f4bb0a87\") " pod="openshift-dns/dns-default-cg4zv" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.103591 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6b363da7-b1c6-495d-a05b-7ff4d7951b3e-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-jqnpq\" (UID: \"6b363da7-b1c6-495d-a05b-7ff4d7951b3e\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-jqnpq" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.103616 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d5d77be3-1836-4c8b-9fc4-ab34196461c5-config\") pod \"service-ca-operator-777779d784-qdxf6\" (UID: \"d5d77be3-1836-4c8b-9fc4-ab34196461c5\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-qdxf6" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.103631 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c7sdh\" (UniqueName: \"kubernetes.io/projected/9bc5a0ed-1419-4ba0-8478-9bcad6a19954-kube-api-access-c7sdh\") pod \"csi-hostpathplugin-r6sdp\" (UID: \"9bc5a0ed-1419-4ba0-8478-9bcad6a19954\") " pod="hostpath-provisioner/csi-hostpathplugin-r6sdp" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.103646 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5db3c754-f675-4ec8-b3aa-65d0528bb9ca-trusted-ca\") pod \"ingress-operator-5b745b69d9-g8n9t\" (UID: \"5db3c754-f675-4ec8-b3aa-65d0528bb9ca\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-g8n9t" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.103662 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l8rxn\" (UniqueName: \"kubernetes.io/projected/5db3c754-f675-4ec8-b3aa-65d0528bb9ca-kube-api-access-l8rxn\") pod \"ingress-operator-5b745b69d9-g8n9t\" (UID: \"5db3c754-f675-4ec8-b3aa-65d0528bb9ca\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-g8n9t" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.103676 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-flzfb\" (UniqueName: \"kubernetes.io/projected/2eddcce3-58a6-45f9-b924-30d95fa897ae-kube-api-access-flzfb\") pod \"machine-config-operator-74547568cd-5h9cj\" (UID: \"2eddcce3-58a6-45f9-b924-30d95fa897ae\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-5h9cj" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.103754 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b4cst\" (UniqueName: \"kubernetes.io/projected/6fe798f2-84e0-4019-a75c-0ee2f2e672f5-kube-api-access-b4cst\") pod \"dns-operator-744455d44c-82qbh\" (UID: \"6fe798f2-84e0-4019-a75c-0ee2f2e672f5\") " pod="openshift-dns-operator/dns-operator-744455d44c-82qbh" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.103796 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/9bc5a0ed-1419-4ba0-8478-9bcad6a19954-plugins-dir\") pod \"csi-hostpathplugin-r6sdp\" (UID: \"9bc5a0ed-1419-4ba0-8478-9bcad6a19954\") " pod="hostpath-provisioner/csi-hostpathplugin-r6sdp" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.103813 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5pq68\" (UniqueName: \"kubernetes.io/projected/f2139e10-e6bf-4193-b242-7e061396f034-kube-api-access-5pq68\") pod \"migrator-59844c95c7-977k9\" (UID: \"f2139e10-e6bf-4193-b242-7e061396f034\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-977k9" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.103829 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1a4f8da7-c099-4a8e-bffc-ca17ca74eb42-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-658pf\" (UID: \"1a4f8da7-c099-4a8e-bffc-ca17ca74eb42\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-658pf" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.103843 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/9bc5a0ed-1419-4ba0-8478-9bcad6a19954-mountpoint-dir\") pod \"csi-hostpathplugin-r6sdp\" (UID: \"9bc5a0ed-1419-4ba0-8478-9bcad6a19954\") " pod="hostpath-provisioner/csi-hostpathplugin-r6sdp" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.103858 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/22010e1d-3ec1-4564-a362-bc963f691f1f-stats-auth\") pod \"router-default-5444994796-jcnpq\" (UID: \"22010e1d-3ec1-4564-a362-bc963f691f1f\") " pod="openshift-ingress/router-default-5444994796-jcnpq" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.103876 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/2eddcce3-58a6-45f9-b924-30d95fa897ae-images\") pod \"machine-config-operator-74547568cd-5h9cj\" (UID: \"2eddcce3-58a6-45f9-b924-30d95fa897ae\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-5h9cj" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.103892 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/76a05ca6-a98c-4313-855c-fd6ffaceee4f-config-volume\") pod \"collect-profiles-29405595-2v6fk\" (UID: \"76a05ca6-a98c-4313-855c-fd6ffaceee4f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405595-2v6fk" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.103909 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4lkgw\" (UniqueName: \"kubernetes.io/projected/089e1482-6475-40b3-a005-c151f4bb0a87-kube-api-access-4lkgw\") pod \"dns-default-cg4zv\" (UID: \"089e1482-6475-40b3-a005-c151f4bb0a87\") " pod="openshift-dns/dns-default-cg4zv" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.103927 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/9bc5a0ed-1419-4ba0-8478-9bcad6a19954-csi-data-dir\") pod \"csi-hostpathplugin-r6sdp\" (UID: \"9bc5a0ed-1419-4ba0-8478-9bcad6a19954\") " pod="hostpath-provisioner/csi-hostpathplugin-r6sdp" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.103950 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/73b80803-714a-446e-9df2-6e3ccf5430ee-proxy-tls\") pod \"machine-config-controller-84d6567774-dxzdh\" (UID: \"73b80803-714a-446e-9df2-6e3ccf5430ee\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-dxzdh" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.103968 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6b363da7-b1c6-495d-a05b-7ff4d7951b3e-config\") pod \"kube-apiserver-operator-766d6c64bb-jqnpq\" (UID: \"6b363da7-b1c6-495d-a05b-7ff4d7951b3e\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-jqnpq" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.103989 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b54w6\" (UniqueName: \"kubernetes.io/projected/76a05ca6-a98c-4313-855c-fd6ffaceee4f-kube-api-access-b54w6\") pod \"collect-profiles-29405595-2v6fk\" (UID: \"76a05ca6-a98c-4313-855c-fd6ffaceee4f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405595-2v6fk" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.104006 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/51595599-be30-46b6-bab6-ee647410efbc-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-fx9mb\" (UID: \"51595599-be30-46b6-bab6-ee647410efbc\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-fx9mb" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.104027 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gw52s\" (UniqueName: \"kubernetes.io/projected/22010e1d-3ec1-4564-a362-bc963f691f1f-kube-api-access-gw52s\") pod \"router-default-5444994796-jcnpq\" (UID: \"22010e1d-3ec1-4564-a362-bc963f691f1f\") " pod="openshift-ingress/router-default-5444994796-jcnpq" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.104049 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/6b363da7-b1c6-495d-a05b-7ff4d7951b3e-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-jqnpq\" (UID: \"6b363da7-b1c6-495d-a05b-7ff4d7951b3e\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-jqnpq" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.104066 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j9wzh\" (UniqueName: \"kubernetes.io/projected/9504d491-6c27-4387-b6da-fa0192cc7a05-kube-api-access-j9wzh\") pod \"cni-sysctl-allowlist-ds-44pwx\" (UID: \"9504d491-6c27-4387-b6da-fa0192cc7a05\") " pod="openshift-multus/cni-sysctl-allowlist-ds-44pwx" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.104082 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/760ce379-2153-4d72-acb7-f3f3f08a9028-tmpfs\") pod \"packageserver-d55dfcdfc-896hq\" (UID: \"760ce379-2153-4d72-acb7-f3f3f08a9028\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-896hq" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.104098 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/1039cddf-be22-4ca1-a622-d21967f9d859-srv-cert\") pod \"olm-operator-6b444d44fb-bzj6n\" (UID: \"1039cddf-be22-4ca1-a622-d21967f9d859\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bzj6n" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.104114 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9eb418ce-65b6-4da3-b282-831dd62916cf-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-2bc4b\" (UID: \"9eb418ce-65b6-4da3-b282-831dd62916cf\") " pod="openshift-marketplace/marketplace-operator-79b997595-2bc4b" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.104540 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/22010e1d-3ec1-4564-a362-bc963f691f1f-service-ca-bundle\") pod \"router-default-5444994796-jcnpq\" (UID: \"22010e1d-3ec1-4564-a362-bc963f691f1f\") " pod="openshift-ingress/router-default-5444994796-jcnpq" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.104585 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hfxrc\" (UniqueName: \"kubernetes.io/projected/3134d1ce-629c-4e8b-86c7-b69a24578952-kube-api-access-hfxrc\") pod \"package-server-manager-789f6589d5-mdws7\" (UID: \"3134d1ce-629c-4e8b-86c7-b69a24578952\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-mdws7" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.104605 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/760ce379-2153-4d72-acb7-f3f3f08a9028-webhook-cert\") pod \"packageserver-d55dfcdfc-896hq\" (UID: \"760ce379-2153-4d72-acb7-f3f3f08a9028\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-896hq" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.104621 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/22010e1d-3ec1-4564-a362-bc963f691f1f-metrics-certs\") pod \"router-default-5444994796-jcnpq\" (UID: \"22010e1d-3ec1-4564-a362-bc963f691f1f\") " pod="openshift-ingress/router-default-5444994796-jcnpq" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.104637 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/73b80803-714a-446e-9df2-6e3ccf5430ee-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-dxzdh\" (UID: \"73b80803-714a-446e-9df2-6e3ccf5430ee\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-dxzdh" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.104656 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/9bc5a0ed-1419-4ba0-8478-9bcad6a19954-socket-dir\") pod \"csi-hostpathplugin-r6sdp\" (UID: \"9bc5a0ed-1419-4ba0-8478-9bcad6a19954\") " pod="hostpath-provisioner/csi-hostpathplugin-r6sdp" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.104681 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1a4f8da7-c099-4a8e-bffc-ca17ca74eb42-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-658pf\" (UID: \"1a4f8da7-c099-4a8e-bffc-ca17ca74eb42\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-658pf" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.104699 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1a4f8da7-c099-4a8e-bffc-ca17ca74eb42-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-658pf\" (UID: \"1a4f8da7-c099-4a8e-bffc-ca17ca74eb42\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-658pf" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.104720 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-grc4b\" (UniqueName: \"kubernetes.io/projected/4415ae1d-dbf4-450b-a329-6dc64b0a019f-kube-api-access-grc4b\") pod \"service-ca-9c57cc56f-xltdb\" (UID: \"4415ae1d-dbf4-450b-a329-6dc64b0a019f\") " pod="openshift-service-ca/service-ca-9c57cc56f-xltdb" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.104735 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ghbsr\" (UniqueName: \"kubernetes.io/projected/760ce379-2153-4d72-acb7-f3f3f08a9028-kube-api-access-ghbsr\") pod \"packageserver-d55dfcdfc-896hq\" (UID: \"760ce379-2153-4d72-acb7-f3f3f08a9028\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-896hq" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.104752 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/51595599-be30-46b6-bab6-ee647410efbc-config\") pod \"kube-controller-manager-operator-78b949d7b-fx9mb\" (UID: \"51595599-be30-46b6-bab6-ee647410efbc\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-fx9mb" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.104768 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/2eddcce3-58a6-45f9-b924-30d95fa897ae-auth-proxy-config\") pod \"machine-config-operator-74547568cd-5h9cj\" (UID: \"2eddcce3-58a6-45f9-b924-30d95fa897ae\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-5h9cj" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.104786 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6fccb501-125b-4bc4-8cad-3349f332c2de-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-gk97s\" (UID: \"6fccb501-125b-4bc4-8cad-3349f332c2de\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-gk97s" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.104805 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zlnfn\" (UniqueName: \"kubernetes.io/projected/0709d8d6-5c00-4ec2-8651-df2cd64fe419-kube-api-access-zlnfn\") pod \"machine-config-server-rsxk2\" (UID: \"0709d8d6-5c00-4ec2-8651-df2cd64fe419\") " pod="openshift-machine-config-operator/machine-config-server-rsxk2" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.104822 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/51595599-be30-46b6-bab6-ee647410efbc-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-fx9mb\" (UID: \"51595599-be30-46b6-bab6-ee647410efbc\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-fx9mb" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.104838 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pblk2\" (UniqueName: \"kubernetes.io/projected/d5d77be3-1836-4c8b-9fc4-ab34196461c5-kube-api-access-pblk2\") pod \"service-ca-operator-777779d784-qdxf6\" (UID: \"d5d77be3-1836-4c8b-9fc4-ab34196461c5\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-qdxf6" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.104854 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/22010e1d-3ec1-4564-a362-bc963f691f1f-default-certificate\") pod \"router-default-5444994796-jcnpq\" (UID: \"22010e1d-3ec1-4564-a362-bc963f691f1f\") " pod="openshift-ingress/router-default-5444994796-jcnpq" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.104869 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/0709d8d6-5c00-4ec2-8651-df2cd64fe419-node-bootstrap-token\") pod \"machine-config-server-rsxk2\" (UID: \"0709d8d6-5c00-4ec2-8651-df2cd64fe419\") " pod="openshift-machine-config-operator/machine-config-server-rsxk2" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.104884 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/5db3c754-f675-4ec8-b3aa-65d0528bb9ca-metrics-tls\") pod \"ingress-operator-5b745b69d9-g8n9t\" (UID: \"5db3c754-f675-4ec8-b3aa-65d0528bb9ca\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-g8n9t" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.104903 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gwqqt\" (UniqueName: \"kubernetes.io/projected/2a2f7727-f05e-4deb-8eba-cb3ec3561621-kube-api-access-gwqqt\") pod \"control-plane-machine-set-operator-78cbb6b69f-2v5d6\" (UID: \"2a2f7727-f05e-4deb-8eba-cb3ec3561621\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-2v5d6" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.104931 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5tdhq\" (UniqueName: \"kubernetes.io/projected/6fccb501-125b-4bc4-8cad-3349f332c2de-kube-api-access-5tdhq\") pod \"kube-storage-version-migrator-operator-b67b599dd-gk97s\" (UID: \"6fccb501-125b-4bc4-8cad-3349f332c2de\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-gk97s" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.104956 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3134d1ce-629c-4e8b-86c7-b69a24578952-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-mdws7\" (UID: \"3134d1ce-629c-4e8b-86c7-b69a24578952\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-mdws7" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.104973 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/76a05ca6-a98c-4313-855c-fd6ffaceee4f-secret-volume\") pod \"collect-profiles-29405595-2v6fk\" (UID: \"76a05ca6-a98c-4313-855c-fd6ffaceee4f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405595-2v6fk" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.104990 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/9504d491-6c27-4387-b6da-fa0192cc7a05-tuning-conf-dir\") pod \"cni-sysctl-allowlist-ds-44pwx\" (UID: \"9504d491-6c27-4387-b6da-fa0192cc7a05\") " pod="openshift-multus/cni-sysctl-allowlist-ds-44pwx" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.105009 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/4415ae1d-dbf4-450b-a329-6dc64b0a019f-signing-key\") pod \"service-ca-9c57cc56f-xltdb\" (UID: \"4415ae1d-dbf4-450b-a329-6dc64b0a019f\") " pod="openshift-service-ca/service-ca-9c57cc56f-xltdb" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.105025 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/2eddcce3-58a6-45f9-b924-30d95fa897ae-proxy-tls\") pod \"machine-config-operator-74547568cd-5h9cj\" (UID: \"2eddcce3-58a6-45f9-b924-30d95fa897ae\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-5h9cj" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.105050 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6fccb501-125b-4bc4-8cad-3349f332c2de-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-gk97s\" (UID: \"6fccb501-125b-4bc4-8cad-3349f332c2de\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-gk97s" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.105067 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/2a2f7727-f05e-4deb-8eba-cb3ec3561621-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-2v5d6\" (UID: \"2a2f7727-f05e-4deb-8eba-cb3ec3561621\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-2v5d6" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.105083 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5stxb\" (UniqueName: \"kubernetes.io/projected/73b80803-714a-446e-9df2-6e3ccf5430ee-kube-api-access-5stxb\") pod \"machine-config-controller-84d6567774-dxzdh\" (UID: \"73b80803-714a-446e-9df2-6e3ccf5430ee\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-dxzdh" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.105105 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qfstd\" (UniqueName: \"kubernetes.io/projected/0e2f0bc1-2261-484e-8f4b-34bddacda1bc-kube-api-access-qfstd\") pod \"catalog-operator-68c6474976-hzvzz\" (UID: \"0e2f0bc1-2261-484e-8f4b-34bddacda1bc\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-hzvzz" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.105122 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/4415ae1d-dbf4-450b-a329-6dc64b0a019f-signing-cabundle\") pod \"service-ca-9c57cc56f-xltdb\" (UID: \"4415ae1d-dbf4-450b-a329-6dc64b0a019f\") " pod="openshift-service-ca/service-ca-9c57cc56f-xltdb" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.105139 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/0e2f0bc1-2261-484e-8f4b-34bddacda1bc-srv-cert\") pod \"catalog-operator-68c6474976-hzvzz\" (UID: \"0e2f0bc1-2261-484e-8f4b-34bddacda1bc\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-hzvzz" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.105156 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/0e2f0bc1-2261-484e-8f4b-34bddacda1bc-profile-collector-cert\") pod \"catalog-operator-68c6474976-hzvzz\" (UID: \"0e2f0bc1-2261-484e-8f4b-34bddacda1bc\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-hzvzz" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.105173 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/1039cddf-be22-4ca1-a622-d21967f9d859-profile-collector-cert\") pod \"olm-operator-6b444d44fb-bzj6n\" (UID: \"1039cddf-be22-4ca1-a622-d21967f9d859\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bzj6n" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.105188 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d5d77be3-1836-4c8b-9fc4-ab34196461c5-serving-cert\") pod \"service-ca-operator-777779d784-qdxf6\" (UID: \"d5d77be3-1836-4c8b-9fc4-ab34196461c5\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-qdxf6" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.105203 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/0709d8d6-5c00-4ec2-8651-df2cd64fe419-certs\") pod \"machine-config-server-rsxk2\" (UID: \"0709d8d6-5c00-4ec2-8651-df2cd64fe419\") " pod="openshift-machine-config-operator/machine-config-server-rsxk2" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.105218 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/6fe798f2-84e0-4019-a75c-0ee2f2e672f5-metrics-tls\") pod \"dns-operator-744455d44c-82qbh\" (UID: \"6fe798f2-84e0-4019-a75c-0ee2f2e672f5\") " pod="openshift-dns-operator/dns-operator-744455d44c-82qbh" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.105264 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/760ce379-2153-4d72-acb7-f3f3f08a9028-apiservice-cert\") pod \"packageserver-d55dfcdfc-896hq\" (UID: \"760ce379-2153-4d72-acb7-f3f3f08a9028\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-896hq" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.105300 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/9bc5a0ed-1419-4ba0-8478-9bcad6a19954-registration-dir\") pod \"csi-hostpathplugin-r6sdp\" (UID: \"9bc5a0ed-1419-4ba0-8478-9bcad6a19954\") " pod="hostpath-provisioner/csi-hostpathplugin-r6sdp" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.105646 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/9bc5a0ed-1419-4ba0-8478-9bcad6a19954-registration-dir\") pod \"csi-hostpathplugin-r6sdp\" (UID: \"9bc5a0ed-1419-4ba0-8478-9bcad6a19954\") " pod="hostpath-provisioner/csi-hostpathplugin-r6sdp" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.106000 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ready\" (UniqueName: \"kubernetes.io/empty-dir/9504d491-6c27-4387-b6da-fa0192cc7a05-ready\") pod \"cni-sysctl-allowlist-ds-44pwx\" (UID: \"9504d491-6c27-4387-b6da-fa0192cc7a05\") " pod="openshift-multus/cni-sysctl-allowlist-ds-44pwx" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.106330 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/9bc5a0ed-1419-4ba0-8478-9bcad6a19954-socket-dir\") pod \"csi-hostpathplugin-r6sdp\" (UID: \"9bc5a0ed-1419-4ba0-8478-9bcad6a19954\") " pod="hostpath-provisioner/csi-hostpathplugin-r6sdp" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.106664 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/9bc5a0ed-1419-4ba0-8478-9bcad6a19954-mountpoint-dir\") pod \"csi-hostpathplugin-r6sdp\" (UID: \"9bc5a0ed-1419-4ba0-8478-9bcad6a19954\") " pod="hostpath-provisioner/csi-hostpathplugin-r6sdp" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.106927 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d5d77be3-1836-4c8b-9fc4-ab34196461c5-config\") pod \"service-ca-operator-777779d784-qdxf6\" (UID: \"d5d77be3-1836-4c8b-9fc4-ab34196461c5\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-qdxf6" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.106976 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/73b80803-714a-446e-9df2-6e3ccf5430ee-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-dxzdh\" (UID: \"73b80803-714a-446e-9df2-6e3ccf5430ee\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-dxzdh" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.107146 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/9bc5a0ed-1419-4ba0-8478-9bcad6a19954-plugins-dir\") pod \"csi-hostpathplugin-r6sdp\" (UID: \"9bc5a0ed-1419-4ba0-8478-9bcad6a19954\") " pod="hostpath-provisioner/csi-hostpathplugin-r6sdp" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.119038 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/9bc5a0ed-1419-4ba0-8478-9bcad6a19954-csi-data-dir\") pod \"csi-hostpathplugin-r6sdp\" (UID: \"9bc5a0ed-1419-4ba0-8478-9bcad6a19954\") " pod="hostpath-provisioner/csi-hostpathplugin-r6sdp" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.119736 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/4415ae1d-dbf4-450b-a329-6dc64b0a019f-signing-cabundle\") pod \"service-ca-9c57cc56f-xltdb\" (UID: \"4415ae1d-dbf4-450b-a329-6dc64b0a019f\") " pod="openshift-service-ca/service-ca-9c57cc56f-xltdb" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.120177 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/2eddcce3-58a6-45f9-b924-30d95fa897ae-auth-proxy-config\") pod \"machine-config-operator-74547568cd-5h9cj\" (UID: \"2eddcce3-58a6-45f9-b924-30d95fa897ae\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-5h9cj" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.122979 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/760ce379-2153-4d72-acb7-f3f3f08a9028-tmpfs\") pod \"packageserver-d55dfcdfc-896hq\" (UID: \"760ce379-2153-4d72-acb7-f3f3f08a9028\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-896hq" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.123426 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/51595599-be30-46b6-bab6-ee647410efbc-config\") pod \"kube-controller-manager-operator-78b949d7b-fx9mb\" (UID: \"51595599-be30-46b6-bab6-ee647410efbc\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-fx9mb" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.123562 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/76a05ca6-a98c-4313-855c-fd6ffaceee4f-config-volume\") pod \"collect-profiles-29405595-2v6fk\" (UID: \"76a05ca6-a98c-4313-855c-fd6ffaceee4f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405595-2v6fk" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.123714 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/9504d491-6c27-4387-b6da-fa0192cc7a05-tuning-conf-dir\") pod \"cni-sysctl-allowlist-ds-44pwx\" (UID: \"9504d491-6c27-4387-b6da-fa0192cc7a05\") " pod="openshift-multus/cni-sysctl-allowlist-ds-44pwx" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.132431 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.135644 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.137176 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/760ce379-2153-4d72-acb7-f3f3f08a9028-webhook-cert\") pod \"packageserver-d55dfcdfc-896hq\" (UID: \"760ce379-2153-4d72-acb7-f3f3f08a9028\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-896hq" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.166716 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.170145 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/22010e1d-3ec1-4564-a362-bc963f691f1f-default-certificate\") pod \"router-default-5444994796-jcnpq\" (UID: \"22010e1d-3ec1-4564-a362-bc963f691f1f\") " pod="openshift-ingress/router-default-5444994796-jcnpq" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.187127 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/22010e1d-3ec1-4564-a362-bc963f691f1f-stats-auth\") pod \"router-default-5444994796-jcnpq\" (UID: \"22010e1d-3ec1-4564-a362-bc963f691f1f\") " pod="openshift-ingress/router-default-5444994796-jcnpq" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.189401 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.191018 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/22010e1d-3ec1-4564-a362-bc963f691f1f-service-ca-bundle\") pod \"router-default-5444994796-jcnpq\" (UID: \"22010e1d-3ec1-4564-a362-bc963f691f1f\") " pod="openshift-ingress/router-default-5444994796-jcnpq" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.193139 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d5d77be3-1836-4c8b-9fc4-ab34196461c5-serving-cert\") pod \"service-ca-operator-777779d784-qdxf6\" (UID: \"d5d77be3-1836-4c8b-9fc4-ab34196461c5\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-qdxf6" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.194011 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/51595599-be30-46b6-bab6-ee647410efbc-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-fx9mb\" (UID: \"51595599-be30-46b6-bab6-ee647410efbc\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-fx9mb" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.194496 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/760ce379-2153-4d72-acb7-f3f3f08a9028-apiservice-cert\") pod \"packageserver-d55dfcdfc-896hq\" (UID: \"760ce379-2153-4d72-acb7-f3f3f08a9028\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-896hq" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.200352 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.200813 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3134d1ce-629c-4e8b-86c7-b69a24578952-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-mdws7\" (UID: \"3134d1ce-629c-4e8b-86c7-b69a24578952\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-mdws7" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.211352 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/22010e1d-3ec1-4564-a362-bc963f691f1f-metrics-certs\") pod \"router-default-5444994796-jcnpq\" (UID: \"22010e1d-3ec1-4564-a362-bc963f691f1f\") " pod="openshift-ingress/router-default-5444994796-jcnpq" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.213921 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/4415ae1d-dbf4-450b-a329-6dc64b0a019f-signing-key\") pod \"service-ca-9c57cc56f-xltdb\" (UID: \"4415ae1d-dbf4-450b-a329-6dc64b0a019f\") " pod="openshift-service-ca/service-ca-9c57cc56f-xltdb" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.217606 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.235822 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.246715 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/1039cddf-be22-4ca1-a622-d21967f9d859-profile-collector-cert\") pod \"olm-operator-6b444d44fb-bzj6n\" (UID: \"1039cddf-be22-4ca1-a622-d21967f9d859\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bzj6n" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.249332 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/76a05ca6-a98c-4313-855c-fd6ffaceee4f-secret-volume\") pod \"collect-profiles-29405595-2v6fk\" (UID: \"76a05ca6-a98c-4313-855c-fd6ffaceee4f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405595-2v6fk" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.254490 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.255147 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/0e2f0bc1-2261-484e-8f4b-34bddacda1bc-profile-collector-cert\") pod \"catalog-operator-68c6474976-hzvzz\" (UID: \"0e2f0bc1-2261-484e-8f4b-34bddacda1bc\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-hzvzz" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.267091 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/1039cddf-be22-4ca1-a622-d21967f9d859-srv-cert\") pod \"olm-operator-6b444d44fb-bzj6n\" (UID: \"1039cddf-be22-4ca1-a622-d21967f9d859\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bzj6n" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.276224 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.295533 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.296276 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-8tbnv"] Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.296952 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/089e1482-6475-40b3-a005-c151f4bb0a87-config-volume\") pod \"dns-default-cg4zv\" (UID: \"089e1482-6475-40b3-a005-c151f4bb0a87\") " pod="openshift-dns/dns-default-cg4zv" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.313072 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.324995 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-ddx6v"] Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.326195 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/6fe798f2-84e0-4019-a75c-0ee2f2e672f5-metrics-tls\") pod \"dns-operator-744455d44c-82qbh\" (UID: \"6fe798f2-84e0-4019-a75c-0ee2f2e672f5\") " pod="openshift-dns-operator/dns-operator-744455d44c-82qbh" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.333648 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.352162 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.375623 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.394410 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.400039 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/089e1482-6475-40b3-a005-c151f4bb0a87-metrics-tls\") pod \"dns-default-cg4zv\" (UID: \"089e1482-6475-40b3-a005-c151f4bb0a87\") " pod="openshift-dns/dns-default-cg4zv" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.413631 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.438173 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.451887 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.454348 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/73b80803-714a-446e-9df2-6e3ccf5430ee-proxy-tls\") pod \"machine-config-controller-84d6567774-dxzdh\" (UID: \"73b80803-714a-446e-9df2-6e3ccf5430ee\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-dxzdh" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.465876 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/0709d8d6-5c00-4ec2-8651-df2cd64fe419-certs\") pod \"machine-config-server-rsxk2\" (UID: \"0709d8d6-5c00-4ec2-8651-df2cd64fe419\") " pod="openshift-machine-config-operator/machine-config-server-rsxk2" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.472631 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.494461 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.505103 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/0709d8d6-5c00-4ec2-8651-df2cd64fe419-node-bootstrap-token\") pod \"machine-config-server-rsxk2\" (UID: \"0709d8d6-5c00-4ec2-8651-df2cd64fe419\") " pod="openshift-machine-config-operator/machine-config-server-rsxk2" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.512012 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.512081 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-dd58r" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.512163 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.512158 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.513939 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.532501 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.553070 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.575136 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1a4f8da7-c099-4a8e-bffc-ca17ca74eb42-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-658pf\" (UID: \"1a4f8da7-c099-4a8e-bffc-ca17ca74eb42\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-658pf" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.576115 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-c2rks"] Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.578878 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.587195 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1a4f8da7-c099-4a8e-bffc-ca17ca74eb42-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-658pf\" (UID: \"1a4f8da7-c099-4a8e-bffc-ca17ca74eb42\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-658pf" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.592671 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.595656 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-7gmxj"] Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.609170 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/2a2f7727-f05e-4deb-8eba-cb3ec3561621-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-2v5d6\" (UID: \"2a2f7727-f05e-4deb-8eba-cb3ec3561621\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-2v5d6" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.612713 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Nov 28 13:21:20 crc kubenswrapper[4631]: W1128 13:21:20.620853 4631 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfeeda081_8bb0_4a89_ae6d_2e87f8f73344.slice/crio-b78bb636b47e96620fe9e91db32a29107a4b64c240f259ca8fc16dccc2eb69b8 WatchSource:0}: Error finding container b78bb636b47e96620fe9e91db32a29107a4b64c240f259ca8fc16dccc2eb69b8: Status 404 returned error can't find the container with id b78bb636b47e96620fe9e91db32a29107a4b64c240f259ca8fc16dccc2eb69b8 Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.626644 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-2ch2b"] Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.633166 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-sysctl-allowlist" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.636815 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/9504d491-6c27-4387-b6da-fa0192cc7a05-cni-sysctl-allowlist\") pod \"cni-sysctl-allowlist-ds-44pwx\" (UID: \"9504d491-6c27-4387-b6da-fa0192cc7a05\") " pod="openshift-multus/cni-sysctl-allowlist-ds-44pwx" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.651981 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.702431 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.704474 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.708093 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/2eddcce3-58a6-45f9-b924-30d95fa897ae-images\") pod \"machine-config-operator-74547568cd-5h9cj\" (UID: \"2eddcce3-58a6-45f9-b924-30d95fa897ae\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-5h9cj" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.711841 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.733463 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.738585 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-bf7bx"] Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.741840 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-rwpv9"] Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.746672 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/2eddcce3-58a6-45f9-b924-30d95fa897ae-proxy-tls\") pod \"machine-config-operator-74547568cd-5h9cj\" (UID: \"2eddcce3-58a6-45f9-b924-30d95fa897ae\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-5h9cj" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.751734 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.760941 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-hgsfk"] Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.772636 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.780524 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-jgx6f"] Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.788181 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-vst8c"] Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.809688 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Nov 28 13:21:20 crc kubenswrapper[4631]: W1128 13:21:20.811207 4631 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b540aa_a7ac_4f13_b430_91e1328d6602.slice/crio-80c5c00e2f225553ffec03379cbcc7ec1f8e27c59d130114be292b2028faa1a9 WatchSource:0}: Error finding container 80c5c00e2f225553ffec03379cbcc7ec1f8e27c59d130114be292b2028faa1a9: Status 404 returned error can't find the container with id 80c5c00e2f225553ffec03379cbcc7ec1f8e27c59d130114be292b2028faa1a9 Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.813985 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.818918 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5db3c754-f675-4ec8-b3aa-65d0528bb9ca-trusted-ca\") pod \"ingress-operator-5b745b69d9-g8n9t\" (UID: \"5db3c754-f675-4ec8-b3aa-65d0528bb9ca\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-g8n9t" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.845960 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.846673 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-bf7bx" event={"ID":"d1b540aa-a7ac-4f13-b430-91e1328d6602","Type":"ContainerStarted","Data":"80c5c00e2f225553ffec03379cbcc7ec1f8e27c59d130114be292b2028faa1a9"} Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.848661 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-r5b6b"] Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.848710 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-7977d"] Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.850834 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8tbnv" event={"ID":"aa813d64-b22e-40a0-b72a-b4e220e85f5a","Type":"ContainerStarted","Data":"71de10ed2cbf0827c1a05997074bf08e49747b65283f893d65377db5d39dfc9e"} Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.852167 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.852611 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/5db3c754-f675-4ec8-b3aa-65d0528bb9ca-metrics-tls\") pod \"ingress-operator-5b745b69d9-g8n9t\" (UID: \"5db3c754-f675-4ec8-b3aa-65d0528bb9ca\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-g8n9t" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.856636 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-2ch2b" event={"ID":"889c117d-ea46-4f02-a2e0-42a47c6e4683","Type":"ContainerStarted","Data":"1b7149d3f0d8f48fea1bcb6531e6afa279b496883ba93cc6b3e281d732b443a6"} Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.870139 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ddx6v" event={"ID":"594fba15-6b93-4f34-af64-403a1b2d694c","Type":"ContainerStarted","Data":"eca0b8da471872c541822e9b1c0a121260a0a99163c68ca217f7989ad15d52e0"} Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.870176 4631 request.go:700] Waited for 1.008072463s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/configmaps?fieldSelector=metadata.name%3Dopenshift-service-ca.crt&limit=500&resourceVersion=0 Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.871457 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-hgsfk" event={"ID":"145c4da2-bb50-4900-84d7-974b0cfaa4d8","Type":"ContainerStarted","Data":"4a8d6262961fe4ede84838daf319eff68a0398144877b19b21af09e42cac844e"} Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.871699 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.873388 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-c2rks" event={"ID":"feeda081-8bb0-4a89-ae6d-2e87f8f73344","Type":"ContainerStarted","Data":"b78bb636b47e96620fe9e91db32a29107a4b64c240f259ca8fc16dccc2eb69b8"} Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.876866 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-rwpv9" event={"ID":"cbd1163d-84ad-427f-b270-787520942822","Type":"ContainerStarted","Data":"bde35cd4beb16bad491b29823cab32a0d1dd4066f04fc9c065e69087de25cd35"} Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.878944 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-7gmxj" event={"ID":"e7c7bd85-4add-4bda-a4fa-71a7b75787e4","Type":"ContainerStarted","Data":"807704f91ff09049c560dee0013b23bfdbc5fb270c9483b032886f494feb9fd7"} Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.893965 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.914007 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.922240 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/9eb418ce-65b6-4da3-b282-831dd62916cf-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-2bc4b\" (UID: \"9eb418ce-65b6-4da3-b282-831dd62916cf\") " pod="openshift-marketplace/marketplace-operator-79b997595-2bc4b" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.932962 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.959726 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.968815 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9eb418ce-65b6-4da3-b282-831dd62916cf-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-2bc4b\" (UID: \"9eb418ce-65b6-4da3-b282-831dd62916cf\") " pod="openshift-marketplace/marketplace-operator-79b997595-2bc4b" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.973483 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.978706 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/0e2f0bc1-2261-484e-8f4b-34bddacda1bc-srv-cert\") pod \"catalog-operator-68c6474976-hzvzz\" (UID: \"0e2f0bc1-2261-484e-8f4b-34bddacda1bc\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-hzvzz" Nov 28 13:21:20 crc kubenswrapper[4631]: I1128 13:21:20.993056 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Nov 28 13:21:21 crc kubenswrapper[4631]: I1128 13:21:21.011921 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Nov 28 13:21:21 crc kubenswrapper[4631]: I1128 13:21:21.033058 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Nov 28 13:21:21 crc kubenswrapper[4631]: I1128 13:21:21.054526 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Nov 28 13:21:21 crc kubenswrapper[4631]: I1128 13:21:21.075250 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Nov 28 13:21:21 crc kubenswrapper[4631]: I1128 13:21:21.093070 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Nov 28 13:21:21 crc kubenswrapper[4631]: E1128 13:21:21.106735 4631 secret.go:188] Couldn't get secret openshift-kube-apiserver-operator/kube-apiserver-operator-serving-cert: failed to sync secret cache: timed out waiting for the condition Nov 28 13:21:21 crc kubenswrapper[4631]: E1128 13:21:21.106875 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6b363da7-b1c6-495d-a05b-7ff4d7951b3e-serving-cert podName:6b363da7-b1c6-495d-a05b-7ff4d7951b3e nodeName:}" failed. No retries permitted until 2025-11-28 13:21:21.606840448 +0000 UTC m=+38.414143792 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "serving-cert" (UniqueName: "kubernetes.io/secret/6b363da7-b1c6-495d-a05b-7ff4d7951b3e-serving-cert") pod "kube-apiserver-operator-766d6c64bb-jqnpq" (UID: "6b363da7-b1c6-495d-a05b-7ff4d7951b3e") : failed to sync secret cache: timed out waiting for the condition Nov 28 13:21:21 crc kubenswrapper[4631]: E1128 13:21:21.108400 4631 secret.go:188] Couldn't get secret openshift-kube-storage-version-migrator-operator/serving-cert: failed to sync secret cache: timed out waiting for the condition Nov 28 13:21:21 crc kubenswrapper[4631]: E1128 13:21:21.108548 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6fccb501-125b-4bc4-8cad-3349f332c2de-serving-cert podName:6fccb501-125b-4bc4-8cad-3349f332c2de nodeName:}" failed. No retries permitted until 2025-11-28 13:21:21.608517064 +0000 UTC m=+38.415820408 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "serving-cert" (UniqueName: "kubernetes.io/secret/6fccb501-125b-4bc4-8cad-3349f332c2de-serving-cert") pod "kube-storage-version-migrator-operator-b67b599dd-gk97s" (UID: "6fccb501-125b-4bc4-8cad-3349f332c2de") : failed to sync secret cache: timed out waiting for the condition Nov 28 13:21:21 crc kubenswrapper[4631]: I1128 13:21:21.112901 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Nov 28 13:21:21 crc kubenswrapper[4631]: E1128 13:21:21.119482 4631 configmap.go:193] Couldn't get configMap openshift-kube-apiserver-operator/kube-apiserver-operator-config: failed to sync configmap cache: timed out waiting for the condition Nov 28 13:21:21 crc kubenswrapper[4631]: E1128 13:21:21.119597 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/6b363da7-b1c6-495d-a05b-7ff4d7951b3e-config podName:6b363da7-b1c6-495d-a05b-7ff4d7951b3e nodeName:}" failed. No retries permitted until 2025-11-28 13:21:21.619561162 +0000 UTC m=+38.426864506 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config" (UniqueName: "kubernetes.io/configmap/6b363da7-b1c6-495d-a05b-7ff4d7951b3e-config") pod "kube-apiserver-operator-766d6c64bb-jqnpq" (UID: "6b363da7-b1c6-495d-a05b-7ff4d7951b3e") : failed to sync configmap cache: timed out waiting for the condition Nov 28 13:21:21 crc kubenswrapper[4631]: E1128 13:21:21.131236 4631 configmap.go:193] Couldn't get configMap openshift-kube-storage-version-migrator-operator/config: failed to sync configmap cache: timed out waiting for the condition Nov 28 13:21:21 crc kubenswrapper[4631]: E1128 13:21:21.131406 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/6fccb501-125b-4bc4-8cad-3349f332c2de-config podName:6fccb501-125b-4bc4-8cad-3349f332c2de nodeName:}" failed. No retries permitted until 2025-11-28 13:21:21.631370717 +0000 UTC m=+38.438674051 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config" (UniqueName: "kubernetes.io/configmap/6fccb501-125b-4bc4-8cad-3349f332c2de-config") pod "kube-storage-version-migrator-operator-b67b599dd-gk97s" (UID: "6fccb501-125b-4bc4-8cad-3349f332c2de") : failed to sync configmap cache: timed out waiting for the condition Nov 28 13:21:21 crc kubenswrapper[4631]: I1128 13:21:21.132573 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Nov 28 13:21:21 crc kubenswrapper[4631]: I1128 13:21:21.178791 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hpfgh\" (UniqueName: \"kubernetes.io/projected/b0690955-1897-45fe-94d0-f756117c7cb5-kube-api-access-hpfgh\") pod \"console-operator-58897d9998-qcd8n\" (UID: \"b0690955-1897-45fe-94d0-f756117c7cb5\") " pod="openshift-console-operator/console-operator-58897d9998-qcd8n" Nov 28 13:21:21 crc kubenswrapper[4631]: I1128 13:21:21.190992 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wrx26\" (UniqueName: \"kubernetes.io/projected/8e852fdc-5b82-4b51-b0e5-804a1b3be0ec-kube-api-access-wrx26\") pod \"console-f9d7485db-659jg\" (UID: \"8e852fdc-5b82-4b51-b0e5-804a1b3be0ec\") " pod="openshift-console/console-f9d7485db-659jg" Nov 28 13:21:21 crc kubenswrapper[4631]: I1128 13:21:21.195785 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Nov 28 13:21:21 crc kubenswrapper[4631]: I1128 13:21:21.214386 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Nov 28 13:21:21 crc kubenswrapper[4631]: I1128 13:21:21.233247 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Nov 28 13:21:21 crc kubenswrapper[4631]: I1128 13:21:21.252405 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Nov 28 13:21:21 crc kubenswrapper[4631]: I1128 13:21:21.273840 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Nov 28 13:21:21 crc kubenswrapper[4631]: I1128 13:21:21.295537 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Nov 28 13:21:21 crc kubenswrapper[4631]: I1128 13:21:21.313045 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Nov 28 13:21:21 crc kubenswrapper[4631]: I1128 13:21:21.333156 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Nov 28 13:21:21 crc kubenswrapper[4631]: I1128 13:21:21.367588 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-qcd8n" Nov 28 13:21:21 crc kubenswrapper[4631]: I1128 13:21:21.373483 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Nov 28 13:21:21 crc kubenswrapper[4631]: I1128 13:21:21.374060 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8s5vw\" (UniqueName: \"kubernetes.io/projected/84ff0842-5520-478f-ba3d-bc8f118ab6ef-kube-api-access-8s5vw\") pod \"machine-approver-56656f9798-xqbvh\" (UID: \"84ff0842-5520-478f-ba3d-bc8f118ab6ef\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-xqbvh" Nov 28 13:21:21 crc kubenswrapper[4631]: I1128 13:21:21.393549 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-xqbvh" Nov 28 13:21:21 crc kubenswrapper[4631]: W1128 13:21:21.415125 4631 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod84ff0842_5520_478f_ba3d_bc8f118ab6ef.slice/crio-f166fee5ca62d08b8ca3c1cc9167824848a094b5bcd4e771facdc8b5035dea0b WatchSource:0}: Error finding container f166fee5ca62d08b8ca3c1cc9167824848a094b5bcd4e771facdc8b5035dea0b: Status 404 returned error can't find the container with id f166fee5ca62d08b8ca3c1cc9167824848a094b5bcd4e771facdc8b5035dea0b Nov 28 13:21:21 crc kubenswrapper[4631]: I1128 13:21:21.416731 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5v8wq\" (UniqueName: \"kubernetes.io/projected/adf078af-a8b4-4988-89e9-7daa0a80a4bb-kube-api-access-5v8wq\") pod \"cluster-samples-operator-665b6dd947-vhmzs\" (UID: \"adf078af-a8b4-4988-89e9-7daa0a80a4bb\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-vhmzs" Nov 28 13:21:21 crc kubenswrapper[4631]: I1128 13:21:21.433065 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-659jg" Nov 28 13:21:21 crc kubenswrapper[4631]: I1128 13:21:21.447928 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-vhmzs" Nov 28 13:21:21 crc kubenswrapper[4631]: I1128 13:21:21.488152 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ksp4p\" (UniqueName: \"kubernetes.io/projected/062b3680-8e44-4ab0-8f9b-a319169b53f1-kube-api-access-ksp4p\") pod \"ingress-canary-xr5v6\" (UID: \"062b3680-8e44-4ab0-8f9b-a319169b53f1\") " pod="openshift-ingress-canary/ingress-canary-xr5v6" Nov 28 13:21:21 crc kubenswrapper[4631]: I1128 13:21:21.504125 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tgg82\" (UniqueName: \"kubernetes.io/projected/f8a4f24e-7ccd-4527-9020-2f101c1d4cb5-kube-api-access-tgg82\") pod \"multus-admission-controller-857f4d67dd-52qtd\" (UID: \"f8a4f24e-7ccd-4527-9020-2f101c1d4cb5\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-52qtd" Nov 28 13:21:21 crc kubenswrapper[4631]: I1128 13:21:21.515422 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-xr5v6" Nov 28 13:21:21 crc kubenswrapper[4631]: I1128 13:21:21.533113 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-52qtd" Nov 28 13:21:21 crc kubenswrapper[4631]: I1128 13:21:21.539640 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5pq68\" (UniqueName: \"kubernetes.io/projected/f2139e10-e6bf-4193-b242-7e061396f034-kube-api-access-5pq68\") pod \"migrator-59844c95c7-977k9\" (UID: \"f2139e10-e6bf-4193-b242-7e061396f034\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-977k9" Nov 28 13:21:21 crc kubenswrapper[4631]: I1128 13:21:21.561886 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sssnb\" (UniqueName: \"kubernetes.io/projected/1039cddf-be22-4ca1-a622-d21967f9d859-kube-api-access-sssnb\") pod \"olm-operator-6b444d44fb-bzj6n\" (UID: \"1039cddf-be22-4ca1-a622-d21967f9d859\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bzj6n" Nov 28 13:21:21 crc kubenswrapper[4631]: I1128 13:21:21.581721 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5tdhq\" (UniqueName: \"kubernetes.io/projected/6fccb501-125b-4bc4-8cad-3349f332c2de-kube-api-access-5tdhq\") pod \"kube-storage-version-migrator-operator-b67b599dd-gk97s\" (UID: \"6fccb501-125b-4bc4-8cad-3349f332c2de\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-gk97s" Nov 28 13:21:21 crc kubenswrapper[4631]: I1128 13:21:21.587707 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hfxrc\" (UniqueName: \"kubernetes.io/projected/3134d1ce-629c-4e8b-86c7-b69a24578952-kube-api-access-hfxrc\") pod \"package-server-manager-789f6589d5-mdws7\" (UID: \"3134d1ce-629c-4e8b-86c7-b69a24578952\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-mdws7" Nov 28 13:21:21 crc kubenswrapper[4631]: I1128 13:21:21.604755 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/5db3c754-f675-4ec8-b3aa-65d0528bb9ca-bound-sa-token\") pod \"ingress-operator-5b745b69d9-g8n9t\" (UID: \"5db3c754-f675-4ec8-b3aa-65d0528bb9ca\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-g8n9t" Nov 28 13:21:21 crc kubenswrapper[4631]: I1128 13:21:21.627702 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bgtn5\" (UniqueName: \"kubernetes.io/projected/9eb418ce-65b6-4da3-b282-831dd62916cf-kube-api-access-bgtn5\") pod \"marketplace-operator-79b997595-2bc4b\" (UID: \"9eb418ce-65b6-4da3-b282-831dd62916cf\") " pod="openshift-marketplace/marketplace-operator-79b997595-2bc4b" Nov 28 13:21:21 crc kubenswrapper[4631]: I1128 13:21:21.654156 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6fccb501-125b-4bc4-8cad-3349f332c2de-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-gk97s\" (UID: \"6fccb501-125b-4bc4-8cad-3349f332c2de\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-gk97s" Nov 28 13:21:21 crc kubenswrapper[4631]: I1128 13:21:21.654229 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6b363da7-b1c6-495d-a05b-7ff4d7951b3e-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-jqnpq\" (UID: \"6b363da7-b1c6-495d-a05b-7ff4d7951b3e\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-jqnpq" Nov 28 13:21:21 crc kubenswrapper[4631]: I1128 13:21:21.654320 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6b363da7-b1c6-495d-a05b-7ff4d7951b3e-config\") pod \"kube-apiserver-operator-766d6c64bb-jqnpq\" (UID: \"6b363da7-b1c6-495d-a05b-7ff4d7951b3e\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-jqnpq" Nov 28 13:21:21 crc kubenswrapper[4631]: I1128 13:21:21.654393 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6fccb501-125b-4bc4-8cad-3349f332c2de-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-gk97s\" (UID: \"6fccb501-125b-4bc4-8cad-3349f332c2de\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-gk97s" Nov 28 13:21:21 crc kubenswrapper[4631]: I1128 13:21:21.656937 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6fccb501-125b-4bc4-8cad-3349f332c2de-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-gk97s\" (UID: \"6fccb501-125b-4bc4-8cad-3349f332c2de\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-gk97s" Nov 28 13:21:21 crc kubenswrapper[4631]: I1128 13:21:21.657508 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6b363da7-b1c6-495d-a05b-7ff4d7951b3e-config\") pod \"kube-apiserver-operator-766d6c64bb-jqnpq\" (UID: \"6b363da7-b1c6-495d-a05b-7ff4d7951b3e\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-jqnpq" Nov 28 13:21:21 crc kubenswrapper[4631]: I1128 13:21:21.657723 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-grc4b\" (UniqueName: \"kubernetes.io/projected/4415ae1d-dbf4-450b-a329-6dc64b0a019f-kube-api-access-grc4b\") pod \"service-ca-9c57cc56f-xltdb\" (UID: \"4415ae1d-dbf4-450b-a329-6dc64b0a019f\") " pod="openshift-service-ca/service-ca-9c57cc56f-xltdb" Nov 28 13:21:21 crc kubenswrapper[4631]: I1128 13:21:21.666435 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6fccb501-125b-4bc4-8cad-3349f332c2de-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-gk97s\" (UID: \"6fccb501-125b-4bc4-8cad-3349f332c2de\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-gk97s" Nov 28 13:21:21 crc kubenswrapper[4631]: I1128 13:21:21.669560 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6b363da7-b1c6-495d-a05b-7ff4d7951b3e-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-jqnpq\" (UID: \"6b363da7-b1c6-495d-a05b-7ff4d7951b3e\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-jqnpq" Nov 28 13:21:21 crc kubenswrapper[4631]: I1128 13:21:21.681052 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ghbsr\" (UniqueName: \"kubernetes.io/projected/760ce379-2153-4d72-acb7-f3f3f08a9028-kube-api-access-ghbsr\") pod \"packageserver-d55dfcdfc-896hq\" (UID: \"760ce379-2153-4d72-acb7-f3f3f08a9028\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-896hq" Nov 28 13:21:21 crc kubenswrapper[4631]: I1128 13:21:21.704827 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-896hq" Nov 28 13:21:21 crc kubenswrapper[4631]: I1128 13:21:21.716753 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c7sdh\" (UniqueName: \"kubernetes.io/projected/9bc5a0ed-1419-4ba0-8478-9bcad6a19954-kube-api-access-c7sdh\") pod \"csi-hostpathplugin-r6sdp\" (UID: \"9bc5a0ed-1419-4ba0-8478-9bcad6a19954\") " pod="hostpath-provisioner/csi-hostpathplugin-r6sdp" Nov 28 13:21:21 crc kubenswrapper[4631]: I1128 13:21:21.739516 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-977k9" Nov 28 13:21:21 crc kubenswrapper[4631]: I1128 13:21:21.756608 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1a4f8da7-c099-4a8e-bffc-ca17ca74eb42-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-658pf\" (UID: \"1a4f8da7-c099-4a8e-bffc-ca17ca74eb42\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-658pf" Nov 28 13:21:21 crc kubenswrapper[4631]: I1128 13:21:21.766933 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-mdws7" Nov 28 13:21:21 crc kubenswrapper[4631]: I1128 13:21:21.767028 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l8rxn\" (UniqueName: \"kubernetes.io/projected/5db3c754-f675-4ec8-b3aa-65d0528bb9ca-kube-api-access-l8rxn\") pod \"ingress-operator-5b745b69d9-g8n9t\" (UID: \"5db3c754-f675-4ec8-b3aa-65d0528bb9ca\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-g8n9t" Nov 28 13:21:21 crc kubenswrapper[4631]: I1128 13:21:21.772941 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zlnfn\" (UniqueName: \"kubernetes.io/projected/0709d8d6-5c00-4ec2-8651-df2cd64fe419-kube-api-access-zlnfn\") pod \"machine-config-server-rsxk2\" (UID: \"0709d8d6-5c00-4ec2-8651-df2cd64fe419\") " pod="openshift-machine-config-operator/machine-config-server-rsxk2" Nov 28 13:21:21 crc kubenswrapper[4631]: I1128 13:21:21.784411 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-g8n9t" Nov 28 13:21:21 crc kubenswrapper[4631]: I1128 13:21:21.785566 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/51595599-be30-46b6-bab6-ee647410efbc-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-fx9mb\" (UID: \"51595599-be30-46b6-bab6-ee647410efbc\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-fx9mb" Nov 28 13:21:21 crc kubenswrapper[4631]: I1128 13:21:21.793324 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-2bc4b" Nov 28 13:21:21 crc kubenswrapper[4631]: I1128 13:21:21.801629 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-flzfb\" (UniqueName: \"kubernetes.io/projected/2eddcce3-58a6-45f9-b924-30d95fa897ae-kube-api-access-flzfb\") pod \"machine-config-operator-74547568cd-5h9cj\" (UID: \"2eddcce3-58a6-45f9-b924-30d95fa897ae\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-5h9cj" Nov 28 13:21:21 crc kubenswrapper[4631]: I1128 13:21:21.806099 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b4cst\" (UniqueName: \"kubernetes.io/projected/6fe798f2-84e0-4019-a75c-0ee2f2e672f5-kube-api-access-b4cst\") pod \"dns-operator-744455d44c-82qbh\" (UID: \"6fe798f2-84e0-4019-a75c-0ee2f2e672f5\") " pod="openshift-dns-operator/dns-operator-744455d44c-82qbh" Nov 28 13:21:21 crc kubenswrapper[4631]: I1128 13:21:21.821726 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pblk2\" (UniqueName: \"kubernetes.io/projected/d5d77be3-1836-4c8b-9fc4-ab34196461c5-kube-api-access-pblk2\") pod \"service-ca-operator-777779d784-qdxf6\" (UID: \"d5d77be3-1836-4c8b-9fc4-ab34196461c5\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-qdxf6" Nov 28 13:21:21 crc kubenswrapper[4631]: I1128 13:21:21.822789 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-qdxf6" Nov 28 13:21:21 crc kubenswrapper[4631]: I1128 13:21:21.835542 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-gk97s" Nov 28 13:21:21 crc kubenswrapper[4631]: I1128 13:21:21.842475 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bzj6n" Nov 28 13:21:21 crc kubenswrapper[4631]: I1128 13:21:21.847551 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qfstd\" (UniqueName: \"kubernetes.io/projected/0e2f0bc1-2261-484e-8f4b-34bddacda1bc-kube-api-access-qfstd\") pod \"catalog-operator-68c6474976-hzvzz\" (UID: \"0e2f0bc1-2261-484e-8f4b-34bddacda1bc\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-hzvzz" Nov 28 13:21:21 crc kubenswrapper[4631]: I1128 13:21:21.851335 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-82qbh" Nov 28 13:21:21 crc kubenswrapper[4631]: I1128 13:21:21.866540 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-xltdb" Nov 28 13:21:21 crc kubenswrapper[4631]: I1128 13:21:21.870501 4631 request.go:700] Waited for 1.751539692s due to client-side throttling, not priority and fairness, request: POST:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-dns/serviceaccounts/dns/token Nov 28 13:21:21 crc kubenswrapper[4631]: I1128 13:21:21.872362 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-qcd8n"] Nov 28 13:21:21 crc kubenswrapper[4631]: I1128 13:21:21.891917 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gwqqt\" (UniqueName: \"kubernetes.io/projected/2a2f7727-f05e-4deb-8eba-cb3ec3561621-kube-api-access-gwqqt\") pod \"control-plane-machine-set-operator-78cbb6b69f-2v5d6\" (UID: \"2a2f7727-f05e-4deb-8eba-cb3ec3561621\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-2v5d6" Nov 28 13:21:21 crc kubenswrapper[4631]: I1128 13:21:21.908586 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4lkgw\" (UniqueName: \"kubernetes.io/projected/089e1482-6475-40b3-a005-c151f4bb0a87-kube-api-access-4lkgw\") pod \"dns-default-cg4zv\" (UID: \"089e1482-6475-40b3-a005-c151f4bb0a87\") " pod="openshift-dns/dns-default-cg4zv" Nov 28 13:21:21 crc kubenswrapper[4631]: I1128 13:21:21.913339 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b54w6\" (UniqueName: \"kubernetes.io/projected/76a05ca6-a98c-4313-855c-fd6ffaceee4f-kube-api-access-b54w6\") pod \"collect-profiles-29405595-2v6fk\" (UID: \"76a05ca6-a98c-4313-855c-fd6ffaceee4f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405595-2v6fk" Nov 28 13:21:21 crc kubenswrapper[4631]: I1128 13:21:21.918220 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5stxb\" (UniqueName: \"kubernetes.io/projected/73b80803-714a-446e-9df2-6e3ccf5430ee-kube-api-access-5stxb\") pod \"machine-config-controller-84d6567774-dxzdh\" (UID: \"73b80803-714a-446e-9df2-6e3ccf5430ee\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-dxzdh" Nov 28 13:21:21 crc kubenswrapper[4631]: I1128 13:21:21.919627 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-rsxk2" Nov 28 13:21:21 crc kubenswrapper[4631]: I1128 13:21:21.922519 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-r6sdp" Nov 28 13:21:21 crc kubenswrapper[4631]: I1128 13:21:21.942585 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-658pf" Nov 28 13:21:21 crc kubenswrapper[4631]: I1128 13:21:21.966610 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-2v5d6" Nov 28 13:21:21 crc kubenswrapper[4631]: I1128 13:21:21.967392 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/6b363da7-b1c6-495d-a05b-7ff4d7951b3e-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-jqnpq\" (UID: \"6b363da7-b1c6-495d-a05b-7ff4d7951b3e\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-jqnpq" Nov 28 13:21:21 crc kubenswrapper[4631]: I1128 13:21:21.985751 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Nov 28 13:21:21 crc kubenswrapper[4631]: I1128 13:21:21.987371 4631 generic.go:334] "Generic (PLEG): container finished" podID="aa813d64-b22e-40a0-b72a-b4e220e85f5a" containerID="065587ec10bf1a1bdcdc93c709cd204597405a1e8ae23c26304949b6ca10a44b" exitCode=0 Nov 28 13:21:21 crc kubenswrapper[4631]: I1128 13:21:21.988086 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8tbnv" event={"ID":"aa813d64-b22e-40a0-b72a-b4e220e85f5a","Type":"ContainerDied","Data":"065587ec10bf1a1bdcdc93c709cd204597405a1e8ae23c26304949b6ca10a44b"} Nov 28 13:21:21 crc kubenswrapper[4631]: I1128 13:21:21.992242 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gw52s\" (UniqueName: \"kubernetes.io/projected/22010e1d-3ec1-4564-a362-bc963f691f1f-kube-api-access-gw52s\") pod \"router-default-5444994796-jcnpq\" (UID: \"22010e1d-3ec1-4564-a362-bc963f691f1f\") " pod="openshift-ingress/router-default-5444994796-jcnpq" Nov 28 13:21:21 crc kubenswrapper[4631]: I1128 13:21:21.992622 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Nov 28 13:21:21 crc kubenswrapper[4631]: I1128 13:21:21.997572 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j9wzh\" (UniqueName: \"kubernetes.io/projected/9504d491-6c27-4387-b6da-fa0192cc7a05-kube-api-access-j9wzh\") pod \"cni-sysctl-allowlist-ds-44pwx\" (UID: \"9504d491-6c27-4387-b6da-fa0192cc7a05\") " pod="openshift-multus/cni-sysctl-allowlist-ds-44pwx" Nov 28 13:21:22 crc kubenswrapper[4631]: I1128 13:21:22.011703 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/cni-sysctl-allowlist-ds-44pwx" Nov 28 13:21:22 crc kubenswrapper[4631]: I1128 13:21:22.017729 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Nov 28 13:21:22 crc kubenswrapper[4631]: I1128 13:21:22.018107 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-fx9mb" Nov 28 13:21:22 crc kubenswrapper[4631]: I1128 13:21:22.026924 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-7977d" event={"ID":"ad9ea570-0543-4731-bda7-9fdc4f3f0ef9","Type":"ContainerStarted","Data":"7ecf66e209fc478f10b388559f36e9d8ec0dd0fcad290e794885cab01fad531e"} Nov 28 13:21:22 crc kubenswrapper[4631]: I1128 13:21:22.026977 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-7977d" event={"ID":"ad9ea570-0543-4731-bda7-9fdc4f3f0ef9","Type":"ContainerStarted","Data":"b545010db7fbbf223688299cda546f5bb1bdbed5fa68896f45d4cec1edb4435a"} Nov 28 13:21:22 crc kubenswrapper[4631]: I1128 13:21:22.038099 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Nov 28 13:21:22 crc kubenswrapper[4631]: I1128 13:21:22.054201 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-659jg"] Nov 28 13:21:22 crc kubenswrapper[4631]: I1128 13:21:22.062367 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-5h9cj" Nov 28 13:21:22 crc kubenswrapper[4631]: I1128 13:21:22.062501 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Nov 28 13:21:22 crc kubenswrapper[4631]: I1128 13:21:22.063818 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-hgsfk" event={"ID":"145c4da2-bb50-4900-84d7-974b0cfaa4d8","Type":"ContainerStarted","Data":"ebf95e54de61f41d4a27cf2da056c3eb14c5d892726916d1ad9ca6b1ed823a0f"} Nov 28 13:21:22 crc kubenswrapper[4631]: I1128 13:21:22.064344 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-hgsfk" Nov 28 13:21:22 crc kubenswrapper[4631]: I1128 13:21:22.075490 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Nov 28 13:21:22 crc kubenswrapper[4631]: I1128 13:21:22.075685 4631 generic.go:334] "Generic (PLEG): container finished" podID="cbd1163d-84ad-427f-b270-787520942822" containerID="4d5738052818540a6e9ed846dcbd9ee6b39e2181cf032c7918a2b7b0905052c1" exitCode=0 Nov 28 13:21:22 crc kubenswrapper[4631]: I1128 13:21:22.075739 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-rwpv9" event={"ID":"cbd1163d-84ad-427f-b270-787520942822","Type":"ContainerDied","Data":"4d5738052818540a6e9ed846dcbd9ee6b39e2181cf032c7918a2b7b0905052c1"} Nov 28 13:21:22 crc kubenswrapper[4631]: I1128 13:21:22.080849 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-jcnpq" Nov 28 13:21:22 crc kubenswrapper[4631]: I1128 13:21:22.109676 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-r5b6b" event={"ID":"6947cfe2-0015-42bd-b68e-465e50d41f40","Type":"ContainerStarted","Data":"a045f364b5f68637dd1bad53d57fecbc57e68ebbadc4497286c698f89be09843"} Nov 28 13:21:22 crc kubenswrapper[4631]: I1128 13:21:22.109716 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-r5b6b" event={"ID":"6947cfe2-0015-42bd-b68e-465e50d41f40","Type":"ContainerStarted","Data":"b23acd96da56591cfcc81cdd33692b3eb6388e78ff665b00511a327099cbab15"} Nov 28 13:21:22 crc kubenswrapper[4631]: I1128 13:21:22.128887 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-2ch2b" event={"ID":"889c117d-ea46-4f02-a2e0-42a47c6e4683","Type":"ContainerStarted","Data":"5aac09560f3f8c7df1547ff4f0496ff3a6afe53f1a8cc3fd913e0ed519edf5aa"} Nov 28 13:21:22 crc kubenswrapper[4631]: I1128 13:21:22.128927 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-2ch2b" event={"ID":"889c117d-ea46-4f02-a2e0-42a47c6e4683","Type":"ContainerStarted","Data":"1efe1167b033f9663a88b9ed4baf6b1459a874416ea8e4895fa1734f8a61ea72"} Nov 28 13:21:22 crc kubenswrapper[4631]: I1128 13:21:22.149309 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-vst8c" event={"ID":"af9566e8-f912-41f1-b694-1d9d6d15e91f","Type":"ContainerStarted","Data":"2cfb56622ad37b8cd67ef28ca346df1451cecd2d3ac5e96625e8f18893bf5087"} Nov 28 13:21:22 crc kubenswrapper[4631]: I1128 13:21:22.149349 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-vst8c" event={"ID":"af9566e8-f912-41f1-b694-1d9d6d15e91f","Type":"ContainerStarted","Data":"5a4da67f250e0fac87dd6b4019ee1bc257a4de2aeb883e5868d7daeb3197e691"} Nov 28 13:21:22 crc kubenswrapper[4631]: I1128 13:21:22.152524 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-bf7bx" event={"ID":"d1b540aa-a7ac-4f13-b430-91e1328d6602","Type":"ContainerStarted","Data":"705f6f4a412c59bfd038a8992d52a2d4af677d53af121bf65d3e5c101154eb0a"} Nov 28 13:21:22 crc kubenswrapper[4631]: I1128 13:21:22.153776 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ddx6v" event={"ID":"594fba15-6b93-4f34-af64-403a1b2d694c","Type":"ContainerStarted","Data":"e27b111a3dc6ae2d3a462faffe6d2e2109a2746588eb8eb6745f2f3c59669f45"} Nov 28 13:21:22 crc kubenswrapper[4631]: I1128 13:21:22.154311 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ddx6v" Nov 28 13:21:22 crc kubenswrapper[4631]: I1128 13:21:22.155453 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-xqbvh" event={"ID":"84ff0842-5520-478f-ba3d-bc8f118ab6ef","Type":"ContainerStarted","Data":"f166fee5ca62d08b8ca3c1cc9167824848a094b5bcd4e771facdc8b5035dea0b"} Nov 28 13:21:22 crc kubenswrapper[4631]: I1128 13:21:22.156335 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-c2rks" event={"ID":"feeda081-8bb0-4a89-ae6d-2e87f8f73344","Type":"ContainerStarted","Data":"cd16326d59a7da2867ae9ef65db1c1178be3b3d5d4b8445a3e855638694a0860"} Nov 28 13:21:22 crc kubenswrapper[4631]: I1128 13:21:22.156707 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-c2rks" Nov 28 13:21:22 crc kubenswrapper[4631]: I1128 13:21:22.164613 4631 patch_prober.go:28] interesting pod/downloads-7954f5f757-c2rks container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.7:8080/\": dial tcp 10.217.0.7:8080: connect: connection refused" start-of-body= Nov 28 13:21:22 crc kubenswrapper[4631]: I1128 13:21:22.164657 4631 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-c2rks" podUID="feeda081-8bb0-4a89-ae6d-2e87f8f73344" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.7:8080/\": dial tcp 10.217.0.7:8080: connect: connection refused" Nov 28 13:21:22 crc kubenswrapper[4631]: I1128 13:21:22.173718 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ddx6v" Nov 28 13:21:22 crc kubenswrapper[4631]: I1128 13:21:22.174841 4631 generic.go:334] "Generic (PLEG): container finished" podID="5e56718f-0cda-4524-aa3a-068bc2e7d569" containerID="482fc8e88fd5fb115ee71bef58594c4c645e4e86eae52feee3b91a86bd19212c" exitCode=0 Nov 28 13:21:22 crc kubenswrapper[4631]: I1128 13:21:22.174948 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-jgx6f" event={"ID":"5e56718f-0cda-4524-aa3a-068bc2e7d569","Type":"ContainerDied","Data":"482fc8e88fd5fb115ee71bef58594c4c645e4e86eae52feee3b91a86bd19212c"} Nov 28 13:21:22 crc kubenswrapper[4631]: I1128 13:21:22.174976 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-jgx6f" event={"ID":"5e56718f-0cda-4524-aa3a-068bc2e7d569","Type":"ContainerStarted","Data":"be86415ed3bdab31f0bf4d7ad87dba338420f9ad0da123de5409c23375b01cab"} Nov 28 13:21:22 crc kubenswrapper[4631]: I1128 13:21:22.202724 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-7gmxj" event={"ID":"e7c7bd85-4add-4bda-a4fa-71a7b75787e4","Type":"ContainerStarted","Data":"bd92fabfc49e57b88c4ce7a4b6957c3af7dfd28d76def445487a97dbd2876afc"} Nov 28 13:21:22 crc kubenswrapper[4631]: I1128 13:21:22.203016 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-7gmxj" Nov 28 13:21:22 crc kubenswrapper[4631]: I1128 13:21:22.227649 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-7gmxj" Nov 28 13:21:22 crc kubenswrapper[4631]: I1128 13:21:22.322958 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29405595-2v6fk" Nov 28 13:21:22 crc kubenswrapper[4631]: I1128 13:21:22.323795 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-hzvzz" Nov 28 13:21:22 crc kubenswrapper[4631]: I1128 13:21:22.323977 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-jqnpq" Nov 28 13:21:22 crc kubenswrapper[4631]: I1128 13:21:22.324553 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-cg4zv" Nov 28 13:21:22 crc kubenswrapper[4631]: I1128 13:21:22.324616 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-dxzdh" Nov 28 13:21:22 crc kubenswrapper[4631]: I1128 13:21:22.325718 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2ef91b69-6215-495f-ac26-23a99f34dcc1-serving-cert\") pod \"etcd-operator-b45778765-s7zwj\" (UID: \"2ef91b69-6215-495f-ac26-23a99f34dcc1\") " pod="openshift-etcd-operator/etcd-operator-b45778765-s7zwj" Nov 28 13:21:22 crc kubenswrapper[4631]: I1128 13:21:22.325798 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/2ef91b69-6215-495f-ac26-23a99f34dcc1-etcd-service-ca\") pod \"etcd-operator-b45778765-s7zwj\" (UID: \"2ef91b69-6215-495f-ac26-23a99f34dcc1\") " pod="openshift-etcd-operator/etcd-operator-b45778765-s7zwj" Nov 28 13:21:22 crc kubenswrapper[4631]: I1128 13:21:22.325876 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/14717577-3fab-482b-bdbe-0e604d23d6d8-installation-pull-secrets\") pod \"image-registry-697d97f7c8-5v5xp\" (UID: \"14717577-3fab-482b-bdbe-0e604d23d6d8\") " pod="openshift-image-registry/image-registry-697d97f7c8-5v5xp" Nov 28 13:21:22 crc kubenswrapper[4631]: I1128 13:21:22.325892 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/14717577-3fab-482b-bdbe-0e604d23d6d8-bound-sa-token\") pod \"image-registry-697d97f7c8-5v5xp\" (UID: \"14717577-3fab-482b-bdbe-0e604d23d6d8\") " pod="openshift-image-registry/image-registry-697d97f7c8-5v5xp" Nov 28 13:21:22 crc kubenswrapper[4631]: I1128 13:21:22.325932 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lqvsn\" (UniqueName: \"kubernetes.io/projected/14717577-3fab-482b-bdbe-0e604d23d6d8-kube-api-access-lqvsn\") pod \"image-registry-697d97f7c8-5v5xp\" (UID: \"14717577-3fab-482b-bdbe-0e604d23d6d8\") " pod="openshift-image-registry/image-registry-697d97f7c8-5v5xp" Nov 28 13:21:22 crc kubenswrapper[4631]: I1128 13:21:22.325952 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/14717577-3fab-482b-bdbe-0e604d23d6d8-registry-tls\") pod \"image-registry-697d97f7c8-5v5xp\" (UID: \"14717577-3fab-482b-bdbe-0e604d23d6d8\") " pod="openshift-image-registry/image-registry-697d97f7c8-5v5xp" Nov 28 13:21:22 crc kubenswrapper[4631]: I1128 13:21:22.325977 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/2ef91b69-6215-495f-ac26-23a99f34dcc1-etcd-ca\") pod \"etcd-operator-b45778765-s7zwj\" (UID: \"2ef91b69-6215-495f-ac26-23a99f34dcc1\") " pod="openshift-etcd-operator/etcd-operator-b45778765-s7zwj" Nov 28 13:21:22 crc kubenswrapper[4631]: I1128 13:21:22.325995 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n67lv\" (UniqueName: \"kubernetes.io/projected/2ef91b69-6215-495f-ac26-23a99f34dcc1-kube-api-access-n67lv\") pod \"etcd-operator-b45778765-s7zwj\" (UID: \"2ef91b69-6215-495f-ac26-23a99f34dcc1\") " pod="openshift-etcd-operator/etcd-operator-b45778765-s7zwj" Nov 28 13:21:22 crc kubenswrapper[4631]: I1128 13:21:22.326018 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/14717577-3fab-482b-bdbe-0e604d23d6d8-trusted-ca\") pod \"image-registry-697d97f7c8-5v5xp\" (UID: \"14717577-3fab-482b-bdbe-0e604d23d6d8\") " pod="openshift-image-registry/image-registry-697d97f7c8-5v5xp" Nov 28 13:21:22 crc kubenswrapper[4631]: I1128 13:21:22.326046 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/14717577-3fab-482b-bdbe-0e604d23d6d8-ca-trust-extracted\") pod \"image-registry-697d97f7c8-5v5xp\" (UID: \"14717577-3fab-482b-bdbe-0e604d23d6d8\") " pod="openshift-image-registry/image-registry-697d97f7c8-5v5xp" Nov 28 13:21:22 crc kubenswrapper[4631]: I1128 13:21:22.326062 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/14717577-3fab-482b-bdbe-0e604d23d6d8-registry-certificates\") pod \"image-registry-697d97f7c8-5v5xp\" (UID: \"14717577-3fab-482b-bdbe-0e604d23d6d8\") " pod="openshift-image-registry/image-registry-697d97f7c8-5v5xp" Nov 28 13:21:22 crc kubenswrapper[4631]: I1128 13:21:22.326117 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5v5xp\" (UID: \"14717577-3fab-482b-bdbe-0e604d23d6d8\") " pod="openshift-image-registry/image-registry-697d97f7c8-5v5xp" Nov 28 13:21:22 crc kubenswrapper[4631]: I1128 13:21:22.326134 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/2ef91b69-6215-495f-ac26-23a99f34dcc1-etcd-client\") pod \"etcd-operator-b45778765-s7zwj\" (UID: \"2ef91b69-6215-495f-ac26-23a99f34dcc1\") " pod="openshift-etcd-operator/etcd-operator-b45778765-s7zwj" Nov 28 13:21:22 crc kubenswrapper[4631]: I1128 13:21:22.326150 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2ef91b69-6215-495f-ac26-23a99f34dcc1-config\") pod \"etcd-operator-b45778765-s7zwj\" (UID: \"2ef91b69-6215-495f-ac26-23a99f34dcc1\") " pod="openshift-etcd-operator/etcd-operator-b45778765-s7zwj" Nov 28 13:21:22 crc kubenswrapper[4631]: E1128 13:21:22.326762 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 13:21:22.826743809 +0000 UTC m=+39.634047243 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5v5xp" (UID: "14717577-3fab-482b-bdbe-0e604d23d6d8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 13:21:22 crc kubenswrapper[4631]: I1128 13:21:22.348312 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-977k9"] Nov 28 13:21:22 crc kubenswrapper[4631]: I1128 13:21:22.447964 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-vhmzs"] Nov 28 13:21:22 crc kubenswrapper[4631]: I1128 13:21:22.451374 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 13:21:22 crc kubenswrapper[4631]: E1128 13:21:22.481377 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 13:21:22.981352975 +0000 UTC m=+39.788656319 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 13:21:22 crc kubenswrapper[4631]: I1128 13:21:22.482368 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lqvsn\" (UniqueName: \"kubernetes.io/projected/14717577-3fab-482b-bdbe-0e604d23d6d8-kube-api-access-lqvsn\") pod \"image-registry-697d97f7c8-5v5xp\" (UID: \"14717577-3fab-482b-bdbe-0e604d23d6d8\") " pod="openshift-image-registry/image-registry-697d97f7c8-5v5xp" Nov 28 13:21:22 crc kubenswrapper[4631]: I1128 13:21:22.482423 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/14717577-3fab-482b-bdbe-0e604d23d6d8-registry-tls\") pod \"image-registry-697d97f7c8-5v5xp\" (UID: \"14717577-3fab-482b-bdbe-0e604d23d6d8\") " pod="openshift-image-registry/image-registry-697d97f7c8-5v5xp" Nov 28 13:21:22 crc kubenswrapper[4631]: I1128 13:21:22.482544 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/2ef91b69-6215-495f-ac26-23a99f34dcc1-etcd-ca\") pod \"etcd-operator-b45778765-s7zwj\" (UID: \"2ef91b69-6215-495f-ac26-23a99f34dcc1\") " pod="openshift-etcd-operator/etcd-operator-b45778765-s7zwj" Nov 28 13:21:22 crc kubenswrapper[4631]: I1128 13:21:22.482571 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n67lv\" (UniqueName: \"kubernetes.io/projected/2ef91b69-6215-495f-ac26-23a99f34dcc1-kube-api-access-n67lv\") pod \"etcd-operator-b45778765-s7zwj\" (UID: \"2ef91b69-6215-495f-ac26-23a99f34dcc1\") " pod="openshift-etcd-operator/etcd-operator-b45778765-s7zwj" Nov 28 13:21:22 crc kubenswrapper[4631]: I1128 13:21:22.482723 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/14717577-3fab-482b-bdbe-0e604d23d6d8-trusted-ca\") pod \"image-registry-697d97f7c8-5v5xp\" (UID: \"14717577-3fab-482b-bdbe-0e604d23d6d8\") " pod="openshift-image-registry/image-registry-697d97f7c8-5v5xp" Nov 28 13:21:22 crc kubenswrapper[4631]: I1128 13:21:22.482813 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/14717577-3fab-482b-bdbe-0e604d23d6d8-ca-trust-extracted\") pod \"image-registry-697d97f7c8-5v5xp\" (UID: \"14717577-3fab-482b-bdbe-0e604d23d6d8\") " pod="openshift-image-registry/image-registry-697d97f7c8-5v5xp" Nov 28 13:21:22 crc kubenswrapper[4631]: I1128 13:21:22.482838 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/14717577-3fab-482b-bdbe-0e604d23d6d8-registry-certificates\") pod \"image-registry-697d97f7c8-5v5xp\" (UID: \"14717577-3fab-482b-bdbe-0e604d23d6d8\") " pod="openshift-image-registry/image-registry-697d97f7c8-5v5xp" Nov 28 13:21:22 crc kubenswrapper[4631]: I1128 13:21:22.483051 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5v5xp\" (UID: \"14717577-3fab-482b-bdbe-0e604d23d6d8\") " pod="openshift-image-registry/image-registry-697d97f7c8-5v5xp" Nov 28 13:21:22 crc kubenswrapper[4631]: I1128 13:21:22.483094 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/2ef91b69-6215-495f-ac26-23a99f34dcc1-etcd-client\") pod \"etcd-operator-b45778765-s7zwj\" (UID: \"2ef91b69-6215-495f-ac26-23a99f34dcc1\") " pod="openshift-etcd-operator/etcd-operator-b45778765-s7zwj" Nov 28 13:21:22 crc kubenswrapper[4631]: I1128 13:21:22.483178 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2ef91b69-6215-495f-ac26-23a99f34dcc1-config\") pod \"etcd-operator-b45778765-s7zwj\" (UID: \"2ef91b69-6215-495f-ac26-23a99f34dcc1\") " pod="openshift-etcd-operator/etcd-operator-b45778765-s7zwj" Nov 28 13:21:22 crc kubenswrapper[4631]: I1128 13:21:22.483394 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2ef91b69-6215-495f-ac26-23a99f34dcc1-serving-cert\") pod \"etcd-operator-b45778765-s7zwj\" (UID: \"2ef91b69-6215-495f-ac26-23a99f34dcc1\") " pod="openshift-etcd-operator/etcd-operator-b45778765-s7zwj" Nov 28 13:21:22 crc kubenswrapper[4631]: I1128 13:21:22.483826 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/2ef91b69-6215-495f-ac26-23a99f34dcc1-etcd-service-ca\") pod \"etcd-operator-b45778765-s7zwj\" (UID: \"2ef91b69-6215-495f-ac26-23a99f34dcc1\") " pod="openshift-etcd-operator/etcd-operator-b45778765-s7zwj" Nov 28 13:21:22 crc kubenswrapper[4631]: I1128 13:21:22.484272 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/14717577-3fab-482b-bdbe-0e604d23d6d8-bound-sa-token\") pod \"image-registry-697d97f7c8-5v5xp\" (UID: \"14717577-3fab-482b-bdbe-0e604d23d6d8\") " pod="openshift-image-registry/image-registry-697d97f7c8-5v5xp" Nov 28 13:21:22 crc kubenswrapper[4631]: I1128 13:21:22.484327 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/14717577-3fab-482b-bdbe-0e604d23d6d8-installation-pull-secrets\") pod \"image-registry-697d97f7c8-5v5xp\" (UID: \"14717577-3fab-482b-bdbe-0e604d23d6d8\") " pod="openshift-image-registry/image-registry-697d97f7c8-5v5xp" Nov 28 13:21:22 crc kubenswrapper[4631]: I1128 13:21:22.491549 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2ef91b69-6215-495f-ac26-23a99f34dcc1-config\") pod \"etcd-operator-b45778765-s7zwj\" (UID: \"2ef91b69-6215-495f-ac26-23a99f34dcc1\") " pod="openshift-etcd-operator/etcd-operator-b45778765-s7zwj" Nov 28 13:21:22 crc kubenswrapper[4631]: I1128 13:21:22.494765 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/2ef91b69-6215-495f-ac26-23a99f34dcc1-etcd-service-ca\") pod \"etcd-operator-b45778765-s7zwj\" (UID: \"2ef91b69-6215-495f-ac26-23a99f34dcc1\") " pod="openshift-etcd-operator/etcd-operator-b45778765-s7zwj" Nov 28 13:21:22 crc kubenswrapper[4631]: I1128 13:21:22.499209 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/14717577-3fab-482b-bdbe-0e604d23d6d8-trusted-ca\") pod \"image-registry-697d97f7c8-5v5xp\" (UID: \"14717577-3fab-482b-bdbe-0e604d23d6d8\") " pod="openshift-image-registry/image-registry-697d97f7c8-5v5xp" Nov 28 13:21:22 crc kubenswrapper[4631]: I1128 13:21:22.500327 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/14717577-3fab-482b-bdbe-0e604d23d6d8-ca-trust-extracted\") pod \"image-registry-697d97f7c8-5v5xp\" (UID: \"14717577-3fab-482b-bdbe-0e604d23d6d8\") " pod="openshift-image-registry/image-registry-697d97f7c8-5v5xp" Nov 28 13:21:22 crc kubenswrapper[4631]: I1128 13:21:22.500820 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/14717577-3fab-482b-bdbe-0e604d23d6d8-registry-certificates\") pod \"image-registry-697d97f7c8-5v5xp\" (UID: \"14717577-3fab-482b-bdbe-0e604d23d6d8\") " pod="openshift-image-registry/image-registry-697d97f7c8-5v5xp" Nov 28 13:21:22 crc kubenswrapper[4631]: E1128 13:21:22.503480 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 13:21:23.003466845 +0000 UTC m=+39.810770179 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5v5xp" (UID: "14717577-3fab-482b-bdbe-0e604d23d6d8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 13:21:22 crc kubenswrapper[4631]: I1128 13:21:22.509186 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/2ef91b69-6215-495f-ac26-23a99f34dcc1-etcd-ca\") pod \"etcd-operator-b45778765-s7zwj\" (UID: \"2ef91b69-6215-495f-ac26-23a99f34dcc1\") " pod="openshift-etcd-operator/etcd-operator-b45778765-s7zwj" Nov 28 13:21:22 crc kubenswrapper[4631]: I1128 13:21:22.540823 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/2ef91b69-6215-495f-ac26-23a99f34dcc1-etcd-client\") pod \"etcd-operator-b45778765-s7zwj\" (UID: \"2ef91b69-6215-495f-ac26-23a99f34dcc1\") " pod="openshift-etcd-operator/etcd-operator-b45778765-s7zwj" Nov 28 13:21:22 crc kubenswrapper[4631]: I1128 13:21:22.542238 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lqvsn\" (UniqueName: \"kubernetes.io/projected/14717577-3fab-482b-bdbe-0e604d23d6d8-kube-api-access-lqvsn\") pod \"image-registry-697d97f7c8-5v5xp\" (UID: \"14717577-3fab-482b-bdbe-0e604d23d6d8\") " pod="openshift-image-registry/image-registry-697d97f7c8-5v5xp" Nov 28 13:21:22 crc kubenswrapper[4631]: I1128 13:21:22.546030 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n67lv\" (UniqueName: \"kubernetes.io/projected/2ef91b69-6215-495f-ac26-23a99f34dcc1-kube-api-access-n67lv\") pod \"etcd-operator-b45778765-s7zwj\" (UID: \"2ef91b69-6215-495f-ac26-23a99f34dcc1\") " pod="openshift-etcd-operator/etcd-operator-b45778765-s7zwj" Nov 28 13:21:22 crc kubenswrapper[4631]: I1128 13:21:22.547227 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/14717577-3fab-482b-bdbe-0e604d23d6d8-registry-tls\") pod \"image-registry-697d97f7c8-5v5xp\" (UID: \"14717577-3fab-482b-bdbe-0e604d23d6d8\") " pod="openshift-image-registry/image-registry-697d97f7c8-5v5xp" Nov 28 13:21:22 crc kubenswrapper[4631]: I1128 13:21:22.547871 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/14717577-3fab-482b-bdbe-0e604d23d6d8-installation-pull-secrets\") pod \"image-registry-697d97f7c8-5v5xp\" (UID: \"14717577-3fab-482b-bdbe-0e604d23d6d8\") " pod="openshift-image-registry/image-registry-697d97f7c8-5v5xp" Nov 28 13:21:22 crc kubenswrapper[4631]: I1128 13:21:22.561956 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2ef91b69-6215-495f-ac26-23a99f34dcc1-serving-cert\") pod \"etcd-operator-b45778765-s7zwj\" (UID: \"2ef91b69-6215-495f-ac26-23a99f34dcc1\") " pod="openshift-etcd-operator/etcd-operator-b45778765-s7zwj" Nov 28 13:21:22 crc kubenswrapper[4631]: I1128 13:21:22.583089 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/14717577-3fab-482b-bdbe-0e604d23d6d8-bound-sa-token\") pod \"image-registry-697d97f7c8-5v5xp\" (UID: \"14717577-3fab-482b-bdbe-0e604d23d6d8\") " pod="openshift-image-registry/image-registry-697d97f7c8-5v5xp" Nov 28 13:21:22 crc kubenswrapper[4631]: I1128 13:21:22.585943 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-xr5v6"] Nov 28 13:21:22 crc kubenswrapper[4631]: I1128 13:21:22.586208 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-52qtd"] Nov 28 13:21:22 crc kubenswrapper[4631]: I1128 13:21:22.586225 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 13:21:22 crc kubenswrapper[4631]: E1128 13:21:22.586919 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 13:21:23.086897339 +0000 UTC m=+39.894200693 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 13:21:22 crc kubenswrapper[4631]: I1128 13:21:22.691125 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5v5xp\" (UID: \"14717577-3fab-482b-bdbe-0e604d23d6d8\") " pod="openshift-image-registry/image-registry-697d97f7c8-5v5xp" Nov 28 13:21:22 crc kubenswrapper[4631]: E1128 13:21:22.691613 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 13:21:23.191599044 +0000 UTC m=+39.998902388 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5v5xp" (UID: "14717577-3fab-482b-bdbe-0e604d23d6d8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 13:21:22 crc kubenswrapper[4631]: I1128 13:21:22.722014 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-s7zwj" Nov 28 13:21:22 crc kubenswrapper[4631]: I1128 13:21:22.747401 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-c2rks" podStartSLOduration=19.747388101 podStartE2EDuration="19.747388101s" podCreationTimestamp="2025-11-28 13:21:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 13:21:22.746236374 +0000 UTC m=+39.553539718" watchObservedRunningTime="2025-11-28 13:21:22.747388101 +0000 UTC m=+39.554691445" Nov 28 13:21:22 crc kubenswrapper[4631]: I1128 13:21:22.797977 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 13:21:22 crc kubenswrapper[4631]: E1128 13:21:22.798592 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 13:21:23.298577382 +0000 UTC m=+40.105880726 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 13:21:22 crc kubenswrapper[4631]: I1128 13:21:22.854418 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-2ch2b" podStartSLOduration=18.85440121 podStartE2EDuration="18.85440121s" podCreationTimestamp="2025-11-28 13:21:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 13:21:22.853436247 +0000 UTC m=+39.660739591" watchObservedRunningTime="2025-11-28 13:21:22.85440121 +0000 UTC m=+39.661704554" Nov 28 13:21:22 crc kubenswrapper[4631]: I1128 13:21:22.900493 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5v5xp\" (UID: \"14717577-3fab-482b-bdbe-0e604d23d6d8\") " pod="openshift-image-registry/image-registry-697d97f7c8-5v5xp" Nov 28 13:21:22 crc kubenswrapper[4631]: E1128 13:21:22.900898 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 13:21:23.400885402 +0000 UTC m=+40.208188746 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5v5xp" (UID: "14717577-3fab-482b-bdbe-0e604d23d6d8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 13:21:22 crc kubenswrapper[4631]: I1128 13:21:22.947026 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ddx6v" podStartSLOduration=18.947009775 podStartE2EDuration="18.947009775s" podCreationTimestamp="2025-11-28 13:21:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 13:21:22.94460193 +0000 UTC m=+39.751905274" watchObservedRunningTime="2025-11-28 13:21:22.947009775 +0000 UTC m=+39.754313119" Nov 28 13:21:22 crc kubenswrapper[4631]: I1128 13:21:22.955142 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-mdws7"] Nov 28 13:21:22 crc kubenswrapper[4631]: I1128 13:21:22.955673 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-896hq"] Nov 28 13:21:23 crc kubenswrapper[4631]: I1128 13:21:23.001726 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 13:21:23 crc kubenswrapper[4631]: E1128 13:21:23.002311 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 13:21:23.50227753 +0000 UTC m=+40.309580874 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 13:21:23 crc kubenswrapper[4631]: I1128 13:21:23.067454 4631 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-hgsfk container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.13:6443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Nov 28 13:21:23 crc kubenswrapper[4631]: I1128 13:21:23.067759 4631 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-hgsfk" podUID="145c4da2-bb50-4900-84d7-974b0cfaa4d8" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.13:6443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Nov 28 13:21:23 crc kubenswrapper[4631]: I1128 13:21:23.106781 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5v5xp\" (UID: \"14717577-3fab-482b-bdbe-0e604d23d6d8\") " pod="openshift-image-registry/image-registry-697d97f7c8-5v5xp" Nov 28 13:21:23 crc kubenswrapper[4631]: E1128 13:21:23.107123 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 13:21:23.607112248 +0000 UTC m=+40.414415592 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5v5xp" (UID: "14717577-3fab-482b-bdbe-0e604d23d6d8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 13:21:23 crc kubenswrapper[4631]: W1128 13:21:23.161971 4631 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3134d1ce_629c_4e8b_86c7_b69a24578952.slice/crio-b0717e9464bc954a23f2dd2df69c891cee0572d64c47c9095e9148722a7bbbef WatchSource:0}: Error finding container b0717e9464bc954a23f2dd2df69c891cee0572d64c47c9095e9148722a7bbbef: Status 404 returned error can't find the container with id b0717e9464bc954a23f2dd2df69c891cee0572d64c47c9095e9148722a7bbbef Nov 28 13:21:23 crc kubenswrapper[4631]: I1128 13:21:23.187510 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-2bc4b"] Nov 28 13:21:23 crc kubenswrapper[4631]: I1128 13:21:23.187555 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-g8n9t"] Nov 28 13:21:23 crc kubenswrapper[4631]: I1128 13:21:23.210358 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 13:21:23 crc kubenswrapper[4631]: E1128 13:21:23.211014 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 13:21:23.710995164 +0000 UTC m=+40.518298508 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 13:21:23 crc kubenswrapper[4631]: I1128 13:21:23.290081 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-mdws7" event={"ID":"3134d1ce-629c-4e8b-86c7-b69a24578952","Type":"ContainerStarted","Data":"b0717e9464bc954a23f2dd2df69c891cee0572d64c47c9095e9148722a7bbbef"} Nov 28 13:21:23 crc kubenswrapper[4631]: I1128 13:21:23.312272 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5v5xp\" (UID: \"14717577-3fab-482b-bdbe-0e604d23d6d8\") " pod="openshift-image-registry/image-registry-697d97f7c8-5v5xp" Nov 28 13:21:23 crc kubenswrapper[4631]: E1128 13:21:23.312549 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 13:21:23.812526066 +0000 UTC m=+40.619829410 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5v5xp" (UID: "14717577-3fab-482b-bdbe-0e604d23d6d8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 13:21:23 crc kubenswrapper[4631]: I1128 13:21:23.324927 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-977k9" event={"ID":"f2139e10-e6bf-4193-b242-7e061396f034","Type":"ContainerStarted","Data":"a3fa6b46da2b293ef91a2e4aeb431542ea50741e4c89183ebf7921d6d988584e"} Nov 28 13:21:23 crc kubenswrapper[4631]: I1128 13:21:23.358207 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-vhmzs" event={"ID":"adf078af-a8b4-4988-89e9-7daa0a80a4bb","Type":"ContainerStarted","Data":"01239274cc2abb213298b1f43f08570bfe6f88302b37414fcc0948b0251069f8"} Nov 28 13:21:23 crc kubenswrapper[4631]: W1128 13:21:23.362266 4631 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5db3c754_f675_4ec8_b3aa_65d0528bb9ca.slice/crio-76d7c1a6225ba5c06cdf812165282d3b338986d97e073c15a47213d0c72832fc WatchSource:0}: Error finding container 76d7c1a6225ba5c06cdf812165282d3b338986d97e073c15a47213d0c72832fc: Status 404 returned error can't find the container with id 76d7c1a6225ba5c06cdf812165282d3b338986d97e073c15a47213d0c72832fc Nov 28 13:21:23 crc kubenswrapper[4631]: I1128 13:21:23.363724 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-7gmxj" podStartSLOduration=20.363709286 podStartE2EDuration="20.363709286s" podCreationTimestamp="2025-11-28 13:21:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 13:21:23.331281358 +0000 UTC m=+40.138584702" watchObservedRunningTime="2025-11-28 13:21:23.363709286 +0000 UTC m=+40.171012640" Nov 28 13:21:23 crc kubenswrapper[4631]: I1128 13:21:23.381137 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-xqbvh" event={"ID":"84ff0842-5520-478f-ba3d-bc8f118ab6ef","Type":"ContainerStarted","Data":"512420a71b28893eb20f50c310da6c7f9872b893222ffde1f1f778bf7dd58a0a"} Nov 28 13:21:23 crc kubenswrapper[4631]: I1128 13:21:23.413068 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 13:21:23 crc kubenswrapper[4631]: E1128 13:21:23.413494 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 13:21:23.913478544 +0000 UTC m=+40.720781888 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 13:21:23 crc kubenswrapper[4631]: I1128 13:21:23.444161 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-xr5v6" event={"ID":"062b3680-8e44-4ab0-8f9b-a319169b53f1","Type":"ContainerStarted","Data":"9764768b89d42085de0c413b9b00c469b033e3a80fc641677d058072c5790368"} Nov 28 13:21:23 crc kubenswrapper[4631]: I1128 13:21:23.451166 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-52qtd" event={"ID":"f8a4f24e-7ccd-4527-9020-2f101c1d4cb5","Type":"ContainerStarted","Data":"0d58fad92c8bb05f6ec281596ef6f91507a716bc2301878de7692ebce810ae61"} Nov 28 13:21:23 crc kubenswrapper[4631]: I1128 13:21:23.478832 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-rsxk2" event={"ID":"0709d8d6-5c00-4ec2-8651-df2cd64fe419","Type":"ContainerStarted","Data":"7a97bb829c585017fdaf1dd83204fac8d267c7c8147226bc29db73b2d6a45d0d"} Nov 28 13:21:23 crc kubenswrapper[4631]: I1128 13:21:23.480643 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-jcnpq" event={"ID":"22010e1d-3ec1-4564-a362-bc963f691f1f","Type":"ContainerStarted","Data":"74d15e59d110176dd383a13e82160e1046e36252f2b969abc1d9308052f8e2b9"} Nov 28 13:21:23 crc kubenswrapper[4631]: I1128 13:21:23.500770 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-659jg" event={"ID":"8e852fdc-5b82-4b51-b0e5-804a1b3be0ec","Type":"ContainerStarted","Data":"df150edf9aded98b366ef2ec156a46838176f0602de4074397fc7790e5401dcc"} Nov 28 13:21:23 crc kubenswrapper[4631]: I1128 13:21:23.518946 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5v5xp\" (UID: \"14717577-3fab-482b-bdbe-0e604d23d6d8\") " pod="openshift-image-registry/image-registry-697d97f7c8-5v5xp" Nov 28 13:21:23 crc kubenswrapper[4631]: E1128 13:21:23.519463 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 13:21:24.019445499 +0000 UTC m=+40.826748843 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5v5xp" (UID: "14717577-3fab-482b-bdbe-0e604d23d6d8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 13:21:23 crc kubenswrapper[4631]: I1128 13:21:23.570727 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-r5b6b" podStartSLOduration=20.570709871 podStartE2EDuration="20.570709871s" podCreationTimestamp="2025-11-28 13:21:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 13:21:23.566489223 +0000 UTC m=+40.373792567" watchObservedRunningTime="2025-11-28 13:21:23.570709871 +0000 UTC m=+40.378013215" Nov 28 13:21:23 crc kubenswrapper[4631]: I1128 13:21:23.589381 4631 patch_prober.go:28] interesting pod/downloads-7954f5f757-c2rks container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.7:8080/\": dial tcp 10.217.0.7:8080: connect: connection refused" start-of-body= Nov 28 13:21:23 crc kubenswrapper[4631]: I1128 13:21:23.589421 4631 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-c2rks" podUID="feeda081-8bb0-4a89-ae6d-2e87f8f73344" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.7:8080/\": dial tcp 10.217.0.7:8080: connect: connection refused" Nov 28 13:21:23 crc kubenswrapper[4631]: I1128 13:21:23.638437 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 13:21:23 crc kubenswrapper[4631]: E1128 13:21:23.641790 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 13:21:24.14177044 +0000 UTC m=+40.949073784 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 13:21:23 crc kubenswrapper[4631]: I1128 13:21:23.650839 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5v5xp\" (UID: \"14717577-3fab-482b-bdbe-0e604d23d6d8\") " pod="openshift-image-registry/image-registry-697d97f7c8-5v5xp" Nov 28 13:21:23 crc kubenswrapper[4631]: E1128 13:21:23.654109 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 13:21:24.154092545 +0000 UTC m=+40.961395889 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5v5xp" (UID: "14717577-3fab-482b-bdbe-0e604d23d6d8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 13:21:23 crc kubenswrapper[4631]: I1128 13:21:23.758559 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 13:21:23 crc kubenswrapper[4631]: E1128 13:21:23.759518 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 13:21:24.259492925 +0000 UTC m=+41.066796269 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 13:21:23 crc kubenswrapper[4631]: I1128 13:21:23.763891 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-qcd8n" event={"ID":"b0690955-1897-45fe-94d0-f756117c7cb5","Type":"ContainerStarted","Data":"cf7e769b7a80c7b44d79bb3bb2e9e1436a8aab105a8bf1b230ea6b65d7cb60e0"} Nov 28 13:21:23 crc kubenswrapper[4631]: I1128 13:21:23.763965 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-hgsfk" Nov 28 13:21:23 crc kubenswrapper[4631]: I1128 13:21:23.763977 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/cni-sysctl-allowlist-ds-44pwx" event={"ID":"9504d491-6c27-4387-b6da-fa0192cc7a05","Type":"ContainerStarted","Data":"944aa14a03b8001a1a5f339a895060dcf282c666e08145bda17fd13886b219d5"} Nov 28 13:21:23 crc kubenswrapper[4631]: I1128 13:21:23.763988 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-896hq" event={"ID":"760ce379-2153-4d72-acb7-f3f3f08a9028","Type":"ContainerStarted","Data":"e3aa2a239cbf90c4e620b09c7d98a2d605142ec66ce785338e70aff2b01c73fb"} Nov 28 13:21:23 crc kubenswrapper[4631]: I1128 13:21:23.764009 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-qdxf6"] Nov 28 13:21:23 crc kubenswrapper[4631]: I1128 13:21:23.861069 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5v5xp\" (UID: \"14717577-3fab-482b-bdbe-0e604d23d6d8\") " pod="openshift-image-registry/image-registry-697d97f7c8-5v5xp" Nov 28 13:21:23 crc kubenswrapper[4631]: E1128 13:21:23.861418 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 13:21:24.361406586 +0000 UTC m=+41.168709930 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5v5xp" (UID: "14717577-3fab-482b-bdbe-0e604d23d6d8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 13:21:23 crc kubenswrapper[4631]: I1128 13:21:23.963212 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 13:21:23 crc kubenswrapper[4631]: E1128 13:21:23.963523 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 13:21:24.463503231 +0000 UTC m=+41.270806575 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 13:21:24 crc kubenswrapper[4631]: I1128 13:21:24.067527 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5v5xp\" (UID: \"14717577-3fab-482b-bdbe-0e604d23d6d8\") " pod="openshift-image-registry/image-registry-697d97f7c8-5v5xp" Nov 28 13:21:24 crc kubenswrapper[4631]: E1128 13:21:24.067846 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 13:21:24.567834697 +0000 UTC m=+41.375138041 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5v5xp" (UID: "14717577-3fab-482b-bdbe-0e604d23d6d8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 13:21:24 crc kubenswrapper[4631]: I1128 13:21:24.102828 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-7977d" podStartSLOduration=21.102809064 podStartE2EDuration="21.102809064s" podCreationTimestamp="2025-11-28 13:21:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 13:21:24.086983979 +0000 UTC m=+40.894287323" watchObservedRunningTime="2025-11-28 13:21:24.102809064 +0000 UTC m=+40.910112408" Nov 28 13:21:24 crc kubenswrapper[4631]: I1128 13:21:24.169544 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 13:21:24 crc kubenswrapper[4631]: E1128 13:21:24.170256 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 13:21:24.670240059 +0000 UTC m=+41.477543403 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 13:21:24 crc kubenswrapper[4631]: I1128 13:21:24.206899 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-xltdb"] Nov 28 13:21:24 crc kubenswrapper[4631]: I1128 13:21:24.283261 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5v5xp\" (UID: \"14717577-3fab-482b-bdbe-0e604d23d6d8\") " pod="openshift-image-registry/image-registry-697d97f7c8-5v5xp" Nov 28 13:21:24 crc kubenswrapper[4631]: I1128 13:21:24.283253 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-bf7bx" podStartSLOduration=21.283218175000002 podStartE2EDuration="21.283218175s" podCreationTimestamp="2025-11-28 13:21:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 13:21:24.259247242 +0000 UTC m=+41.066550586" watchObservedRunningTime="2025-11-28 13:21:24.283218175 +0000 UTC m=+41.090521519" Nov 28 13:21:24 crc kubenswrapper[4631]: E1128 13:21:24.284223 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 13:21:24.784207328 +0000 UTC m=+41.591510752 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5v5xp" (UID: "14717577-3fab-482b-bdbe-0e604d23d6d8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 13:21:24 crc kubenswrapper[4631]: I1128 13:21:24.287104 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-2v5d6"] Nov 28 13:21:24 crc kubenswrapper[4631]: I1128 13:21:24.372661 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-658pf"] Nov 28 13:21:24 crc kubenswrapper[4631]: I1128 13:21:24.391784 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 13:21:24 crc kubenswrapper[4631]: E1128 13:21:24.392597 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 13:21:24.892573127 +0000 UTC m=+41.699876461 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 13:21:24 crc kubenswrapper[4631]: I1128 13:21:24.498525 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5v5xp\" (UID: \"14717577-3fab-482b-bdbe-0e604d23d6d8\") " pod="openshift-image-registry/image-registry-697d97f7c8-5v5xp" Nov 28 13:21:24 crc kubenswrapper[4631]: E1128 13:21:24.498901 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 13:21:24.998888109 +0000 UTC m=+41.806191453 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5v5xp" (UID: "14717577-3fab-482b-bdbe-0e604d23d6d8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 13:21:24 crc kubenswrapper[4631]: I1128 13:21:24.513210 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bzj6n"] Nov 28 13:21:24 crc kubenswrapper[4631]: I1128 13:21:24.516588 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-hgsfk" podStartSLOduration=21.516558257 podStartE2EDuration="21.516558257s" podCreationTimestamp="2025-11-28 13:21:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 13:21:24.503976437 +0000 UTC m=+41.311279781" watchObservedRunningTime="2025-11-28 13:21:24.516558257 +0000 UTC m=+41.323861601" Nov 28 13:21:24 crc kubenswrapper[4631]: W1128 13:21:24.555439 4631 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4415ae1d_dbf4_450b_a329_6dc64b0a019f.slice/crio-51a61ff4d6de346d4fb91b6265b4dd545d4826f5f1f0d5c52a9ebb98e0de0b35 WatchSource:0}: Error finding container 51a61ff4d6de346d4fb91b6265b4dd545d4826f5f1f0d5c52a9ebb98e0de0b35: Status 404 returned error can't find the container with id 51a61ff4d6de346d4fb91b6265b4dd545d4826f5f1f0d5c52a9ebb98e0de0b35 Nov 28 13:21:24 crc kubenswrapper[4631]: I1128 13:21:24.597251 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-gk97s"] Nov 28 13:21:24 crc kubenswrapper[4631]: I1128 13:21:24.606617 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 13:21:24 crc kubenswrapper[4631]: E1128 13:21:24.607216 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 13:21:25.107198707 +0000 UTC m=+41.914502051 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 13:21:24 crc kubenswrapper[4631]: I1128 13:21:24.613121 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-vst8c" podStartSLOduration=21.613097144 podStartE2EDuration="21.613097144s" podCreationTimestamp="2025-11-28 13:21:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 13:21:24.613100024 +0000 UTC m=+41.420403368" watchObservedRunningTime="2025-11-28 13:21:24.613097144 +0000 UTC m=+41.420400488" Nov 28 13:21:24 crc kubenswrapper[4631]: I1128 13:21:24.717516 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5v5xp\" (UID: \"14717577-3fab-482b-bdbe-0e604d23d6d8\") " pod="openshift-image-registry/image-registry-697d97f7c8-5v5xp" Nov 28 13:21:24 crc kubenswrapper[4631]: E1128 13:21:24.718194 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 13:21:25.218181358 +0000 UTC m=+42.025484702 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5v5xp" (UID: "14717577-3fab-482b-bdbe-0e604d23d6d8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 13:21:24 crc kubenswrapper[4631]: I1128 13:21:24.800970 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-659jg" event={"ID":"8e852fdc-5b82-4b51-b0e5-804a1b3be0ec","Type":"ContainerStarted","Data":"6d1ef439fe58f9af39a531e0eaf632db22a7ed41276b202ee16433099c70ecc3"} Nov 28 13:21:24 crc kubenswrapper[4631]: I1128 13:21:24.821569 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 13:21:24 crc kubenswrapper[4631]: E1128 13:21:24.822177 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 13:21:25.322153886 +0000 UTC m=+42.129457230 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 13:21:24 crc kubenswrapper[4631]: I1128 13:21:24.856617 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-xr5v6" event={"ID":"062b3680-8e44-4ab0-8f9b-a319169b53f1","Type":"ContainerStarted","Data":"80b7c4f5570b70d5b0bcd2696fec9ca64fd5a9a480566d9f3c889e9b09ef18e7"} Nov 28 13:21:24 crc kubenswrapper[4631]: I1128 13:21:24.900547 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-xltdb" event={"ID":"4415ae1d-dbf4-450b-a329-6dc64b0a019f","Type":"ContainerStarted","Data":"51a61ff4d6de346d4fb91b6265b4dd545d4826f5f1f0d5c52a9ebb98e0de0b35"} Nov 28 13:21:24 crc kubenswrapper[4631]: I1128 13:21:24.925203 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5v5xp\" (UID: \"14717577-3fab-482b-bdbe-0e604d23d6d8\") " pod="openshift-image-registry/image-registry-697d97f7c8-5v5xp" Nov 28 13:21:24 crc kubenswrapper[4631]: E1128 13:21:24.927581 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 13:21:25.427562367 +0000 UTC m=+42.234865711 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5v5xp" (UID: "14717577-3fab-482b-bdbe-0e604d23d6d8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 13:21:24 crc kubenswrapper[4631]: I1128 13:21:24.928815 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-fx9mb"] Nov 28 13:21:24 crc kubenswrapper[4631]: I1128 13:21:24.970863 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-qdxf6" event={"ID":"d5d77be3-1836-4c8b-9fc4-ab34196461c5","Type":"ContainerStarted","Data":"5f60e9f926227ea6286415d65e3efccf0b7fbadbbbbadcfb748248de8dfecaa2"} Nov 28 13:21:25 crc kubenswrapper[4631]: I1128 13:21:25.027156 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 13:21:25 crc kubenswrapper[4631]: E1128 13:21:25.040357 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 13:21:25.540306508 +0000 UTC m=+42.347609852 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 13:21:25 crc kubenswrapper[4631]: I1128 13:21:25.042277 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-2bc4b" event={"ID":"9eb418ce-65b6-4da3-b282-831dd62916cf","Type":"ContainerStarted","Data":"da066ebca14c9bee0f45b1f08ff7c3dc5c1383c4e540419c9c9b754a8dac7dcb"} Nov 28 13:21:25 crc kubenswrapper[4631]: I1128 13:21:25.045997 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-jqnpq"] Nov 28 13:21:25 crc kubenswrapper[4631]: I1128 13:21:25.075775 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-qcd8n" event={"ID":"b0690955-1897-45fe-94d0-f756117c7cb5","Type":"ContainerStarted","Data":"1454aa3161964863febe98244ed3cc0078bbe193637ff65eed3a10b6d29b0511"} Nov 28 13:21:25 crc kubenswrapper[4631]: I1128 13:21:25.076608 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-qcd8n" Nov 28 13:21:25 crc kubenswrapper[4631]: I1128 13:21:25.125409 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-5h9cj"] Nov 28 13:21:25 crc kubenswrapper[4631]: I1128 13:21:25.135407 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-658pf" event={"ID":"1a4f8da7-c099-4a8e-bffc-ca17ca74eb42","Type":"ContainerStarted","Data":"9db3f87f02165239ba8b14954084fb738418b11d53be31944be6a267fed66815"} Nov 28 13:21:25 crc kubenswrapper[4631]: I1128 13:21:25.141814 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5v5xp\" (UID: \"14717577-3fab-482b-bdbe-0e604d23d6d8\") " pod="openshift-image-registry/image-registry-697d97f7c8-5v5xp" Nov 28 13:21:25 crc kubenswrapper[4631]: I1128 13:21:25.141896 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8716de2d-f33e-4434-9076-6345669aaff8-metrics-certs\") pod \"network-metrics-daemon-dd58r\" (UID: \"8716de2d-f33e-4434-9076-6345669aaff8\") " pod="openshift-multus/network-metrics-daemon-dd58r" Nov 28 13:21:25 crc kubenswrapper[4631]: E1128 13:21:25.143136 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 13:21:25.643122189 +0000 UTC m=+42.450425533 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5v5xp" (UID: "14717577-3fab-482b-bdbe-0e604d23d6d8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 13:21:25 crc kubenswrapper[4631]: I1128 13:21:25.168574 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-82qbh"] Nov 28 13:21:25 crc kubenswrapper[4631]: I1128 13:21:25.190321 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-g8n9t" event={"ID":"5db3c754-f675-4ec8-b3aa-65d0528bb9ca","Type":"ContainerStarted","Data":"76d7c1a6225ba5c06cdf812165282d3b338986d97e073c15a47213d0c72832fc"} Nov 28 13:21:25 crc kubenswrapper[4631]: I1128 13:21:25.199254 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-qcd8n" podStartSLOduration=22.199238683 podStartE2EDuration="22.199238683s" podCreationTimestamp="2025-11-28 13:21:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 13:21:25.189959559 +0000 UTC m=+41.997262913" watchObservedRunningTime="2025-11-28 13:21:25.199238683 +0000 UTC m=+42.006542027" Nov 28 13:21:25 crc kubenswrapper[4631]: I1128 13:21:25.201371 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8716de2d-f33e-4434-9076-6345669aaff8-metrics-certs\") pod \"network-metrics-daemon-dd58r\" (UID: \"8716de2d-f33e-4434-9076-6345669aaff8\") " pod="openshift-multus/network-metrics-daemon-dd58r" Nov 28 13:21:25 crc kubenswrapper[4631]: I1128 13:21:25.244274 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 13:21:25 crc kubenswrapper[4631]: E1128 13:21:25.244672 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 13:21:25.744657101 +0000 UTC m=+42.551960445 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 13:21:25 crc kubenswrapper[4631]: I1128 13:21:25.252164 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-896hq" Nov 28 13:21:25 crc kubenswrapper[4631]: I1128 13:21:25.266574 4631 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-896hq container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.24:5443/healthz\": dial tcp 10.217.0.24:5443: connect: connection refused" start-of-body= Nov 28 13:21:25 crc kubenswrapper[4631]: I1128 13:21:25.266700 4631 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-896hq" podUID="760ce379-2153-4d72-acb7-f3f3f08a9028" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.24:5443/healthz\": dial tcp 10.217.0.24:5443: connect: connection refused" Nov 28 13:21:25 crc kubenswrapper[4631]: I1128 13:21:25.329337 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-dd58r" Nov 28 13:21:25 crc kubenswrapper[4631]: I1128 13:21:25.347842 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5v5xp\" (UID: \"14717577-3fab-482b-bdbe-0e604d23d6d8\") " pod="openshift-image-registry/image-registry-697d97f7c8-5v5xp" Nov 28 13:21:25 crc kubenswrapper[4631]: E1128 13:21:25.350969 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 13:21:25.850952633 +0000 UTC m=+42.658255977 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5v5xp" (UID: "14717577-3fab-482b-bdbe-0e604d23d6d8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 13:21:25 crc kubenswrapper[4631]: I1128 13:21:25.367750 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-659jg" podStartSLOduration=22.36772853 podStartE2EDuration="22.36772853s" podCreationTimestamp="2025-11-28 13:21:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 13:21:25.284133621 +0000 UTC m=+42.091436995" watchObservedRunningTime="2025-11-28 13:21:25.36772853 +0000 UTC m=+42.175031864" Nov 28 13:21:25 crc kubenswrapper[4631]: I1128 13:21:25.382487 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-xr5v6" podStartSLOduration=7.382466359 podStartE2EDuration="7.382466359s" podCreationTimestamp="2025-11-28 13:21:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 13:21:25.363903811 +0000 UTC m=+42.171207155" watchObservedRunningTime="2025-11-28 13:21:25.382466359 +0000 UTC m=+42.189769703" Nov 28 13:21:25 crc kubenswrapper[4631]: I1128 13:21:25.407517 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-r6sdp"] Nov 28 13:21:25 crc kubenswrapper[4631]: I1128 13:21:25.432695 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-896hq" podStartSLOduration=21.432679937 podStartE2EDuration="21.432679937s" podCreationTimestamp="2025-11-28 13:21:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 13:21:25.430679722 +0000 UTC m=+42.237983066" watchObservedRunningTime="2025-11-28 13:21:25.432679937 +0000 UTC m=+42.239983281" Nov 28 13:21:25 crc kubenswrapper[4631]: I1128 13:21:25.449559 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 13:21:25 crc kubenswrapper[4631]: E1128 13:21:25.449877 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 13:21:25.949842463 +0000 UTC m=+42.757145807 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 13:21:25 crc kubenswrapper[4631]: I1128 13:21:25.481180 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-s7zwj"] Nov 28 13:21:25 crc kubenswrapper[4631]: I1128 13:21:25.482646 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-hzvzz"] Nov 28 13:21:25 crc kubenswrapper[4631]: I1128 13:21:25.494994 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-dxzdh"] Nov 28 13:21:25 crc kubenswrapper[4631]: I1128 13:21:25.543026 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-cg4zv"] Nov 28 13:21:25 crc kubenswrapper[4631]: I1128 13:21:25.551466 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5v5xp\" (UID: \"14717577-3fab-482b-bdbe-0e604d23d6d8\") " pod="openshift-image-registry/image-registry-697d97f7c8-5v5xp" Nov 28 13:21:25 crc kubenswrapper[4631]: E1128 13:21:25.551980 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 13:21:26.051962358 +0000 UTC m=+42.859265702 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5v5xp" (UID: "14717577-3fab-482b-bdbe-0e604d23d6d8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 13:21:25 crc kubenswrapper[4631]: I1128 13:21:25.600804 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29405595-2v6fk"] Nov 28 13:21:25 crc kubenswrapper[4631]: I1128 13:21:25.652398 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 13:21:25 crc kubenswrapper[4631]: E1128 13:21:25.652565 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 13:21:26.152537708 +0000 UTC m=+42.959841052 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 13:21:25 crc kubenswrapper[4631]: I1128 13:21:25.652819 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5v5xp\" (UID: \"14717577-3fab-482b-bdbe-0e604d23d6d8\") " pod="openshift-image-registry/image-registry-697d97f7c8-5v5xp" Nov 28 13:21:25 crc kubenswrapper[4631]: E1128 13:21:25.653170 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 13:21:26.153156232 +0000 UTC m=+42.960459576 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5v5xp" (UID: "14717577-3fab-482b-bdbe-0e604d23d6d8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 13:21:25 crc kubenswrapper[4631]: W1128 13:21:25.723511 4631 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod73b80803_714a_446e_9df2_6e3ccf5430ee.slice/crio-1848a47444ba71c094ba5757124a0ed0c43f577f6664f416f018015f41bf4d63 WatchSource:0}: Error finding container 1848a47444ba71c094ba5757124a0ed0c43f577f6664f416f018015f41bf4d63: Status 404 returned error can't find the container with id 1848a47444ba71c094ba5757124a0ed0c43f577f6664f416f018015f41bf4d63 Nov 28 13:21:25 crc kubenswrapper[4631]: I1128 13:21:25.762474 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 13:21:25 crc kubenswrapper[4631]: E1128 13:21:25.762850 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 13:21:26.262831653 +0000 UTC m=+43.070134997 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 13:21:25 crc kubenswrapper[4631]: I1128 13:21:25.865054 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5v5xp\" (UID: \"14717577-3fab-482b-bdbe-0e604d23d6d8\") " pod="openshift-image-registry/image-registry-697d97f7c8-5v5xp" Nov 28 13:21:25 crc kubenswrapper[4631]: E1128 13:21:25.865363 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 13:21:26.365352287 +0000 UTC m=+43.172655621 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5v5xp" (UID: "14717577-3fab-482b-bdbe-0e604d23d6d8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 13:21:25 crc kubenswrapper[4631]: I1128 13:21:25.966817 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 13:21:25 crc kubenswrapper[4631]: E1128 13:21:25.967162 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 13:21:26.467119495 +0000 UTC m=+43.274422839 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 13:21:25 crc kubenswrapper[4631]: I1128 13:21:25.967792 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5v5xp\" (UID: \"14717577-3fab-482b-bdbe-0e604d23d6d8\") " pod="openshift-image-registry/image-registry-697d97f7c8-5v5xp" Nov 28 13:21:25 crc kubenswrapper[4631]: E1128 13:21:25.968371 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 13:21:26.468353873 +0000 UTC m=+43.275657217 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5v5xp" (UID: "14717577-3fab-482b-bdbe-0e604d23d6d8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 13:21:26 crc kubenswrapper[4631]: I1128 13:21:26.071472 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 13:21:26 crc kubenswrapper[4631]: E1128 13:21:26.071799 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 13:21:26.571763008 +0000 UTC m=+43.379066352 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 13:21:26 crc kubenswrapper[4631]: I1128 13:21:26.079418 4631 patch_prober.go:28] interesting pod/console-operator-58897d9998-qcd8n container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.12:8443/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Nov 28 13:21:26 crc kubenswrapper[4631]: I1128 13:21:26.079510 4631 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-qcd8n" podUID="b0690955-1897-45fe-94d0-f756117c7cb5" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.12:8443/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 28 13:21:26 crc kubenswrapper[4631]: I1128 13:21:26.176713 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5v5xp\" (UID: \"14717577-3fab-482b-bdbe-0e604d23d6d8\") " pod="openshift-image-registry/image-registry-697d97f7c8-5v5xp" Nov 28 13:21:26 crc kubenswrapper[4631]: E1128 13:21:26.177721 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 13:21:26.677694521 +0000 UTC m=+43.484997865 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5v5xp" (UID: "14717577-3fab-482b-bdbe-0e604d23d6d8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 13:21:26 crc kubenswrapper[4631]: I1128 13:21:26.278418 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 13:21:26 crc kubenswrapper[4631]: E1128 13:21:26.278625 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 13:21:26.778601229 +0000 UTC m=+43.585904573 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 13:21:26 crc kubenswrapper[4631]: I1128 13:21:26.279129 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5v5xp\" (UID: \"14717577-3fab-482b-bdbe-0e604d23d6d8\") " pod="openshift-image-registry/image-registry-697d97f7c8-5v5xp" Nov 28 13:21:26 crc kubenswrapper[4631]: E1128 13:21:26.279644 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 13:21:26.779634773 +0000 UTC m=+43.586938117 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5v5xp" (UID: "14717577-3fab-482b-bdbe-0e604d23d6d8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 13:21:26 crc kubenswrapper[4631]: I1128 13:21:26.280752 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-jqnpq" event={"ID":"6b363da7-b1c6-495d-a05b-7ff4d7951b3e","Type":"ContainerStarted","Data":"dce604266cd83125a11bdbfb011aa52555c51f4ea5af545ffc952cd56d72dee8"} Nov 28 13:21:26 crc kubenswrapper[4631]: I1128 13:21:26.283539 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bzj6n" event={"ID":"1039cddf-be22-4ca1-a622-d21967f9d859","Type":"ContainerStarted","Data":"0f48a0cda4e027727f9399cfdf7ce53cc0a3189ed9e4f651fb8400b84cf5eb04"} Nov 28 13:21:26 crc kubenswrapper[4631]: I1128 13:21:26.287805 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-r6sdp" event={"ID":"9bc5a0ed-1419-4ba0-8478-9bcad6a19954","Type":"ContainerStarted","Data":"96c44f36d26b0201e7740fb82b5e705557bbe88d56961c2f085dc1b52910bd71"} Nov 28 13:21:26 crc kubenswrapper[4631]: I1128 13:21:26.290701 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-82qbh" event={"ID":"6fe798f2-84e0-4019-a75c-0ee2f2e672f5","Type":"ContainerStarted","Data":"08717d71ede3bcb9dd4d416d90856f192ddbb73bd411034644acbf8a01ccb83e"} Nov 28 13:21:26 crc kubenswrapper[4631]: I1128 13:21:26.293933 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29405595-2v6fk" event={"ID":"76a05ca6-a98c-4313-855c-fd6ffaceee4f","Type":"ContainerStarted","Data":"f75613555c90f9c4c2505bf8ec5be1a675aed747146be2b68c6f8ada4d4ba88c"} Nov 28 13:21:26 crc kubenswrapper[4631]: I1128 13:21:26.299014 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-mdws7" event={"ID":"3134d1ce-629c-4e8b-86c7-b69a24578952","Type":"ContainerStarted","Data":"cad3366ce701e3ccc7a0a6bc244ce7c970b933fbce8c21949a33618301d78d1f"} Nov 28 13:21:26 crc kubenswrapper[4631]: I1128 13:21:26.300925 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-hzvzz" event={"ID":"0e2f0bc1-2261-484e-8f4b-34bddacda1bc","Type":"ContainerStarted","Data":"79e9998ec17d706a6b7a26c83a9b275b4217a7c24995fb840f534b7fe5225bce"} Nov 28 13:21:26 crc kubenswrapper[4631]: I1128 13:21:26.302751 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-fx9mb" event={"ID":"51595599-be30-46b6-bab6-ee647410efbc","Type":"ContainerStarted","Data":"2da0173cc2c957542470209f104b9a9bf3bd25a0e3f1f7b96dca25d0b6d2448d"} Nov 28 13:21:26 crc kubenswrapper[4631]: I1128 13:21:26.303844 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-gk97s" event={"ID":"6fccb501-125b-4bc4-8cad-3349f332c2de","Type":"ContainerStarted","Data":"c92a5ebb9b91d4ced909fa14b7ca09892564929ffbaddb7b2fd5c386aa59ec67"} Nov 28 13:21:26 crc kubenswrapper[4631]: I1128 13:21:26.306734 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-977k9" event={"ID":"f2139e10-e6bf-4193-b242-7e061396f034","Type":"ContainerStarted","Data":"8040d820fdd3f26c08c90fd48b62d2f962eb6c5429fe20ce8b8fe0f2d51c6f11"} Nov 28 13:21:26 crc kubenswrapper[4631]: I1128 13:21:26.312784 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-5h9cj" event={"ID":"2eddcce3-58a6-45f9-b924-30d95fa897ae","Type":"ContainerStarted","Data":"25aa18fdfc0e1498263ceec4213bc6272fd31af16d9005c497720a4477a49702"} Nov 28 13:21:26 crc kubenswrapper[4631]: I1128 13:21:26.312842 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-dd58r"] Nov 28 13:21:26 crc kubenswrapper[4631]: I1128 13:21:26.315084 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-jcnpq" event={"ID":"22010e1d-3ec1-4564-a362-bc963f691f1f","Type":"ContainerStarted","Data":"d51f2d88f3f9bef51d696b0e7ac9608808b8630fe2118af32d1ae3bee845f989"} Nov 28 13:21:26 crc kubenswrapper[4631]: I1128 13:21:26.316393 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-2v5d6" event={"ID":"2a2f7727-f05e-4deb-8eba-cb3ec3561621","Type":"ContainerStarted","Data":"1d4371f7e6dfb2718147d1de16d59543d8ec3a5eea0df4f5575f009241136486"} Nov 28 13:21:26 crc kubenswrapper[4631]: I1128 13:21:26.318878 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-rwpv9" event={"ID":"cbd1163d-84ad-427f-b270-787520942822","Type":"ContainerStarted","Data":"8fc1d36238803225310e8b722435c76bacb63deaee84eb3dfed6fa586d00b18b"} Nov 28 13:21:26 crc kubenswrapper[4631]: I1128 13:21:26.319019 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-rwpv9" Nov 28 13:21:26 crc kubenswrapper[4631]: I1128 13:21:26.320085 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-896hq" event={"ID":"760ce379-2153-4d72-acb7-f3f3f08a9028","Type":"ContainerStarted","Data":"e202e478ef3155eb1c46b783c92e9d47723c074a64a726e16276a10f57c615de"} Nov 28 13:21:26 crc kubenswrapper[4631]: I1128 13:21:26.320733 4631 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-896hq container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.24:5443/healthz\": dial tcp 10.217.0.24:5443: connect: connection refused" start-of-body= Nov 28 13:21:26 crc kubenswrapper[4631]: I1128 13:21:26.320783 4631 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-896hq" podUID="760ce379-2153-4d72-acb7-f3f3f08a9028" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.24:5443/healthz\": dial tcp 10.217.0.24:5443: connect: connection refused" Nov 28 13:21:26 crc kubenswrapper[4631]: I1128 13:21:26.323572 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-cg4zv" event={"ID":"089e1482-6475-40b3-a005-c151f4bb0a87","Type":"ContainerStarted","Data":"79c28b75019752ee696ed7b9ec2c28fbcfeefc91f005bec9b2621eebc28c416e"} Nov 28 13:21:26 crc kubenswrapper[4631]: I1128 13:21:26.327455 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-s7zwj" event={"ID":"2ef91b69-6215-495f-ac26-23a99f34dcc1","Type":"ContainerStarted","Data":"0176a1ace175bf184bef910935062f4f9dcaebdf8dcd2dde97ad065c47504673"} Nov 28 13:21:26 crc kubenswrapper[4631]: I1128 13:21:26.335410 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-dxzdh" event={"ID":"73b80803-714a-446e-9df2-6e3ccf5430ee","Type":"ContainerStarted","Data":"1848a47444ba71c094ba5757124a0ed0c43f577f6664f416f018015f41bf4d63"} Nov 28 13:21:26 crc kubenswrapper[4631]: I1128 13:21:26.380616 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 13:21:26 crc kubenswrapper[4631]: E1128 13:21:26.380813 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 13:21:26.880787216 +0000 UTC m=+43.688090560 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 13:21:26 crc kubenswrapper[4631]: I1128 13:21:26.380935 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5v5xp\" (UID: \"14717577-3fab-482b-bdbe-0e604d23d6d8\") " pod="openshift-image-registry/image-registry-697d97f7c8-5v5xp" Nov 28 13:21:26 crc kubenswrapper[4631]: E1128 13:21:26.383068 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 13:21:26.883059688 +0000 UTC m=+43.690363032 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5v5xp" (UID: "14717577-3fab-482b-bdbe-0e604d23d6d8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 13:21:26 crc kubenswrapper[4631]: W1128 13:21:26.456197 4631 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8716de2d_f33e_4434_9076_6345669aaff8.slice/crio-ed25a29723900f96cfae660b9cf1d4f35959624a6a1c1b2dcd1df24788fba3f2 WatchSource:0}: Error finding container ed25a29723900f96cfae660b9cf1d4f35959624a6a1c1b2dcd1df24788fba3f2: Status 404 returned error can't find the container with id ed25a29723900f96cfae660b9cf1d4f35959624a6a1c1b2dcd1df24788fba3f2 Nov 28 13:21:26 crc kubenswrapper[4631]: I1128 13:21:26.482638 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 13:21:26 crc kubenswrapper[4631]: E1128 13:21:26.483524 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 13:21:26.983496704 +0000 UTC m=+43.790800048 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 13:21:26 crc kubenswrapper[4631]: I1128 13:21:26.585133 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5v5xp\" (UID: \"14717577-3fab-482b-bdbe-0e604d23d6d8\") " pod="openshift-image-registry/image-registry-697d97f7c8-5v5xp" Nov 28 13:21:26 crc kubenswrapper[4631]: E1128 13:21:26.585745 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 13:21:27.085719432 +0000 UTC m=+43.893022776 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5v5xp" (UID: "14717577-3fab-482b-bdbe-0e604d23d6d8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 13:21:26 crc kubenswrapper[4631]: I1128 13:21:26.685928 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 13:21:26 crc kubenswrapper[4631]: E1128 13:21:26.686323 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 13:21:27.186307072 +0000 UTC m=+43.993610416 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 13:21:26 crc kubenswrapper[4631]: I1128 13:21:26.785251 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-qcd8n" Nov 28 13:21:26 crc kubenswrapper[4631]: I1128 13:21:26.787751 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5v5xp\" (UID: \"14717577-3fab-482b-bdbe-0e604d23d6d8\") " pod="openshift-image-registry/image-registry-697d97f7c8-5v5xp" Nov 28 13:21:26 crc kubenswrapper[4631]: E1128 13:21:26.788080 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 13:21:27.28806801 +0000 UTC m=+44.095371354 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5v5xp" (UID: "14717577-3fab-482b-bdbe-0e604d23d6d8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 13:21:26 crc kubenswrapper[4631]: I1128 13:21:26.812141 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-rwpv9" podStartSLOduration=23.812109464 podStartE2EDuration="23.812109464s" podCreationTimestamp="2025-11-28 13:21:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 13:21:26.335660215 +0000 UTC m=+43.142963569" watchObservedRunningTime="2025-11-28 13:21:26.812109464 +0000 UTC m=+43.619412808" Nov 28 13:21:26 crc kubenswrapper[4631]: I1128 13:21:26.889386 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 13:21:26 crc kubenswrapper[4631]: E1128 13:21:26.889537 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 13:21:27.389512869 +0000 UTC m=+44.196816213 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 13:21:26 crc kubenswrapper[4631]: I1128 13:21:26.889632 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5v5xp\" (UID: \"14717577-3fab-482b-bdbe-0e604d23d6d8\") " pod="openshift-image-registry/image-registry-697d97f7c8-5v5xp" Nov 28 13:21:26 crc kubenswrapper[4631]: E1128 13:21:26.889942 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 13:21:27.389931039 +0000 UTC m=+44.197234383 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5v5xp" (UID: "14717577-3fab-482b-bdbe-0e604d23d6d8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 13:21:27 crc kubenswrapper[4631]: I1128 13:21:27.025753 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 13:21:27 crc kubenswrapper[4631]: E1128 13:21:27.026308 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 13:21:27.526282104 +0000 UTC m=+44.333585448 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 13:21:27 crc kubenswrapper[4631]: I1128 13:21:27.127649 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5v5xp\" (UID: \"14717577-3fab-482b-bdbe-0e604d23d6d8\") " pod="openshift-image-registry/image-registry-697d97f7c8-5v5xp" Nov 28 13:21:27 crc kubenswrapper[4631]: E1128 13:21:27.128038 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 13:21:27.628024791 +0000 UTC m=+44.435328135 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5v5xp" (UID: "14717577-3fab-482b-bdbe-0e604d23d6d8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 13:21:27 crc kubenswrapper[4631]: I1128 13:21:27.228240 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 13:21:27 crc kubenswrapper[4631]: E1128 13:21:27.228453 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 13:21:27.728430277 +0000 UTC m=+44.535733621 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 13:21:27 crc kubenswrapper[4631]: I1128 13:21:27.228588 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5v5xp\" (UID: \"14717577-3fab-482b-bdbe-0e604d23d6d8\") " pod="openshift-image-registry/image-registry-697d97f7c8-5v5xp" Nov 28 13:21:27 crc kubenswrapper[4631]: E1128 13:21:27.228866 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 13:21:27.728854217 +0000 UTC m=+44.536157561 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5v5xp" (UID: "14717577-3fab-482b-bdbe-0e604d23d6d8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 13:21:27 crc kubenswrapper[4631]: I1128 13:21:27.330499 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 13:21:27 crc kubenswrapper[4631]: E1128 13:21:27.330808 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 13:21:27.830766208 +0000 UTC m=+44.638069582 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 13:21:27 crc kubenswrapper[4631]: I1128 13:21:27.330927 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5v5xp\" (UID: \"14717577-3fab-482b-bdbe-0e604d23d6d8\") " pod="openshift-image-registry/image-registry-697d97f7c8-5v5xp" Nov 28 13:21:27 crc kubenswrapper[4631]: E1128 13:21:27.331450 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 13:21:27.831431482 +0000 UTC m=+44.638734826 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5v5xp" (UID: "14717577-3fab-482b-bdbe-0e604d23d6d8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 13:21:27 crc kubenswrapper[4631]: I1128 13:21:27.353207 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-g8n9t" event={"ID":"5db3c754-f675-4ec8-b3aa-65d0528bb9ca","Type":"ContainerStarted","Data":"cf4feb79018ccbdaf3b288eee779f03f889aef576a31ffdcf013436405070d4c"} Nov 28 13:21:27 crc kubenswrapper[4631]: I1128 13:21:27.356299 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-rsxk2" event={"ID":"0709d8d6-5c00-4ec2-8651-df2cd64fe419","Type":"ContainerStarted","Data":"e9e18bbf9b8d18cbb605313e38880ccb6bf7ec222477a6e2ebb1d9fe1d9d4e57"} Nov 28 13:21:27 crc kubenswrapper[4631]: I1128 13:21:27.358341 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-dd58r" event={"ID":"8716de2d-f33e-4434-9076-6345669aaff8","Type":"ContainerStarted","Data":"ed25a29723900f96cfae660b9cf1d4f35959624a6a1c1b2dcd1df24788fba3f2"} Nov 28 13:21:27 crc kubenswrapper[4631]: I1128 13:21:27.392794 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-jcnpq" podStartSLOduration=23.392756877 podStartE2EDuration="23.392756877s" podCreationTimestamp="2025-11-28 13:21:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 13:21:27.386542543 +0000 UTC m=+44.193845887" watchObservedRunningTime="2025-11-28 13:21:27.392756877 +0000 UTC m=+44.200060231" Nov 28 13:21:27 crc kubenswrapper[4631]: I1128 13:21:27.432596 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 13:21:27 crc kubenswrapper[4631]: E1128 13:21:27.432820 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 13:21:27.93279203 +0000 UTC m=+44.740095364 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 13:21:27 crc kubenswrapper[4631]: I1128 13:21:27.432950 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5v5xp\" (UID: \"14717577-3fab-482b-bdbe-0e604d23d6d8\") " pod="openshift-image-registry/image-registry-697d97f7c8-5v5xp" Nov 28 13:21:27 crc kubenswrapper[4631]: E1128 13:21:27.433804 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 13:21:27.933766083 +0000 UTC m=+44.741069467 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5v5xp" (UID: "14717577-3fab-482b-bdbe-0e604d23d6d8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 13:21:27 crc kubenswrapper[4631]: I1128 13:21:27.534502 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 13:21:27 crc kubenswrapper[4631]: E1128 13:21:27.534717 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 13:21:28.03468826 +0000 UTC m=+44.841991614 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 13:21:27 crc kubenswrapper[4631]: I1128 13:21:27.535015 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5v5xp\" (UID: \"14717577-3fab-482b-bdbe-0e604d23d6d8\") " pod="openshift-image-registry/image-registry-697d97f7c8-5v5xp" Nov 28 13:21:27 crc kubenswrapper[4631]: E1128 13:21:27.535438 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 13:21:28.035427818 +0000 UTC m=+44.842731172 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5v5xp" (UID: "14717577-3fab-482b-bdbe-0e604d23d6d8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 13:21:27 crc kubenswrapper[4631]: I1128 13:21:27.636626 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 13:21:27 crc kubenswrapper[4631]: E1128 13:21:27.637155 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 13:21:28.137137523 +0000 UTC m=+44.944440877 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 13:21:27 crc kubenswrapper[4631]: I1128 13:21:27.738608 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5v5xp\" (UID: \"14717577-3fab-482b-bdbe-0e604d23d6d8\") " pod="openshift-image-registry/image-registry-697d97f7c8-5v5xp" Nov 28 13:21:27 crc kubenswrapper[4631]: E1128 13:21:27.739132 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 13:21:28.239109325 +0000 UTC m=+45.046412709 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5v5xp" (UID: "14717577-3fab-482b-bdbe-0e604d23d6d8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 13:21:27 crc kubenswrapper[4631]: I1128 13:21:27.839938 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 13:21:27 crc kubenswrapper[4631]: E1128 13:21:27.840172 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 13:21:28.340125815 +0000 UTC m=+45.147429169 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 13:21:27 crc kubenswrapper[4631]: I1128 13:21:27.840497 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5v5xp\" (UID: \"14717577-3fab-482b-bdbe-0e604d23d6d8\") " pod="openshift-image-registry/image-registry-697d97f7c8-5v5xp" Nov 28 13:21:27 crc kubenswrapper[4631]: E1128 13:21:27.840945 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 13:21:28.340924184 +0000 UTC m=+45.148227528 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5v5xp" (UID: "14717577-3fab-482b-bdbe-0e604d23d6d8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 13:21:27 crc kubenswrapper[4631]: I1128 13:21:27.941828 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 13:21:27 crc kubenswrapper[4631]: E1128 13:21:27.942127 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 13:21:28.442111838 +0000 UTC m=+45.249415182 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 13:21:28 crc kubenswrapper[4631]: I1128 13:21:28.043185 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5v5xp\" (UID: \"14717577-3fab-482b-bdbe-0e604d23d6d8\") " pod="openshift-image-registry/image-registry-697d97f7c8-5v5xp" Nov 28 13:21:28 crc kubenswrapper[4631]: E1128 13:21:28.043528 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 13:21:28.543516587 +0000 UTC m=+45.350819931 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5v5xp" (UID: "14717577-3fab-482b-bdbe-0e604d23d6d8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 13:21:28 crc kubenswrapper[4631]: I1128 13:21:28.081466 4631 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-jcnpq" Nov 28 13:21:28 crc kubenswrapper[4631]: I1128 13:21:28.092060 4631 patch_prober.go:28] interesting pod/router-default-5444994796-jcnpq container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 28 13:21:28 crc kubenswrapper[4631]: [-]has-synced failed: reason withheld Nov 28 13:21:28 crc kubenswrapper[4631]: [+]process-running ok Nov 28 13:21:28 crc kubenswrapper[4631]: healthz check failed Nov 28 13:21:28 crc kubenswrapper[4631]: I1128 13:21:28.092311 4631 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-jcnpq" podUID="22010e1d-3ec1-4564-a362-bc963f691f1f" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 28 13:21:28 crc kubenswrapper[4631]: I1128 13:21:28.144265 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 13:21:28 crc kubenswrapper[4631]: E1128 13:21:28.144613 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 13:21:28.644598139 +0000 UTC m=+45.451901483 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 13:21:28 crc kubenswrapper[4631]: I1128 13:21:28.246102 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5v5xp\" (UID: \"14717577-3fab-482b-bdbe-0e604d23d6d8\") " pod="openshift-image-registry/image-registry-697d97f7c8-5v5xp" Nov 28 13:21:28 crc kubenswrapper[4631]: E1128 13:21:28.246745 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 13:21:28.746719074 +0000 UTC m=+45.554022418 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5v5xp" (UID: "14717577-3fab-482b-bdbe-0e604d23d6d8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 13:21:28 crc kubenswrapper[4631]: I1128 13:21:28.347485 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 13:21:28 crc kubenswrapper[4631]: E1128 13:21:28.348072 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 13:21:28.848044441 +0000 UTC m=+45.655347795 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 13:21:28 crc kubenswrapper[4631]: I1128 13:21:28.358745 4631 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-896hq container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.24:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Nov 28 13:21:28 crc kubenswrapper[4631]: I1128 13:21:28.358879 4631 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-896hq" podUID="760ce379-2153-4d72-acb7-f3f3f08a9028" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.24:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Nov 28 13:21:28 crc kubenswrapper[4631]: I1128 13:21:28.363810 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-977k9" event={"ID":"f2139e10-e6bf-4193-b242-7e061396f034","Type":"ContainerStarted","Data":"c23c86a6fc35a65bbadb6aeca06ff733ce9363402ec01862cd1b32986b048f60"} Nov 28 13:21:28 crc kubenswrapper[4631]: I1128 13:21:28.364992 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/cni-sysctl-allowlist-ds-44pwx" event={"ID":"9504d491-6c27-4387-b6da-fa0192cc7a05","Type":"ContainerStarted","Data":"3b12a8f186ea08cfb45b04b2d0d67de069fb809dcbf835a4dd4e008e4e24a4ae"} Nov 28 13:21:28 crc kubenswrapper[4631]: I1128 13:21:28.367352 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-qdxf6" event={"ID":"d5d77be3-1836-4c8b-9fc4-ab34196461c5","Type":"ContainerStarted","Data":"8358cab16c35212f0aed0295c366c6ebc7135622e9ac7c3e73c307863dbc96af"} Nov 28 13:21:28 crc kubenswrapper[4631]: I1128 13:21:28.368802 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-dxzdh" event={"ID":"73b80803-714a-446e-9df2-6e3ccf5430ee","Type":"ContainerStarted","Data":"4996f6f0b810b45400a5b192c86b1af646b121f9386dcc2a656936c0d3231f9c"} Nov 28 13:21:28 crc kubenswrapper[4631]: I1128 13:21:28.370182 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-52qtd" event={"ID":"f8a4f24e-7ccd-4527-9020-2f101c1d4cb5","Type":"ContainerStarted","Data":"273749f8439d74bf1cf460ad755a4a095992ff92be1d841eb7c2cdd9e1bc72ae"} Nov 28 13:21:28 crc kubenswrapper[4631]: I1128 13:21:28.371976 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-xqbvh" event={"ID":"84ff0842-5520-478f-ba3d-bc8f118ab6ef","Type":"ContainerStarted","Data":"6eb1576af52177d25e994c90280f879ffd9b4aaab909cee87614aa8d151f67d5"} Nov 28 13:21:28 crc kubenswrapper[4631]: I1128 13:21:28.373253 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-2bc4b" event={"ID":"9eb418ce-65b6-4da3-b282-831dd62916cf","Type":"ContainerStarted","Data":"48793a8f9c2488bf82d5c7596352e34ede9baff8e45779b3bb05fecc4a16daef"} Nov 28 13:21:28 crc kubenswrapper[4631]: I1128 13:21:28.375163 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-jgx6f" event={"ID":"5e56718f-0cda-4524-aa3a-068bc2e7d569","Type":"ContainerStarted","Data":"67d24c652f75d000cce88153ab68204b0e6f57e940e1ca87344dcbf5bf6c46ae"} Nov 28 13:21:28 crc kubenswrapper[4631]: I1128 13:21:28.377287 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8tbnv" event={"ID":"aa813d64-b22e-40a0-b72a-b4e220e85f5a","Type":"ContainerStarted","Data":"0b4e423322047b7f86b701c48d7fbf74a354c96428baa5d25ab59695485ff597"} Nov 28 13:21:28 crc kubenswrapper[4631]: I1128 13:21:28.379275 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-vhmzs" event={"ID":"adf078af-a8b4-4988-89e9-7daa0a80a4bb","Type":"ContainerStarted","Data":"cd4229c126e11f88c81a428b36c92a1f563c347d16bcdbe93ec516d8459ffa46"} Nov 28 13:21:28 crc kubenswrapper[4631]: I1128 13:21:28.449167 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5v5xp\" (UID: \"14717577-3fab-482b-bdbe-0e604d23d6d8\") " pod="openshift-image-registry/image-registry-697d97f7c8-5v5xp" Nov 28 13:21:28 crc kubenswrapper[4631]: E1128 13:21:28.449749 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 13:21:28.949727517 +0000 UTC m=+45.757030861 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5v5xp" (UID: "14717577-3fab-482b-bdbe-0e604d23d6d8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 13:21:28 crc kubenswrapper[4631]: I1128 13:21:28.554806 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 13:21:28 crc kubenswrapper[4631]: E1128 13:21:28.555132 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 13:21:29.055109077 +0000 UTC m=+45.862412421 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 13:21:28 crc kubenswrapper[4631]: I1128 13:21:28.555463 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5v5xp\" (UID: \"14717577-3fab-482b-bdbe-0e604d23d6d8\") " pod="openshift-image-registry/image-registry-697d97f7c8-5v5xp" Nov 28 13:21:28 crc kubenswrapper[4631]: E1128 13:21:28.555845 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 13:21:29.055836124 +0000 UTC m=+45.863139468 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5v5xp" (UID: "14717577-3fab-482b-bdbe-0e604d23d6d8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 13:21:28 crc kubenswrapper[4631]: I1128 13:21:28.664238 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 13:21:28 crc kubenswrapper[4631]: E1128 13:21:28.664622 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 13:21:29.164599842 +0000 UTC m=+45.971903186 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 13:21:28 crc kubenswrapper[4631]: I1128 13:21:28.664878 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5v5xp\" (UID: \"14717577-3fab-482b-bdbe-0e604d23d6d8\") " pod="openshift-image-registry/image-registry-697d97f7c8-5v5xp" Nov 28 13:21:28 crc kubenswrapper[4631]: E1128 13:21:28.665343 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 13:21:29.165304479 +0000 UTC m=+45.972607823 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5v5xp" (UID: "14717577-3fab-482b-bdbe-0e604d23d6d8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 13:21:28 crc kubenswrapper[4631]: I1128 13:21:28.766569 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 13:21:28 crc kubenswrapper[4631]: E1128 13:21:28.767293 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 13:21:29.267006514 +0000 UTC m=+46.074309868 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 13:21:28 crc kubenswrapper[4631]: I1128 13:21:28.767685 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5v5xp\" (UID: \"14717577-3fab-482b-bdbe-0e604d23d6d8\") " pod="openshift-image-registry/image-registry-697d97f7c8-5v5xp" Nov 28 13:21:28 crc kubenswrapper[4631]: E1128 13:21:28.768147 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 13:21:29.26812382 +0000 UTC m=+46.075427154 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5v5xp" (UID: "14717577-3fab-482b-bdbe-0e604d23d6d8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 13:21:28 crc kubenswrapper[4631]: I1128 13:21:28.835174 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-rwpv9" Nov 28 13:21:28 crc kubenswrapper[4631]: I1128 13:21:28.869231 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 13:21:28 crc kubenswrapper[4631]: E1128 13:21:28.869447 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 13:21:29.369410777 +0000 UTC m=+46.176714111 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 13:21:28 crc kubenswrapper[4631]: I1128 13:21:28.869554 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5v5xp\" (UID: \"14717577-3fab-482b-bdbe-0e604d23d6d8\") " pod="openshift-image-registry/image-registry-697d97f7c8-5v5xp" Nov 28 13:21:28 crc kubenswrapper[4631]: E1128 13:21:28.869834 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 13:21:29.369827976 +0000 UTC m=+46.177131320 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5v5xp" (UID: "14717577-3fab-482b-bdbe-0e604d23d6d8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 13:21:28 crc kubenswrapper[4631]: I1128 13:21:28.972653 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 13:21:28 crc kubenswrapper[4631]: E1128 13:21:28.972999 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 13:21:29.472967906 +0000 UTC m=+46.280271250 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 13:21:28 crc kubenswrapper[4631]: I1128 13:21:28.973509 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5v5xp\" (UID: \"14717577-3fab-482b-bdbe-0e604d23d6d8\") " pod="openshift-image-registry/image-registry-697d97f7c8-5v5xp" Nov 28 13:21:28 crc kubenswrapper[4631]: E1128 13:21:28.974894 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 13:21:29.474876719 +0000 UTC m=+46.282180063 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5v5xp" (UID: "14717577-3fab-482b-bdbe-0e604d23d6d8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 13:21:29 crc kubenswrapper[4631]: I1128 13:21:29.079357 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 13:21:29 crc kubenswrapper[4631]: E1128 13:21:29.079961 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 13:21:29.579930782 +0000 UTC m=+46.387234126 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 13:21:29 crc kubenswrapper[4631]: I1128 13:21:29.087767 4631 patch_prober.go:28] interesting pod/router-default-5444994796-jcnpq container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 28 13:21:29 crc kubenswrapper[4631]: [-]has-synced failed: reason withheld Nov 28 13:21:29 crc kubenswrapper[4631]: [+]process-running ok Nov 28 13:21:29 crc kubenswrapper[4631]: healthz check failed Nov 28 13:21:29 crc kubenswrapper[4631]: I1128 13:21:29.087823 4631 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-jcnpq" podUID="22010e1d-3ec1-4564-a362-bc963f691f1f" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 28 13:21:29 crc kubenswrapper[4631]: I1128 13:21:29.181716 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5v5xp\" (UID: \"14717577-3fab-482b-bdbe-0e604d23d6d8\") " pod="openshift-image-registry/image-registry-697d97f7c8-5v5xp" Nov 28 13:21:29 crc kubenswrapper[4631]: E1128 13:21:29.182326 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 13:21:29.682273643 +0000 UTC m=+46.489576987 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5v5xp" (UID: "14717577-3fab-482b-bdbe-0e604d23d6d8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 13:21:29 crc kubenswrapper[4631]: I1128 13:21:29.282451 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 13:21:29 crc kubenswrapper[4631]: E1128 13:21:29.283174 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 13:21:29.78315631 +0000 UTC m=+46.590459644 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 13:21:29 crc kubenswrapper[4631]: I1128 13:21:29.385189 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5v5xp\" (UID: \"14717577-3fab-482b-bdbe-0e604d23d6d8\") " pod="openshift-image-registry/image-registry-697d97f7c8-5v5xp" Nov 28 13:21:29 crc kubenswrapper[4631]: E1128 13:21:29.391086 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 13:21:29.891067679 +0000 UTC m=+46.698371023 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5v5xp" (UID: "14717577-3fab-482b-bdbe-0e604d23d6d8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 13:21:29 crc kubenswrapper[4631]: I1128 13:21:29.450800 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-2v5d6" event={"ID":"2a2f7727-f05e-4deb-8eba-cb3ec3561621","Type":"ContainerStarted","Data":"bb5dbc944d5d38bd3f5892e149a20b00deee91385665f1e95a656f8a4c8a3f85"} Nov 28 13:21:29 crc kubenswrapper[4631]: I1128 13:21:29.470125 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-gk97s" event={"ID":"6fccb501-125b-4bc4-8cad-3349f332c2de","Type":"ContainerStarted","Data":"a6ecdaa357c96afa3760fade45a7ec891a483ed9f7b8591691908ad3484c29a7"} Nov 28 13:21:29 crc kubenswrapper[4631]: I1128 13:21:29.471169 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-2v5d6" podStartSLOduration=25.471158636 podStartE2EDuration="25.471158636s" podCreationTimestamp="2025-11-28 13:21:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 13:21:29.467191574 +0000 UTC m=+46.274494918" watchObservedRunningTime="2025-11-28 13:21:29.471158636 +0000 UTC m=+46.278461980" Nov 28 13:21:29 crc kubenswrapper[4631]: I1128 13:21:29.475953 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-r6sdp" event={"ID":"9bc5a0ed-1419-4ba0-8478-9bcad6a19954","Type":"ContainerStarted","Data":"7974d22e49a8d8998dd5786268ff8f081c17b7edebe8135a9a2b1ed947705636"} Nov 28 13:21:29 crc kubenswrapper[4631]: I1128 13:21:29.477126 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-xltdb" event={"ID":"4415ae1d-dbf4-450b-a329-6dc64b0a019f","Type":"ContainerStarted","Data":"4ed587be4e7dfe9e5d42f33dfb492283f45774121315ce73b7e70656ca5f551d"} Nov 28 13:21:29 crc kubenswrapper[4631]: E1128 13:21:29.488453 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 13:21:29.988424765 +0000 UTC m=+46.795728109 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 13:21:29 crc kubenswrapper[4631]: I1128 13:21:29.488493 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 13:21:29 crc kubenswrapper[4631]: I1128 13:21:29.488773 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5v5xp\" (UID: \"14717577-3fab-482b-bdbe-0e604d23d6d8\") " pod="openshift-image-registry/image-registry-697d97f7c8-5v5xp" Nov 28 13:21:29 crc kubenswrapper[4631]: E1128 13:21:29.489079 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 13:21:29.989071719 +0000 UTC m=+46.796375063 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5v5xp" (UID: "14717577-3fab-482b-bdbe-0e604d23d6d8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 13:21:29 crc kubenswrapper[4631]: I1128 13:21:29.521757 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-gk97s" podStartSLOduration=25.521736882 podStartE2EDuration="25.521736882s" podCreationTimestamp="2025-11-28 13:21:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 13:21:29.515889957 +0000 UTC m=+46.323193291" watchObservedRunningTime="2025-11-28 13:21:29.521736882 +0000 UTC m=+46.329040226" Nov 28 13:21:29 crc kubenswrapper[4631]: I1128 13:21:29.542660 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bzj6n" event={"ID":"1039cddf-be22-4ca1-a622-d21967f9d859","Type":"ContainerStarted","Data":"05781d0c3dc00563fb9ba21b4c2409736235a7c57b1d221c2aa5d781dbe220e6"} Nov 28 13:21:29 crc kubenswrapper[4631]: I1128 13:21:29.542733 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bzj6n" Nov 28 13:21:29 crc kubenswrapper[4631]: I1128 13:21:29.562947 4631 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-bzj6n container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.29:8443/healthz\": dial tcp 10.217.0.29:8443: connect: connection refused" start-of-body= Nov 28 13:21:29 crc kubenswrapper[4631]: I1128 13:21:29.562993 4631 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bzj6n" podUID="1039cddf-be22-4ca1-a622-d21967f9d859" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.29:8443/healthz\": dial tcp 10.217.0.29:8443: connect: connection refused" Nov 28 13:21:29 crc kubenswrapper[4631]: I1128 13:21:29.570114 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-cg4zv" event={"ID":"089e1482-6475-40b3-a005-c151f4bb0a87","Type":"ContainerStarted","Data":"928063ae1cbe1ec1b59fa0bed13c6fa42a91fccf037fb5bf1d8feb250fa8dd4e"} Nov 28 13:21:29 crc kubenswrapper[4631]: I1128 13:21:29.592722 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 13:21:29 crc kubenswrapper[4631]: E1128 13:21:29.594982 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 13:21:30.094948761 +0000 UTC m=+46.902252105 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 13:21:29 crc kubenswrapper[4631]: I1128 13:21:29.600592 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-xltdb" podStartSLOduration=25.600565611 podStartE2EDuration="25.600565611s" podCreationTimestamp="2025-11-28 13:21:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 13:21:29.553923084 +0000 UTC m=+46.361226428" watchObservedRunningTime="2025-11-28 13:21:29.600565611 +0000 UTC m=+46.407868965" Nov 28 13:21:29 crc kubenswrapper[4631]: I1128 13:21:29.629709 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-s7zwj" event={"ID":"2ef91b69-6215-495f-ac26-23a99f34dcc1","Type":"ContainerStarted","Data":"dbcb5157bc79968edb44af92794635f7a6891b569aa13389a1a728613c169e46"} Nov 28 13:21:29 crc kubenswrapper[4631]: I1128 13:21:29.653153 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-5h9cj" event={"ID":"2eddcce3-58a6-45f9-b924-30d95fa897ae","Type":"ContainerStarted","Data":"912de0700157bf30468e286ac77689aa4b5d81188a74d7c180b688d7e9e405d4"} Nov 28 13:21:29 crc kubenswrapper[4631]: I1128 13:21:29.666146 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bzj6n" podStartSLOduration=25.666127393 podStartE2EDuration="25.666127393s" podCreationTimestamp="2025-11-28 13:21:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 13:21:29.606521748 +0000 UTC m=+46.413825092" watchObservedRunningTime="2025-11-28 13:21:29.666127393 +0000 UTC m=+46.473430737" Nov 28 13:21:29 crc kubenswrapper[4631]: I1128 13:21:29.667817 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-s7zwj" podStartSLOduration=25.667809952 podStartE2EDuration="25.667809952s" podCreationTimestamp="2025-11-28 13:21:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 13:21:29.665075149 +0000 UTC m=+46.472378493" watchObservedRunningTime="2025-11-28 13:21:29.667809952 +0000 UTC m=+46.475113296" Nov 28 13:21:29 crc kubenswrapper[4631]: I1128 13:21:29.671538 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-fx9mb" event={"ID":"51595599-be30-46b6-bab6-ee647410efbc","Type":"ContainerStarted","Data":"787f64b2a3fd1cd6950ea70be1b77d06635c31ba03b6f27b2819d9312c80c324"} Nov 28 13:21:29 crc kubenswrapper[4631]: I1128 13:21:29.693727 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5v5xp\" (UID: \"14717577-3fab-482b-bdbe-0e604d23d6d8\") " pod="openshift-image-registry/image-registry-697d97f7c8-5v5xp" Nov 28 13:21:29 crc kubenswrapper[4631]: E1128 13:21:29.695166 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 13:21:30.195154412 +0000 UTC m=+47.002457756 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5v5xp" (UID: "14717577-3fab-482b-bdbe-0e604d23d6d8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 13:21:29 crc kubenswrapper[4631]: I1128 13:21:29.730629 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-dd58r" event={"ID":"8716de2d-f33e-4434-9076-6345669aaff8","Type":"ContainerStarted","Data":"e3b69f5679a93fb305c51f6a6f51a107aacffd90bb97696f2355adcbec1bc6ed"} Nov 28 13:21:29 crc kubenswrapper[4631]: I1128 13:21:29.764123 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-hzvzz" event={"ID":"0e2f0bc1-2261-484e-8f4b-34bddacda1bc","Type":"ContainerStarted","Data":"077996910a5a07e790ec0ae73a0b3771ca0641ae0b680f2d5f76b24dd85c494f"} Nov 28 13:21:29 crc kubenswrapper[4631]: I1128 13:21:29.764553 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-hzvzz" Nov 28 13:21:29 crc kubenswrapper[4631]: I1128 13:21:29.778422 4631 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-hzvzz container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.37:8443/healthz\": dial tcp 10.217.0.37:8443: connect: connection refused" start-of-body= Nov 28 13:21:29 crc kubenswrapper[4631]: I1128 13:21:29.778471 4631 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-hzvzz" podUID="0e2f0bc1-2261-484e-8f4b-34bddacda1bc" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.37:8443/healthz\": dial tcp 10.217.0.37:8443: connect: connection refused" Nov 28 13:21:29 crc kubenswrapper[4631]: I1128 13:21:29.812294 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-658pf" event={"ID":"1a4f8da7-c099-4a8e-bffc-ca17ca74eb42","Type":"ContainerStarted","Data":"e001975f37c7f1522fb7d3161f3f17eaf772a411aa36bf5d6a27b446769a76f9"} Nov 28 13:21:29 crc kubenswrapper[4631]: I1128 13:21:29.814841 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 13:21:29 crc kubenswrapper[4631]: E1128 13:21:29.825807 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 13:21:30.325763535 +0000 UTC m=+47.133066879 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 13:21:29 crc kubenswrapper[4631]: I1128 13:21:29.841552 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-fx9mb" podStartSLOduration=25.841521429 podStartE2EDuration="25.841521429s" podCreationTimestamp="2025-11-28 13:21:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 13:21:29.743243682 +0000 UTC m=+46.550547026" watchObservedRunningTime="2025-11-28 13:21:29.841521429 +0000 UTC m=+46.648824773" Nov 28 13:21:29 crc kubenswrapper[4631]: I1128 13:21:29.874442 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-82qbh" event={"ID":"6fe798f2-84e0-4019-a75c-0ee2f2e672f5","Type":"ContainerStarted","Data":"cbf4da7071732b2ac23afff8a0b9b4157e949b0e8dcc3b3c90a58cc9ece1ea52"} Nov 28 13:21:29 crc kubenswrapper[4631]: I1128 13:21:29.889986 4631 patch_prober.go:28] interesting pod/downloads-7954f5f757-c2rks container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.7:8080/\": dial tcp 10.217.0.7:8080: connect: connection refused" start-of-body= Nov 28 13:21:29 crc kubenswrapper[4631]: I1128 13:21:29.890035 4631 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-c2rks" podUID="feeda081-8bb0-4a89-ae6d-2e87f8f73344" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.7:8080/\": dial tcp 10.217.0.7:8080: connect: connection refused" Nov 28 13:21:29 crc kubenswrapper[4631]: I1128 13:21:29.890504 4631 patch_prober.go:28] interesting pod/downloads-7954f5f757-c2rks container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.7:8080/\": dial tcp 10.217.0.7:8080: connect: connection refused" start-of-body= Nov 28 13:21:29 crc kubenswrapper[4631]: I1128 13:21:29.890536 4631 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-c2rks" podUID="feeda081-8bb0-4a89-ae6d-2e87f8f73344" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.7:8080/\": dial tcp 10.217.0.7:8080: connect: connection refused" Nov 28 13:21:29 crc kubenswrapper[4631]: I1128 13:21:29.895557 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-hzvzz" podStartSLOduration=25.895542124 podStartE2EDuration="25.895542124s" podCreationTimestamp="2025-11-28 13:21:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 13:21:29.85202716 +0000 UTC m=+46.659330504" watchObservedRunningTime="2025-11-28 13:21:29.895542124 +0000 UTC m=+46.702845468" Nov 28 13:21:29 crc kubenswrapper[4631]: I1128 13:21:29.899412 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-jqnpq" event={"ID":"6b363da7-b1c6-495d-a05b-7ff4d7951b3e","Type":"ContainerStarted","Data":"e60041e3d6e7c2a7aecc5749ff079a1521f8c44b24c71fc16f7e87308081a5c7"} Nov 28 13:21:29 crc kubenswrapper[4631]: I1128 13:21:29.901236 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-mdws7" event={"ID":"3134d1ce-629c-4e8b-86c7-b69a24578952","Type":"ContainerStarted","Data":"66206f441e284f7ba75b17e99b4d71e944ebac02bb59d483cb3404b0f0bd1bd5"} Nov 28 13:21:29 crc kubenswrapper[4631]: I1128 13:21:29.901436 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-mdws7" Nov 28 13:21:29 crc kubenswrapper[4631]: I1128 13:21:29.904113 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29405595-2v6fk" event={"ID":"76a05ca6-a98c-4313-855c-fd6ffaceee4f","Type":"ContainerStarted","Data":"0f3af793008c1a47e8009812f16cb978e3930b0e22afc292634c25e32a447f21"} Nov 28 13:21:29 crc kubenswrapper[4631]: I1128 13:21:29.905886 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-2bc4b" Nov 28 13:21:29 crc kubenswrapper[4631]: I1128 13:21:29.906354 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-multus/cni-sysctl-allowlist-ds-44pwx" Nov 28 13:21:29 crc kubenswrapper[4631]: I1128 13:21:29.910555 4631 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-2bc4b container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.42:8080/healthz\": dial tcp 10.217.0.42:8080: connect: connection refused" start-of-body= Nov 28 13:21:29 crc kubenswrapper[4631]: I1128 13:21:29.910604 4631 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-2bc4b" podUID="9eb418ce-65b6-4da3-b282-831dd62916cf" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.42:8080/healthz\": dial tcp 10.217.0.42:8080: connect: connection refused" Nov 28 13:21:29 crc kubenswrapper[4631]: I1128 13:21:29.921228 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5v5xp\" (UID: \"14717577-3fab-482b-bdbe-0e604d23d6d8\") " pod="openshift-image-registry/image-registry-697d97f7c8-5v5xp" Nov 28 13:21:29 crc kubenswrapper[4631]: E1128 13:21:29.927819 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 13:21:30.427803808 +0000 UTC m=+47.235107152 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5v5xp" (UID: "14717577-3fab-482b-bdbe-0e604d23d6d8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 13:21:29 crc kubenswrapper[4631]: I1128 13:21:29.944119 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-658pf" podStartSLOduration=25.944100124 podStartE2EDuration="25.944100124s" podCreationTimestamp="2025-11-28 13:21:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 13:21:29.896840464 +0000 UTC m=+46.704143808" watchObservedRunningTime="2025-11-28 13:21:29.944100124 +0000 UTC m=+46.751403468" Nov 28 13:21:29 crc kubenswrapper[4631]: I1128 13:21:29.959533 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-multus/cni-sysctl-allowlist-ds-44pwx" Nov 28 13:21:29 crc kubenswrapper[4631]: I1128 13:21:29.976936 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-jqnpq" podStartSLOduration=25.976919112 podStartE2EDuration="25.976919112s" podCreationTimestamp="2025-11-28 13:21:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 13:21:29.944643017 +0000 UTC m=+46.751946361" watchObservedRunningTime="2025-11-28 13:21:29.976919112 +0000 UTC m=+46.784222456" Nov 28 13:21:29 crc kubenswrapper[4631]: I1128 13:21:29.978149 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-mdws7" podStartSLOduration=25.978144099 podStartE2EDuration="25.978144099s" podCreationTimestamp="2025-11-28 13:21:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 13:21:29.976715466 +0000 UTC m=+46.784018810" watchObservedRunningTime="2025-11-28 13:21:29.978144099 +0000 UTC m=+46.785447453" Nov 28 13:21:30 crc kubenswrapper[4631]: I1128 13:21:30.016512 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29405595-2v6fk" podStartSLOduration=27.016493664 podStartE2EDuration="27.016493664s" podCreationTimestamp="2025-11-28 13:21:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 13:21:30.014662052 +0000 UTC m=+46.821965396" watchObservedRunningTime="2025-11-28 13:21:30.016493664 +0000 UTC m=+46.823797008" Nov 28 13:21:30 crc kubenswrapper[4631]: I1128 13:21:30.026590 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 13:21:30 crc kubenswrapper[4631]: E1128 13:21:30.026973 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 13:21:30.526959035 +0000 UTC m=+47.334262379 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 13:21:30 crc kubenswrapper[4631]: I1128 13:21:30.033095 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-rsxk2" podStartSLOduration=12.033082367 podStartE2EDuration="12.033082367s" podCreationTimestamp="2025-11-28 13:21:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 13:21:30.031293425 +0000 UTC m=+46.838596769" watchObservedRunningTime="2025-11-28 13:21:30.033082367 +0000 UTC m=+46.840385711" Nov 28 13:21:30 crc kubenswrapper[4631]: I1128 13:21:30.087146 4631 patch_prober.go:28] interesting pod/router-default-5444994796-jcnpq container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 28 13:21:30 crc kubenswrapper[4631]: [-]has-synced failed: reason withheld Nov 28 13:21:30 crc kubenswrapper[4631]: [+]process-running ok Nov 28 13:21:30 crc kubenswrapper[4631]: healthz check failed Nov 28 13:21:30 crc kubenswrapper[4631]: I1128 13:21:30.087234 4631 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-jcnpq" podUID="22010e1d-3ec1-4564-a362-bc963f691f1f" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 28 13:21:30 crc kubenswrapper[4631]: I1128 13:21:30.094677 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-xqbvh" podStartSLOduration=28.094646677 podStartE2EDuration="28.094646677s" podCreationTimestamp="2025-11-28 13:21:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 13:21:30.092780994 +0000 UTC m=+46.900084328" watchObservedRunningTime="2025-11-28 13:21:30.094646677 +0000 UTC m=+46.901950021" Nov 28 13:21:30 crc kubenswrapper[4631]: I1128 13:21:30.095244 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-2bc4b" podStartSLOduration=26.09522472 podStartE2EDuration="26.09522472s" podCreationTimestamp="2025-11-28 13:21:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 13:21:30.059533797 +0000 UTC m=+46.866837141" watchObservedRunningTime="2025-11-28 13:21:30.09522472 +0000 UTC m=+46.902528054" Nov 28 13:21:30 crc kubenswrapper[4631]: I1128 13:21:30.121876 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-977k9" podStartSLOduration=26.121846244 podStartE2EDuration="26.121846244s" podCreationTimestamp="2025-11-28 13:21:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 13:21:30.118145639 +0000 UTC m=+46.925448983" watchObservedRunningTime="2025-11-28 13:21:30.121846244 +0000 UTC m=+46.929149588" Nov 28 13:21:30 crc kubenswrapper[4631]: I1128 13:21:30.127745 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5v5xp\" (UID: \"14717577-3fab-482b-bdbe-0e604d23d6d8\") " pod="openshift-image-registry/image-registry-697d97f7c8-5v5xp" Nov 28 13:21:30 crc kubenswrapper[4631]: E1128 13:21:30.128036 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 13:21:30.628025336 +0000 UTC m=+47.435328680 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5v5xp" (UID: "14717577-3fab-482b-bdbe-0e604d23d6d8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 13:21:30 crc kubenswrapper[4631]: I1128 13:21:30.147711 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/cni-sysctl-allowlist-ds-44pwx" podStartSLOduration=12.147696191 podStartE2EDuration="12.147696191s" podCreationTimestamp="2025-11-28 13:21:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 13:21:30.146641596 +0000 UTC m=+46.953944940" watchObservedRunningTime="2025-11-28 13:21:30.147696191 +0000 UTC m=+46.954999535" Nov 28 13:21:30 crc kubenswrapper[4631]: I1128 13:21:30.171814 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8tbnv" podStartSLOduration=26.171797667 podStartE2EDuration="26.171797667s" podCreationTimestamp="2025-11-28 13:21:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 13:21:30.169704288 +0000 UTC m=+46.977007632" watchObservedRunningTime="2025-11-28 13:21:30.171797667 +0000 UTC m=+46.979101011" Nov 28 13:21:30 crc kubenswrapper[4631]: I1128 13:21:30.228350 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 13:21:30 crc kubenswrapper[4631]: E1128 13:21:30.228674 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 13:21:30.728658078 +0000 UTC m=+47.535961422 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 13:21:30 crc kubenswrapper[4631]: I1128 13:21:30.267729 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-qdxf6" podStartSLOduration=26.267711298 podStartE2EDuration="26.267711298s" podCreationTimestamp="2025-11-28 13:21:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 13:21:30.213215581 +0000 UTC m=+47.020518925" watchObservedRunningTime="2025-11-28 13:21:30.267711298 +0000 UTC m=+47.075014642" Nov 28 13:21:30 crc kubenswrapper[4631]: I1128 13:21:30.330238 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5v5xp\" (UID: \"14717577-3fab-482b-bdbe-0e604d23d6d8\") " pod="openshift-image-registry/image-registry-697d97f7c8-5v5xp" Nov 28 13:21:30 crc kubenswrapper[4631]: E1128 13:21:30.330589 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 13:21:30.830576329 +0000 UTC m=+47.637879673 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5v5xp" (UID: "14717577-3fab-482b-bdbe-0e604d23d6d8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 13:21:30 crc kubenswrapper[4631]: I1128 13:21:30.431590 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 13:21:30 crc kubenswrapper[4631]: E1128 13:21:30.431904 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 13:21:30.931889725 +0000 UTC m=+47.739193069 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 13:21:30 crc kubenswrapper[4631]: I1128 13:21:30.533276 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5v5xp\" (UID: \"14717577-3fab-482b-bdbe-0e604d23d6d8\") " pod="openshift-image-registry/image-registry-697d97f7c8-5v5xp" Nov 28 13:21:30 crc kubenswrapper[4631]: E1128 13:21:30.533578 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 13:21:31.033565521 +0000 UTC m=+47.840868865 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5v5xp" (UID: "14717577-3fab-482b-bdbe-0e604d23d6d8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 13:21:30 crc kubenswrapper[4631]: I1128 13:21:30.634115 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 13:21:30 crc kubenswrapper[4631]: E1128 13:21:30.634347 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 13:21:31.134321565 +0000 UTC m=+47.941624899 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 13:21:30 crc kubenswrapper[4631]: I1128 13:21:30.634736 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5v5xp\" (UID: \"14717577-3fab-482b-bdbe-0e604d23d6d8\") " pod="openshift-image-registry/image-registry-697d97f7c8-5v5xp" Nov 28 13:21:30 crc kubenswrapper[4631]: E1128 13:21:30.635036 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 13:21:31.135023131 +0000 UTC m=+47.942326475 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5v5xp" (UID: "14717577-3fab-482b-bdbe-0e604d23d6d8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 13:21:30 crc kubenswrapper[4631]: I1128 13:21:30.736060 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 13:21:30 crc kubenswrapper[4631]: E1128 13:21:30.736203 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 13:21:31.236180154 +0000 UTC m=+48.043483498 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 13:21:30 crc kubenswrapper[4631]: I1128 13:21:30.736334 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5v5xp\" (UID: \"14717577-3fab-482b-bdbe-0e604d23d6d8\") " pod="openshift-image-registry/image-registry-697d97f7c8-5v5xp" Nov 28 13:21:30 crc kubenswrapper[4631]: E1128 13:21:30.736639 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 13:21:31.236626194 +0000 UTC m=+48.043929538 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5v5xp" (UID: "14717577-3fab-482b-bdbe-0e604d23d6d8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 13:21:30 crc kubenswrapper[4631]: I1128 13:21:30.740229 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-multus/cni-sysctl-allowlist-ds-44pwx"] Nov 28 13:21:30 crc kubenswrapper[4631]: I1128 13:21:30.836919 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 13:21:30 crc kubenswrapper[4631]: E1128 13:21:30.837057 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 13:21:31.33702795 +0000 UTC m=+48.144331294 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 13:21:30 crc kubenswrapper[4631]: I1128 13:21:30.837222 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5v5xp\" (UID: \"14717577-3fab-482b-bdbe-0e604d23d6d8\") " pod="openshift-image-registry/image-registry-697d97f7c8-5v5xp" Nov 28 13:21:30 crc kubenswrapper[4631]: E1128 13:21:30.837628 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 13:21:31.337615353 +0000 UTC m=+48.144918697 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5v5xp" (UID: "14717577-3fab-482b-bdbe-0e604d23d6d8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 13:21:30 crc kubenswrapper[4631]: I1128 13:21:30.909234 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-dd58r" event={"ID":"8716de2d-f33e-4434-9076-6345669aaff8","Type":"ContainerStarted","Data":"227a1e5b1627ac20c9847d2d2d4f634100a827d148f43a4255a935c733bcca43"} Nov 28 13:21:30 crc kubenswrapper[4631]: I1128 13:21:30.911081 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-cg4zv" event={"ID":"089e1482-6475-40b3-a005-c151f4bb0a87","Type":"ContainerStarted","Data":"3544bc12973ba11f89fd3da8e797dcb17b8b036eabc870a1af696fec30689721"} Nov 28 13:21:30 crc kubenswrapper[4631]: I1128 13:21:30.911449 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-cg4zv" Nov 28 13:21:30 crc kubenswrapper[4631]: I1128 13:21:30.912592 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-g8n9t" event={"ID":"5db3c754-f675-4ec8-b3aa-65d0528bb9ca","Type":"ContainerStarted","Data":"2b318951d129a4038c73351dc2fb1230d0cf26a515539d76625e3d89992e4107"} Nov 28 13:21:30 crc kubenswrapper[4631]: I1128 13:21:30.914684 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-jgx6f" event={"ID":"5e56718f-0cda-4524-aa3a-068bc2e7d569","Type":"ContainerStarted","Data":"a45d2a06695a5ab2e642eb67d4469eb26de75338ddeaf8a15aef9e1f9ed6c45f"} Nov 28 13:21:30 crc kubenswrapper[4631]: I1128 13:21:30.916097 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-vhmzs" event={"ID":"adf078af-a8b4-4988-89e9-7daa0a80a4bb","Type":"ContainerStarted","Data":"8966d2e081639b809ff42b796e5754f4fc7d154ba3ab0536f48842b88a05a94e"} Nov 28 13:21:30 crc kubenswrapper[4631]: I1128 13:21:30.917569 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-dxzdh" event={"ID":"73b80803-714a-446e-9df2-6e3ccf5430ee","Type":"ContainerStarted","Data":"1d752da2b1f482d0d5b4f942288ba28079ccf583679becfad5354c9bc61de01e"} Nov 28 13:21:30 crc kubenswrapper[4631]: I1128 13:21:30.919156 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-5h9cj" event={"ID":"2eddcce3-58a6-45f9-b924-30d95fa897ae","Type":"ContainerStarted","Data":"759c2ee68d11d150a2f61d4da9f8ed926f65823ac9e2ea13a97cd318169839b5"} Nov 28 13:21:30 crc kubenswrapper[4631]: I1128 13:21:30.934311 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-52qtd" event={"ID":"f8a4f24e-7ccd-4527-9020-2f101c1d4cb5","Type":"ContainerStarted","Data":"b649fe5be70b6eece647c43d685093b17f05eccc45cf23085739ec57293f39ef"} Nov 28 13:21:30 crc kubenswrapper[4631]: I1128 13:21:30.938081 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-82qbh" event={"ID":"6fe798f2-84e0-4019-a75c-0ee2f2e672f5","Type":"ContainerStarted","Data":"5f7b42836b2bfdcdfeff8dadf0ca6268cb4be1e50c59f91621f0cb3fb1c0106a"} Nov 28 13:21:30 crc kubenswrapper[4631]: I1128 13:21:30.939166 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 13:21:30 crc kubenswrapper[4631]: E1128 13:21:30.940433 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 13:21:31.440412135 +0000 UTC m=+48.247715489 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 13:21:30 crc kubenswrapper[4631]: I1128 13:21:30.941113 4631 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-2bc4b container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.42:8080/healthz\": dial tcp 10.217.0.42:8080: connect: connection refused" start-of-body= Nov 28 13:21:30 crc kubenswrapper[4631]: I1128 13:21:30.941149 4631 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-2bc4b" podUID="9eb418ce-65b6-4da3-b282-831dd62916cf" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.42:8080/healthz\": dial tcp 10.217.0.42:8080: connect: connection refused" Nov 28 13:21:30 crc kubenswrapper[4631]: I1128 13:21:30.947620 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-hzvzz" Nov 28 13:21:30 crc kubenswrapper[4631]: I1128 13:21:30.982710 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-dd58r" podStartSLOduration=27.98269413 podStartE2EDuration="27.98269413s" podCreationTimestamp="2025-11-28 13:21:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 13:21:30.962057314 +0000 UTC m=+47.769360658" watchObservedRunningTime="2025-11-28 13:21:30.98269413 +0000 UTC m=+47.789997474" Nov 28 13:21:30 crc kubenswrapper[4631]: I1128 13:21:30.982888 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-dxzdh" podStartSLOduration=26.982884504 podStartE2EDuration="26.982884504s" podCreationTimestamp="2025-11-28 13:21:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 13:21:30.981208086 +0000 UTC m=+47.788511430" watchObservedRunningTime="2025-11-28 13:21:30.982884504 +0000 UTC m=+47.790187838" Nov 28 13:21:30 crc kubenswrapper[4631]: I1128 13:21:30.995635 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bzj6n" Nov 28 13:21:31 crc kubenswrapper[4631]: I1128 13:21:31.043479 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5v5xp\" (UID: \"14717577-3fab-482b-bdbe-0e604d23d6d8\") " pod="openshift-image-registry/image-registry-697d97f7c8-5v5xp" Nov 28 13:21:31 crc kubenswrapper[4631]: E1128 13:21:31.043810 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 13:21:31.54379374 +0000 UTC m=+48.351097084 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5v5xp" (UID: "14717577-3fab-482b-bdbe-0e604d23d6d8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 13:21:31 crc kubenswrapper[4631]: I1128 13:21:31.073430 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-vhmzs" podStartSLOduration=28.073415563 podStartE2EDuration="28.073415563s" podCreationTimestamp="2025-11-28 13:21:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 13:21:31.07069918 +0000 UTC m=+47.878002524" watchObservedRunningTime="2025-11-28 13:21:31.073415563 +0000 UTC m=+47.880718907" Nov 28 13:21:31 crc kubenswrapper[4631]: I1128 13:21:31.086565 4631 patch_prober.go:28] interesting pod/router-default-5444994796-jcnpq container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 28 13:21:31 crc kubenswrapper[4631]: [-]has-synced failed: reason withheld Nov 28 13:21:31 crc kubenswrapper[4631]: [+]process-running ok Nov 28 13:21:31 crc kubenswrapper[4631]: healthz check failed Nov 28 13:21:31 crc kubenswrapper[4631]: I1128 13:21:31.086648 4631 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-jcnpq" podUID="22010e1d-3ec1-4564-a362-bc963f691f1f" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 28 13:21:31 crc kubenswrapper[4631]: I1128 13:21:31.133050 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-82qbh" podStartSLOduration=27.133033057 podStartE2EDuration="27.133033057s" podCreationTimestamp="2025-11-28 13:21:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 13:21:31.122905624 +0000 UTC m=+47.930208958" watchObservedRunningTime="2025-11-28 13:21:31.133033057 +0000 UTC m=+47.940336401" Nov 28 13:21:31 crc kubenswrapper[4631]: I1128 13:21:31.147147 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 13:21:31 crc kubenswrapper[4631]: E1128 13:21:31.147341 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 13:21:31.647311896 +0000 UTC m=+48.454615240 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 13:21:31 crc kubenswrapper[4631]: I1128 13:21:31.147556 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5v5xp\" (UID: \"14717577-3fab-482b-bdbe-0e604d23d6d8\") " pod="openshift-image-registry/image-registry-697d97f7c8-5v5xp" Nov 28 13:21:31 crc kubenswrapper[4631]: E1128 13:21:31.147862 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 13:21:31.647850949 +0000 UTC m=+48.455154363 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5v5xp" (UID: "14717577-3fab-482b-bdbe-0e604d23d6d8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 13:21:31 crc kubenswrapper[4631]: I1128 13:21:31.249078 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 13:21:31 crc kubenswrapper[4631]: E1128 13:21:31.249356 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 13:21:31.74932729 +0000 UTC m=+48.556630634 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 13:21:31 crc kubenswrapper[4631]: I1128 13:21:31.336146 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-g8n9t" podStartSLOduration=27.336131552 podStartE2EDuration="27.336131552s" podCreationTimestamp="2025-11-28 13:21:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 13:21:31.221419466 +0000 UTC m=+48.028722810" watchObservedRunningTime="2025-11-28 13:21:31.336131552 +0000 UTC m=+48.143434896" Nov 28 13:21:31 crc kubenswrapper[4631]: I1128 13:21:31.350934 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5v5xp\" (UID: \"14717577-3fab-482b-bdbe-0e604d23d6d8\") " pod="openshift-image-registry/image-registry-697d97f7c8-5v5xp" Nov 28 13:21:31 crc kubenswrapper[4631]: E1128 13:21:31.351279 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 13:21:31.851268001 +0000 UTC m=+48.658571345 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5v5xp" (UID: "14717577-3fab-482b-bdbe-0e604d23d6d8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 13:21:31 crc kubenswrapper[4631]: I1128 13:21:31.433853 4631 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-659jg" Nov 28 13:21:31 crc kubenswrapper[4631]: I1128 13:21:31.433898 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-659jg" Nov 28 13:21:31 crc kubenswrapper[4631]: I1128 13:21:31.449135 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-5h9cj" podStartSLOduration=27.449118518 podStartE2EDuration="27.449118518s" podCreationTimestamp="2025-11-28 13:21:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 13:21:31.408018551 +0000 UTC m=+48.215321885" watchObservedRunningTime="2025-11-28 13:21:31.449118518 +0000 UTC m=+48.256421862" Nov 28 13:21:31 crc kubenswrapper[4631]: I1128 13:21:31.452001 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 13:21:31 crc kubenswrapper[4631]: E1128 13:21:31.453398 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 13:21:31.953382846 +0000 UTC m=+48.760686190 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 13:21:31 crc kubenswrapper[4631]: I1128 13:21:31.464939 4631 patch_prober.go:28] interesting pod/console-f9d7485db-659jg container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.15:8443/health\": dial tcp 10.217.0.15:8443: connect: connection refused" start-of-body= Nov 28 13:21:31 crc kubenswrapper[4631]: I1128 13:21:31.464997 4631 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-659jg" podUID="8e852fdc-5b82-4b51-b0e5-804a1b3be0ec" containerName="console" probeResult="failure" output="Get \"https://10.217.0.15:8443/health\": dial tcp 10.217.0.15:8443: connect: connection refused" Nov 28 13:21:31 crc kubenswrapper[4631]: I1128 13:21:31.558798 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5v5xp\" (UID: \"14717577-3fab-482b-bdbe-0e604d23d6d8\") " pod="openshift-image-registry/image-registry-697d97f7c8-5v5xp" Nov 28 13:21:31 crc kubenswrapper[4631]: E1128 13:21:31.559119 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 13:21:32.059106705 +0000 UTC m=+48.866410049 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5v5xp" (UID: "14717577-3fab-482b-bdbe-0e604d23d6d8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 13:21:31 crc kubenswrapper[4631]: I1128 13:21:31.586629 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-52qtd" podStartSLOduration=27.586610649 podStartE2EDuration="27.586610649s" podCreationTimestamp="2025-11-28 13:21:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 13:21:31.464714668 +0000 UTC m=+48.272018012" watchObservedRunningTime="2025-11-28 13:21:31.586610649 +0000 UTC m=+48.393913993" Nov 28 13:21:31 crc kubenswrapper[4631]: I1128 13:21:31.660321 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 13:21:31 crc kubenswrapper[4631]: E1128 13:21:31.660479 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 13:21:32.160455503 +0000 UTC m=+48.967758847 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 13:21:31 crc kubenswrapper[4631]: I1128 13:21:31.660583 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5v5xp\" (UID: \"14717577-3fab-482b-bdbe-0e604d23d6d8\") " pod="openshift-image-registry/image-registry-697d97f7c8-5v5xp" Nov 28 13:21:31 crc kubenswrapper[4631]: E1128 13:21:31.660855 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 13:21:32.160841171 +0000 UTC m=+48.968144515 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5v5xp" (UID: "14717577-3fab-482b-bdbe-0e604d23d6d8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 13:21:31 crc kubenswrapper[4631]: I1128 13:21:31.680180 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-cg4zv" podStartSLOduration=13.680160977 podStartE2EDuration="13.680160977s" podCreationTimestamp="2025-11-28 13:21:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 13:21:31.678247433 +0000 UTC m=+48.485550777" watchObservedRunningTime="2025-11-28 13:21:31.680160977 +0000 UTC m=+48.487464331" Nov 28 13:21:31 crc kubenswrapper[4631]: I1128 13:21:31.680676 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-jgx6f" podStartSLOduration=28.680668478 podStartE2EDuration="28.680668478s" podCreationTimestamp="2025-11-28 13:21:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 13:21:31.587977241 +0000 UTC m=+48.395280595" watchObservedRunningTime="2025-11-28 13:21:31.680668478 +0000 UTC m=+48.487971822" Nov 28 13:21:31 crc kubenswrapper[4631]: I1128 13:21:31.712976 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-896hq" Nov 28 13:21:31 crc kubenswrapper[4631]: I1128 13:21:31.761369 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 13:21:31 crc kubenswrapper[4631]: E1128 13:21:31.762062 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 13:21:32.262036965 +0000 UTC m=+49.069340309 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 13:21:31 crc kubenswrapper[4631]: I1128 13:21:31.794478 4631 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-2bc4b container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.42:8080/healthz\": dial tcp 10.217.0.42:8080: connect: connection refused" start-of-body= Nov 28 13:21:31 crc kubenswrapper[4631]: I1128 13:21:31.794520 4631 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-2bc4b container/marketplace-operator namespace/openshift-marketplace: Liveness probe status=failure output="Get \"http://10.217.0.42:8080/healthz\": dial tcp 10.217.0.42:8080: connect: connection refused" start-of-body= Nov 28 13:21:31 crc kubenswrapper[4631]: I1128 13:21:31.794531 4631 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-2bc4b" podUID="9eb418ce-65b6-4da3-b282-831dd62916cf" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.42:8080/healthz\": dial tcp 10.217.0.42:8080: connect: connection refused" Nov 28 13:21:31 crc kubenswrapper[4631]: I1128 13:21:31.794568 4631 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-marketplace/marketplace-operator-79b997595-2bc4b" podUID="9eb418ce-65b6-4da3-b282-831dd62916cf" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.42:8080/healthz\": dial tcp 10.217.0.42:8080: connect: connection refused" Nov 28 13:21:31 crc kubenswrapper[4631]: I1128 13:21:31.863167 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5v5xp\" (UID: \"14717577-3fab-482b-bdbe-0e604d23d6d8\") " pod="openshift-image-registry/image-registry-697d97f7c8-5v5xp" Nov 28 13:21:31 crc kubenswrapper[4631]: E1128 13:21:31.863529 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 13:21:32.363512836 +0000 UTC m=+49.170816180 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5v5xp" (UID: "14717577-3fab-482b-bdbe-0e604d23d6d8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 13:21:31 crc kubenswrapper[4631]: I1128 13:21:31.957862 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-r6sdp" event={"ID":"9bc5a0ed-1419-4ba0-8478-9bcad6a19954","Type":"ContainerStarted","Data":"870e1213b19e2b4554aead8df3df550176a19bdf8047ef98a2c637a58eb9871e"} Nov 28 13:21:31 crc kubenswrapper[4631]: I1128 13:21:31.958200 4631 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-multus/cni-sysctl-allowlist-ds-44pwx" podUID="9504d491-6c27-4387-b6da-fa0192cc7a05" containerName="kube-multus-additional-cni-plugins" containerID="cri-o://3b12a8f186ea08cfb45b04b2d0d67de069fb809dcbf835a4dd4e008e4e24a4ae" gracePeriod=30 Nov 28 13:21:31 crc kubenswrapper[4631]: I1128 13:21:31.976330 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 13:21:31 crc kubenswrapper[4631]: E1128 13:21:31.976950 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 13:21:32.476930382 +0000 UTC m=+49.284233726 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 13:21:32 crc kubenswrapper[4631]: E1128 13:21:32.028959 4631 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="3b12a8f186ea08cfb45b04b2d0d67de069fb809dcbf835a4dd4e008e4e24a4ae" cmd=["/bin/bash","-c","test -f /ready/ready"] Nov 28 13:21:32 crc kubenswrapper[4631]: E1128 13:21:32.049428 4631 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="3b12a8f186ea08cfb45b04b2d0d67de069fb809dcbf835a4dd4e008e4e24a4ae" cmd=["/bin/bash","-c","test -f /ready/ready"] Nov 28 13:21:32 crc kubenswrapper[4631]: E1128 13:21:32.053710 4631 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="3b12a8f186ea08cfb45b04b2d0d67de069fb809dcbf835a4dd4e008e4e24a4ae" cmd=["/bin/bash","-c","test -f /ready/ready"] Nov 28 13:21:32 crc kubenswrapper[4631]: E1128 13:21:32.053936 4631 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openshift-multus/cni-sysctl-allowlist-ds-44pwx" podUID="9504d491-6c27-4387-b6da-fa0192cc7a05" containerName="kube-multus-additional-cni-plugins" Nov 28 13:21:32 crc kubenswrapper[4631]: I1128 13:21:32.079137 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5v5xp\" (UID: \"14717577-3fab-482b-bdbe-0e604d23d6d8\") " pod="openshift-image-registry/image-registry-697d97f7c8-5v5xp" Nov 28 13:21:32 crc kubenswrapper[4631]: I1128 13:21:32.083803 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-jcnpq" Nov 28 13:21:32 crc kubenswrapper[4631]: E1128 13:21:32.085144 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 13:21:32.585120227 +0000 UTC m=+49.392423571 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5v5xp" (UID: "14717577-3fab-482b-bdbe-0e604d23d6d8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 13:21:32 crc kubenswrapper[4631]: I1128 13:21:32.096460 4631 patch_prober.go:28] interesting pod/router-default-5444994796-jcnpq container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 28 13:21:32 crc kubenswrapper[4631]: [-]has-synced failed: reason withheld Nov 28 13:21:32 crc kubenswrapper[4631]: [+]process-running ok Nov 28 13:21:32 crc kubenswrapper[4631]: healthz check failed Nov 28 13:21:32 crc kubenswrapper[4631]: I1128 13:21:32.096516 4631 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-jcnpq" podUID="22010e1d-3ec1-4564-a362-bc963f691f1f" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 28 13:21:32 crc kubenswrapper[4631]: I1128 13:21:32.184680 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 13:21:32 crc kubenswrapper[4631]: E1128 13:21:32.184910 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 13:21:32.684875098 +0000 UTC m=+49.492178452 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 13:21:32 crc kubenswrapper[4631]: I1128 13:21:32.186532 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5v5xp\" (UID: \"14717577-3fab-482b-bdbe-0e604d23d6d8\") " pod="openshift-image-registry/image-registry-697d97f7c8-5v5xp" Nov 28 13:21:32 crc kubenswrapper[4631]: E1128 13:21:32.186955 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 13:21:32.686940096 +0000 UTC m=+49.494243590 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5v5xp" (UID: "14717577-3fab-482b-bdbe-0e604d23d6d8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 13:21:32 crc kubenswrapper[4631]: I1128 13:21:32.287994 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 13:21:32 crc kubenswrapper[4631]: E1128 13:21:32.288540 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 13:21:32.788513599 +0000 UTC m=+49.595816943 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 13:21:32 crc kubenswrapper[4631]: I1128 13:21:32.394413 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5v5xp\" (UID: \"14717577-3fab-482b-bdbe-0e604d23d6d8\") " pod="openshift-image-registry/image-registry-697d97f7c8-5v5xp" Nov 28 13:21:32 crc kubenswrapper[4631]: E1128 13:21:32.394789 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 13:21:32.894768769 +0000 UTC m=+49.702072113 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5v5xp" (UID: "14717577-3fab-482b-bdbe-0e604d23d6d8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 13:21:32 crc kubenswrapper[4631]: I1128 13:21:32.496055 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 13:21:32 crc kubenswrapper[4631]: E1128 13:21:32.496419 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 13:21:32.996400964 +0000 UTC m=+49.803704308 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 13:21:32 crc kubenswrapper[4631]: I1128 13:21:32.598060 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5v5xp\" (UID: \"14717577-3fab-482b-bdbe-0e604d23d6d8\") " pod="openshift-image-registry/image-registry-697d97f7c8-5v5xp" Nov 28 13:21:32 crc kubenswrapper[4631]: E1128 13:21:32.598467 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 13:21:33.098450238 +0000 UTC m=+49.905753582 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5v5xp" (UID: "14717577-3fab-482b-bdbe-0e604d23d6d8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 13:21:32 crc kubenswrapper[4631]: I1128 13:21:32.673862 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-tpjm2"] Nov 28 13:21:32 crc kubenswrapper[4631]: I1128 13:21:32.674948 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-tpjm2" Nov 28 13:21:32 crc kubenswrapper[4631]: I1128 13:21:32.683914 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Nov 28 13:21:32 crc kubenswrapper[4631]: I1128 13:21:32.692256 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-tpjm2"] Nov 28 13:21:32 crc kubenswrapper[4631]: I1128 13:21:32.698933 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 13:21:32 crc kubenswrapper[4631]: E1128 13:21:32.699095 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 13:21:33.199062149 +0000 UTC m=+50.006365493 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 13:21:32 crc kubenswrapper[4631]: I1128 13:21:32.699165 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5v5xp\" (UID: \"14717577-3fab-482b-bdbe-0e604d23d6d8\") " pod="openshift-image-registry/image-registry-697d97f7c8-5v5xp" Nov 28 13:21:32 crc kubenswrapper[4631]: E1128 13:21:32.699496 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 13:21:33.199484398 +0000 UTC m=+50.006787732 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5v5xp" (UID: "14717577-3fab-482b-bdbe-0e604d23d6d8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 13:21:32 crc kubenswrapper[4631]: I1128 13:21:32.784608 4631 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Nov 28 13:21:32 crc kubenswrapper[4631]: I1128 13:21:32.800549 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 13:21:32 crc kubenswrapper[4631]: E1128 13:21:32.800647 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 13:21:33.300631861 +0000 UTC m=+50.107935205 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 13:21:32 crc kubenswrapper[4631]: I1128 13:21:32.801376 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2fdd01cb-8e67-4bc4-aa2b-e023882103ce-utilities\") pod \"certified-operators-tpjm2\" (UID: \"2fdd01cb-8e67-4bc4-aa2b-e023882103ce\") " pod="openshift-marketplace/certified-operators-tpjm2" Nov 28 13:21:32 crc kubenswrapper[4631]: I1128 13:21:32.801487 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5v5xp\" (UID: \"14717577-3fab-482b-bdbe-0e604d23d6d8\") " pod="openshift-image-registry/image-registry-697d97f7c8-5v5xp" Nov 28 13:21:32 crc kubenswrapper[4631]: I1128 13:21:32.801514 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d52xh\" (UniqueName: \"kubernetes.io/projected/2fdd01cb-8e67-4bc4-aa2b-e023882103ce-kube-api-access-d52xh\") pod \"certified-operators-tpjm2\" (UID: \"2fdd01cb-8e67-4bc4-aa2b-e023882103ce\") " pod="openshift-marketplace/certified-operators-tpjm2" Nov 28 13:21:32 crc kubenswrapper[4631]: I1128 13:21:32.801554 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2fdd01cb-8e67-4bc4-aa2b-e023882103ce-catalog-content\") pod \"certified-operators-tpjm2\" (UID: \"2fdd01cb-8e67-4bc4-aa2b-e023882103ce\") " pod="openshift-marketplace/certified-operators-tpjm2" Nov 28 13:21:32 crc kubenswrapper[4631]: E1128 13:21:32.801840 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 13:21:33.301831339 +0000 UTC m=+50.109134683 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5v5xp" (UID: "14717577-3fab-482b-bdbe-0e604d23d6d8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 13:21:32 crc kubenswrapper[4631]: I1128 13:21:32.870507 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-4trst"] Nov 28 13:21:32 crc kubenswrapper[4631]: I1128 13:21:32.871693 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4trst" Nov 28 13:21:32 crc kubenswrapper[4631]: I1128 13:21:32.875901 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Nov 28 13:21:32 crc kubenswrapper[4631]: I1128 13:21:32.902877 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 13:21:32 crc kubenswrapper[4631]: I1128 13:21:32.903059 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2fdd01cb-8e67-4bc4-aa2b-e023882103ce-utilities\") pod \"certified-operators-tpjm2\" (UID: \"2fdd01cb-8e67-4bc4-aa2b-e023882103ce\") " pod="openshift-marketplace/certified-operators-tpjm2" Nov 28 13:21:32 crc kubenswrapper[4631]: I1128 13:21:32.903129 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d52xh\" (UniqueName: \"kubernetes.io/projected/2fdd01cb-8e67-4bc4-aa2b-e023882103ce-kube-api-access-d52xh\") pod \"certified-operators-tpjm2\" (UID: \"2fdd01cb-8e67-4bc4-aa2b-e023882103ce\") " pod="openshift-marketplace/certified-operators-tpjm2" Nov 28 13:21:32 crc kubenswrapper[4631]: I1128 13:21:32.903164 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2fdd01cb-8e67-4bc4-aa2b-e023882103ce-catalog-content\") pod \"certified-operators-tpjm2\" (UID: \"2fdd01cb-8e67-4bc4-aa2b-e023882103ce\") " pod="openshift-marketplace/certified-operators-tpjm2" Nov 28 13:21:32 crc kubenswrapper[4631]: I1128 13:21:32.903585 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2fdd01cb-8e67-4bc4-aa2b-e023882103ce-catalog-content\") pod \"certified-operators-tpjm2\" (UID: \"2fdd01cb-8e67-4bc4-aa2b-e023882103ce\") " pod="openshift-marketplace/certified-operators-tpjm2" Nov 28 13:21:32 crc kubenswrapper[4631]: E1128 13:21:32.903658 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 13:21:33.403641487 +0000 UTC m=+50.210944831 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 13:21:32 crc kubenswrapper[4631]: I1128 13:21:32.903917 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2fdd01cb-8e67-4bc4-aa2b-e023882103ce-utilities\") pod \"certified-operators-tpjm2\" (UID: \"2fdd01cb-8e67-4bc4-aa2b-e023882103ce\") " pod="openshift-marketplace/certified-operators-tpjm2" Nov 28 13:21:32 crc kubenswrapper[4631]: I1128 13:21:32.929096 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d52xh\" (UniqueName: \"kubernetes.io/projected/2fdd01cb-8e67-4bc4-aa2b-e023882103ce-kube-api-access-d52xh\") pod \"certified-operators-tpjm2\" (UID: \"2fdd01cb-8e67-4bc4-aa2b-e023882103ce\") " pod="openshift-marketplace/certified-operators-tpjm2" Nov 28 13:21:32 crc kubenswrapper[4631]: I1128 13:21:32.929467 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-4trst"] Nov 28 13:21:33 crc kubenswrapper[4631]: I1128 13:21:33.004279 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f6675fed-3a90-4779-9b95-1acba5248fd2-utilities\") pod \"community-operators-4trst\" (UID: \"f6675fed-3a90-4779-9b95-1acba5248fd2\") " pod="openshift-marketplace/community-operators-4trst" Nov 28 13:21:33 crc kubenswrapper[4631]: I1128 13:21:33.004518 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-njf98\" (UniqueName: \"kubernetes.io/projected/f6675fed-3a90-4779-9b95-1acba5248fd2-kube-api-access-njf98\") pod \"community-operators-4trst\" (UID: \"f6675fed-3a90-4779-9b95-1acba5248fd2\") " pod="openshift-marketplace/community-operators-4trst" Nov 28 13:21:33 crc kubenswrapper[4631]: I1128 13:21:33.004579 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5v5xp\" (UID: \"14717577-3fab-482b-bdbe-0e604d23d6d8\") " pod="openshift-image-registry/image-registry-697d97f7c8-5v5xp" Nov 28 13:21:33 crc kubenswrapper[4631]: I1128 13:21:33.004667 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f6675fed-3a90-4779-9b95-1acba5248fd2-catalog-content\") pod \"community-operators-4trst\" (UID: \"f6675fed-3a90-4779-9b95-1acba5248fd2\") " pod="openshift-marketplace/community-operators-4trst" Nov 28 13:21:33 crc kubenswrapper[4631]: E1128 13:21:33.004849 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 13:21:33.504835421 +0000 UTC m=+50.312138765 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5v5xp" (UID: "14717577-3fab-482b-bdbe-0e604d23d6d8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 13:21:33 crc kubenswrapper[4631]: I1128 13:21:33.011081 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-r6sdp" event={"ID":"9bc5a0ed-1419-4ba0-8478-9bcad6a19954","Type":"ContainerStarted","Data":"0741c9d414e6054bb7678d8b2d61203b2fc3ca093daeafe3a1eb24947b1a50a9"} Nov 28 13:21:33 crc kubenswrapper[4631]: I1128 13:21:33.011131 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-r6sdp" event={"ID":"9bc5a0ed-1419-4ba0-8478-9bcad6a19954","Type":"ContainerStarted","Data":"e2f199c853f788a341549cc0d3cd46cd8516dc382d74216d07abc5b48fad2ed0"} Nov 28 13:21:33 crc kubenswrapper[4631]: I1128 13:21:33.014279 4631 generic.go:334] "Generic (PLEG): container finished" podID="76a05ca6-a98c-4313-855c-fd6ffaceee4f" containerID="0f3af793008c1a47e8009812f16cb978e3930b0e22afc292634c25e32a447f21" exitCode=0 Nov 28 13:21:33 crc kubenswrapper[4631]: I1128 13:21:33.014361 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29405595-2v6fk" event={"ID":"76a05ca6-a98c-4313-855c-fd6ffaceee4f","Type":"ContainerDied","Data":"0f3af793008c1a47e8009812f16cb978e3930b0e22afc292634c25e32a447f21"} Nov 28 13:21:33 crc kubenswrapper[4631]: I1128 13:21:33.023540 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-tpjm2" Nov 28 13:21:33 crc kubenswrapper[4631]: I1128 13:21:33.029534 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-268tw" Nov 28 13:21:33 crc kubenswrapper[4631]: I1128 13:21:33.029708 4631 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 28 13:21:33 crc kubenswrapper[4631]: I1128 13:21:33.030182 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-r6sdp" podStartSLOduration=14.030166155 podStartE2EDuration="14.030166155s" podCreationTimestamp="2025-11-28 13:21:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 13:21:33.028461066 +0000 UTC m=+49.835764410" watchObservedRunningTime="2025-11-28 13:21:33.030166155 +0000 UTC m=+49.837469499" Nov 28 13:21:33 crc kubenswrapper[4631]: I1128 13:21:33.061319 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-268tw" Nov 28 13:21:33 crc kubenswrapper[4631]: I1128 13:21:33.086086 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-zkczn"] Nov 28 13:21:33 crc kubenswrapper[4631]: I1128 13:21:33.088930 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zkczn" Nov 28 13:21:33 crc kubenswrapper[4631]: I1128 13:21:33.089334 4631 patch_prober.go:28] interesting pod/router-default-5444994796-jcnpq container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 28 13:21:33 crc kubenswrapper[4631]: [-]has-synced failed: reason withheld Nov 28 13:21:33 crc kubenswrapper[4631]: [+]process-running ok Nov 28 13:21:33 crc kubenswrapper[4631]: healthz check failed Nov 28 13:21:33 crc kubenswrapper[4631]: I1128 13:21:33.089391 4631 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-jcnpq" podUID="22010e1d-3ec1-4564-a362-bc963f691f1f" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 28 13:21:33 crc kubenswrapper[4631]: I1128 13:21:33.105433 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 13:21:33 crc kubenswrapper[4631]: E1128 13:21:33.105561 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 13:21:33.605540734 +0000 UTC m=+50.412844078 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 13:21:33 crc kubenswrapper[4631]: I1128 13:21:33.105902 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f6675fed-3a90-4779-9b95-1acba5248fd2-utilities\") pod \"community-operators-4trst\" (UID: \"f6675fed-3a90-4779-9b95-1acba5248fd2\") " pod="openshift-marketplace/community-operators-4trst" Nov 28 13:21:33 crc kubenswrapper[4631]: I1128 13:21:33.106022 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-njf98\" (UniqueName: \"kubernetes.io/projected/f6675fed-3a90-4779-9b95-1acba5248fd2-kube-api-access-njf98\") pod \"community-operators-4trst\" (UID: \"f6675fed-3a90-4779-9b95-1acba5248fd2\") " pod="openshift-marketplace/community-operators-4trst" Nov 28 13:21:33 crc kubenswrapper[4631]: I1128 13:21:33.106043 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5v5xp\" (UID: \"14717577-3fab-482b-bdbe-0e604d23d6d8\") " pod="openshift-image-registry/image-registry-697d97f7c8-5v5xp" Nov 28 13:21:33 crc kubenswrapper[4631]: I1128 13:21:33.106117 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f6675fed-3a90-4779-9b95-1acba5248fd2-catalog-content\") pod \"community-operators-4trst\" (UID: \"f6675fed-3a90-4779-9b95-1acba5248fd2\") " pod="openshift-marketplace/community-operators-4trst" Nov 28 13:21:33 crc kubenswrapper[4631]: E1128 13:21:33.106576 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 13:21:33.606560168 +0000 UTC m=+50.413863512 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5v5xp" (UID: "14717577-3fab-482b-bdbe-0e604d23d6d8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 13:21:33 crc kubenswrapper[4631]: I1128 13:21:33.107341 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f6675fed-3a90-4779-9b95-1acba5248fd2-utilities\") pod \"community-operators-4trst\" (UID: \"f6675fed-3a90-4779-9b95-1acba5248fd2\") " pod="openshift-marketplace/community-operators-4trst" Nov 28 13:21:33 crc kubenswrapper[4631]: I1128 13:21:33.107465 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f6675fed-3a90-4779-9b95-1acba5248fd2-catalog-content\") pod \"community-operators-4trst\" (UID: \"f6675fed-3a90-4779-9b95-1acba5248fd2\") " pod="openshift-marketplace/community-operators-4trst" Nov 28 13:21:33 crc kubenswrapper[4631]: I1128 13:21:33.126339 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-zkczn"] Nov 28 13:21:33 crc kubenswrapper[4631]: I1128 13:21:33.130617 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-njf98\" (UniqueName: \"kubernetes.io/projected/f6675fed-3a90-4779-9b95-1acba5248fd2-kube-api-access-njf98\") pod \"community-operators-4trst\" (UID: \"f6675fed-3a90-4779-9b95-1acba5248fd2\") " pod="openshift-marketplace/community-operators-4trst" Nov 28 13:21:33 crc kubenswrapper[4631]: I1128 13:21:33.184623 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4trst" Nov 28 13:21:33 crc kubenswrapper[4631]: I1128 13:21:33.207382 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 13:21:33 crc kubenswrapper[4631]: I1128 13:21:33.207531 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/10dbdccc-7009-407e-846d-8a2976a58fad-utilities\") pod \"certified-operators-zkczn\" (UID: \"10dbdccc-7009-407e-846d-8a2976a58fad\") " pod="openshift-marketplace/certified-operators-zkczn" Nov 28 13:21:33 crc kubenswrapper[4631]: I1128 13:21:33.207556 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tfshw\" (UniqueName: \"kubernetes.io/projected/10dbdccc-7009-407e-846d-8a2976a58fad-kube-api-access-tfshw\") pod \"certified-operators-zkczn\" (UID: \"10dbdccc-7009-407e-846d-8a2976a58fad\") " pod="openshift-marketplace/certified-operators-zkczn" Nov 28 13:21:33 crc kubenswrapper[4631]: I1128 13:21:33.207608 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/10dbdccc-7009-407e-846d-8a2976a58fad-catalog-content\") pod \"certified-operators-zkczn\" (UID: \"10dbdccc-7009-407e-846d-8a2976a58fad\") " pod="openshift-marketplace/certified-operators-zkczn" Nov 28 13:21:33 crc kubenswrapper[4631]: E1128 13:21:33.207735 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 13:21:33.707720781 +0000 UTC m=+50.515024125 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 13:21:33 crc kubenswrapper[4631]: I1128 13:21:33.273613 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-6nrc5"] Nov 28 13:21:33 crc kubenswrapper[4631]: I1128 13:21:33.277357 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-6nrc5"] Nov 28 13:21:33 crc kubenswrapper[4631]: I1128 13:21:33.277448 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6nrc5" Nov 28 13:21:33 crc kubenswrapper[4631]: I1128 13:21:33.308648 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/10dbdccc-7009-407e-846d-8a2976a58fad-utilities\") pod \"certified-operators-zkczn\" (UID: \"10dbdccc-7009-407e-846d-8a2976a58fad\") " pod="openshift-marketplace/certified-operators-zkczn" Nov 28 13:21:33 crc kubenswrapper[4631]: I1128 13:21:33.308688 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tfshw\" (UniqueName: \"kubernetes.io/projected/10dbdccc-7009-407e-846d-8a2976a58fad-kube-api-access-tfshw\") pod \"certified-operators-zkczn\" (UID: \"10dbdccc-7009-407e-846d-8a2976a58fad\") " pod="openshift-marketplace/certified-operators-zkczn" Nov 28 13:21:33 crc kubenswrapper[4631]: I1128 13:21:33.308725 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qpqxs\" (UniqueName: \"kubernetes.io/projected/72b60e37-5508-451f-8602-9dcebd55324e-kube-api-access-qpqxs\") pod \"community-operators-6nrc5\" (UID: \"72b60e37-5508-451f-8602-9dcebd55324e\") " pod="openshift-marketplace/community-operators-6nrc5" Nov 28 13:21:33 crc kubenswrapper[4631]: I1128 13:21:33.308760 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/10dbdccc-7009-407e-846d-8a2976a58fad-catalog-content\") pod \"certified-operators-zkczn\" (UID: \"10dbdccc-7009-407e-846d-8a2976a58fad\") " pod="openshift-marketplace/certified-operators-zkczn" Nov 28 13:21:33 crc kubenswrapper[4631]: I1128 13:21:33.308788 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/72b60e37-5508-451f-8602-9dcebd55324e-utilities\") pod \"community-operators-6nrc5\" (UID: \"72b60e37-5508-451f-8602-9dcebd55324e\") " pod="openshift-marketplace/community-operators-6nrc5" Nov 28 13:21:33 crc kubenswrapper[4631]: I1128 13:21:33.308811 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/72b60e37-5508-451f-8602-9dcebd55324e-catalog-content\") pod \"community-operators-6nrc5\" (UID: \"72b60e37-5508-451f-8602-9dcebd55324e\") " pod="openshift-marketplace/community-operators-6nrc5" Nov 28 13:21:33 crc kubenswrapper[4631]: I1128 13:21:33.308838 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5v5xp\" (UID: \"14717577-3fab-482b-bdbe-0e604d23d6d8\") " pod="openshift-image-registry/image-registry-697d97f7c8-5v5xp" Nov 28 13:21:33 crc kubenswrapper[4631]: E1128 13:21:33.309093 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 13:21:33.809081968 +0000 UTC m=+50.616385312 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5v5xp" (UID: "14717577-3fab-482b-bdbe-0e604d23d6d8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 13:21:33 crc kubenswrapper[4631]: I1128 13:21:33.309795 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/10dbdccc-7009-407e-846d-8a2976a58fad-utilities\") pod \"certified-operators-zkczn\" (UID: \"10dbdccc-7009-407e-846d-8a2976a58fad\") " pod="openshift-marketplace/certified-operators-zkczn" Nov 28 13:21:33 crc kubenswrapper[4631]: I1128 13:21:33.310215 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/10dbdccc-7009-407e-846d-8a2976a58fad-catalog-content\") pod \"certified-operators-zkczn\" (UID: \"10dbdccc-7009-407e-846d-8a2976a58fad\") " pod="openshift-marketplace/certified-operators-zkczn" Nov 28 13:21:33 crc kubenswrapper[4631]: I1128 13:21:33.336039 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tfshw\" (UniqueName: \"kubernetes.io/projected/10dbdccc-7009-407e-846d-8a2976a58fad-kube-api-access-tfshw\") pod \"certified-operators-zkczn\" (UID: \"10dbdccc-7009-407e-846d-8a2976a58fad\") " pod="openshift-marketplace/certified-operators-zkczn" Nov 28 13:21:33 crc kubenswrapper[4631]: I1128 13:21:33.393837 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-tpjm2"] Nov 28 13:21:33 crc kubenswrapper[4631]: I1128 13:21:33.407899 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zkczn" Nov 28 13:21:33 crc kubenswrapper[4631]: I1128 13:21:33.409538 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 13:21:33 crc kubenswrapper[4631]: I1128 13:21:33.409842 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qpqxs\" (UniqueName: \"kubernetes.io/projected/72b60e37-5508-451f-8602-9dcebd55324e-kube-api-access-qpqxs\") pod \"community-operators-6nrc5\" (UID: \"72b60e37-5508-451f-8602-9dcebd55324e\") " pod="openshift-marketplace/community-operators-6nrc5" Nov 28 13:21:33 crc kubenswrapper[4631]: I1128 13:21:33.409891 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/72b60e37-5508-451f-8602-9dcebd55324e-utilities\") pod \"community-operators-6nrc5\" (UID: \"72b60e37-5508-451f-8602-9dcebd55324e\") " pod="openshift-marketplace/community-operators-6nrc5" Nov 28 13:21:33 crc kubenswrapper[4631]: I1128 13:21:33.409914 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/72b60e37-5508-451f-8602-9dcebd55324e-catalog-content\") pod \"community-operators-6nrc5\" (UID: \"72b60e37-5508-451f-8602-9dcebd55324e\") " pod="openshift-marketplace/community-operators-6nrc5" Nov 28 13:21:33 crc kubenswrapper[4631]: I1128 13:21:33.410329 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/72b60e37-5508-451f-8602-9dcebd55324e-catalog-content\") pod \"community-operators-6nrc5\" (UID: \"72b60e37-5508-451f-8602-9dcebd55324e\") " pod="openshift-marketplace/community-operators-6nrc5" Nov 28 13:21:33 crc kubenswrapper[4631]: E1128 13:21:33.410399 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 13:21:33.910385495 +0000 UTC m=+50.717688839 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 13:21:33 crc kubenswrapper[4631]: I1128 13:21:33.410844 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/72b60e37-5508-451f-8602-9dcebd55324e-utilities\") pod \"community-operators-6nrc5\" (UID: \"72b60e37-5508-451f-8602-9dcebd55324e\") " pod="openshift-marketplace/community-operators-6nrc5" Nov 28 13:21:33 crc kubenswrapper[4631]: I1128 13:21:33.429998 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qpqxs\" (UniqueName: \"kubernetes.io/projected/72b60e37-5508-451f-8602-9dcebd55324e-kube-api-access-qpqxs\") pod \"community-operators-6nrc5\" (UID: \"72b60e37-5508-451f-8602-9dcebd55324e\") " pod="openshift-marketplace/community-operators-6nrc5" Nov 28 13:21:33 crc kubenswrapper[4631]: I1128 13:21:33.510899 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5v5xp\" (UID: \"14717577-3fab-482b-bdbe-0e604d23d6d8\") " pod="openshift-image-registry/image-registry-697d97f7c8-5v5xp" Nov 28 13:21:33 crc kubenswrapper[4631]: E1128 13:21:33.511152 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 13:21:34.011140469 +0000 UTC m=+50.818443813 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5v5xp" (UID: "14717577-3fab-482b-bdbe-0e604d23d6d8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 13:21:33 crc kubenswrapper[4631]: I1128 13:21:33.597559 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6nrc5" Nov 28 13:21:33 crc kubenswrapper[4631]: I1128 13:21:33.616337 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 13:21:33 crc kubenswrapper[4631]: E1128 13:21:33.616660 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 13:21:34.116621692 +0000 UTC m=+50.923925046 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 13:21:33 crc kubenswrapper[4631]: I1128 13:21:33.616754 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5v5xp\" (UID: \"14717577-3fab-482b-bdbe-0e604d23d6d8\") " pod="openshift-image-registry/image-registry-697d97f7c8-5v5xp" Nov 28 13:21:33 crc kubenswrapper[4631]: E1128 13:21:33.617087 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 13:21:34.117074772 +0000 UTC m=+50.924378106 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5v5xp" (UID: "14717577-3fab-482b-bdbe-0e604d23d6d8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 13:21:33 crc kubenswrapper[4631]: I1128 13:21:33.619972 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-zkczn"] Nov 28 13:21:33 crc kubenswrapper[4631]: I1128 13:21:33.650242 4631 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2025-11-28T13:21:32.784631692Z","Handler":null,"Name":""} Nov 28 13:21:33 crc kubenswrapper[4631]: I1128 13:21:33.677995 4631 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Nov 28 13:21:33 crc kubenswrapper[4631]: I1128 13:21:33.678028 4631 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Nov 28 13:21:33 crc kubenswrapper[4631]: I1128 13:21:33.717814 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 13:21:33 crc kubenswrapper[4631]: I1128 13:21:33.723840 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Nov 28 13:21:33 crc kubenswrapper[4631]: I1128 13:21:33.726609 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-4trst"] Nov 28 13:21:33 crc kubenswrapper[4631]: I1128 13:21:33.819615 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5v5xp\" (UID: \"14717577-3fab-482b-bdbe-0e604d23d6d8\") " pod="openshift-image-registry/image-registry-697d97f7c8-5v5xp" Nov 28 13:21:33 crc kubenswrapper[4631]: I1128 13:21:33.847086 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-6nrc5"] Nov 28 13:21:33 crc kubenswrapper[4631]: I1128 13:21:33.847523 4631 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Nov 28 13:21:33 crc kubenswrapper[4631]: I1128 13:21:33.847565 4631 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5v5xp\" (UID: \"14717577-3fab-482b-bdbe-0e604d23d6d8\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-5v5xp" Nov 28 13:21:33 crc kubenswrapper[4631]: I1128 13:21:33.877108 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5v5xp\" (UID: \"14717577-3fab-482b-bdbe-0e604d23d6d8\") " pod="openshift-image-registry/image-registry-697d97f7c8-5v5xp" Nov 28 13:21:33 crc kubenswrapper[4631]: W1128 13:21:33.925610 4631 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod72b60e37_5508_451f_8602_9dcebd55324e.slice/crio-98e79530ae258d5c9d111947af766c51808513037656c4bc1cb0801f7a69bbb3 WatchSource:0}: Error finding container 98e79530ae258d5c9d111947af766c51808513037656c4bc1cb0801f7a69bbb3: Status 404 returned error can't find the container with id 98e79530ae258d5c9d111947af766c51808513037656c4bc1cb0801f7a69bbb3 Nov 28 13:21:34 crc kubenswrapper[4631]: I1128 13:21:34.044918 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-5v5xp" Nov 28 13:21:34 crc kubenswrapper[4631]: I1128 13:21:34.049314 4631 generic.go:334] "Generic (PLEG): container finished" podID="10dbdccc-7009-407e-846d-8a2976a58fad" containerID="a844c003570d7e42551cdfb77ab5afad3faade0c2383aaa0bba7016962d0d765" exitCode=0 Nov 28 13:21:34 crc kubenswrapper[4631]: I1128 13:21:34.049382 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zkczn" event={"ID":"10dbdccc-7009-407e-846d-8a2976a58fad","Type":"ContainerDied","Data":"a844c003570d7e42551cdfb77ab5afad3faade0c2383aaa0bba7016962d0d765"} Nov 28 13:21:34 crc kubenswrapper[4631]: I1128 13:21:34.049406 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zkczn" event={"ID":"10dbdccc-7009-407e-846d-8a2976a58fad","Type":"ContainerStarted","Data":"f4d5d88ce6700a920b115de3d73ffb17405cef6dae06bfcdb52097b5ab21fe95"} Nov 28 13:21:34 crc kubenswrapper[4631]: I1128 13:21:34.057643 4631 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 28 13:21:34 crc kubenswrapper[4631]: I1128 13:21:34.059887 4631 generic.go:334] "Generic (PLEG): container finished" podID="2fdd01cb-8e67-4bc4-aa2b-e023882103ce" containerID="b402075f9cf569ae47999d890b945c40cafe4f81f7a434335ab9de307a1e0b90" exitCode=0 Nov 28 13:21:34 crc kubenswrapper[4631]: I1128 13:21:34.059949 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tpjm2" event={"ID":"2fdd01cb-8e67-4bc4-aa2b-e023882103ce","Type":"ContainerDied","Data":"b402075f9cf569ae47999d890b945c40cafe4f81f7a434335ab9de307a1e0b90"} Nov 28 13:21:34 crc kubenswrapper[4631]: I1128 13:21:34.059975 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tpjm2" event={"ID":"2fdd01cb-8e67-4bc4-aa2b-e023882103ce","Type":"ContainerStarted","Data":"380a08fd5f1ebb2bde91ea8aa5fea36fbeac716fb80ea119fefec1ea3997cea3"} Nov 28 13:21:34 crc kubenswrapper[4631]: I1128 13:21:34.062987 4631 generic.go:334] "Generic (PLEG): container finished" podID="f6675fed-3a90-4779-9b95-1acba5248fd2" containerID="0a6ee6c8cd381cbe0f8320e8ebb7c0d4d336a59006118461ae8a6458498aa3d6" exitCode=0 Nov 28 13:21:34 crc kubenswrapper[4631]: I1128 13:21:34.063043 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4trst" event={"ID":"f6675fed-3a90-4779-9b95-1acba5248fd2","Type":"ContainerDied","Data":"0a6ee6c8cd381cbe0f8320e8ebb7c0d4d336a59006118461ae8a6458498aa3d6"} Nov 28 13:21:34 crc kubenswrapper[4631]: I1128 13:21:34.063130 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4trst" event={"ID":"f6675fed-3a90-4779-9b95-1acba5248fd2","Type":"ContainerStarted","Data":"70d6a2a429d0a2a96315f02b1a28caede0861023f7293afcc5c58f2927d41a1c"} Nov 28 13:21:34 crc kubenswrapper[4631]: I1128 13:21:34.077181 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6nrc5" event={"ID":"72b60e37-5508-451f-8602-9dcebd55324e","Type":"ContainerStarted","Data":"e0cadea03b1d6a0341c5e2f626ac4163ccb8542fb464bf13aeffc51cc43f1ede"} Nov 28 13:21:34 crc kubenswrapper[4631]: I1128 13:21:34.077474 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6nrc5" event={"ID":"72b60e37-5508-451f-8602-9dcebd55324e","Type":"ContainerStarted","Data":"98e79530ae258d5c9d111947af766c51808513037656c4bc1cb0801f7a69bbb3"} Nov 28 13:21:34 crc kubenswrapper[4631]: I1128 13:21:34.096407 4631 patch_prober.go:28] interesting pod/router-default-5444994796-jcnpq container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 28 13:21:34 crc kubenswrapper[4631]: [-]has-synced failed: reason withheld Nov 28 13:21:34 crc kubenswrapper[4631]: [+]process-running ok Nov 28 13:21:34 crc kubenswrapper[4631]: healthz check failed Nov 28 13:21:34 crc kubenswrapper[4631]: I1128 13:21:34.096457 4631 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-jcnpq" podUID="22010e1d-3ec1-4564-a362-bc963f691f1f" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 28 13:21:34 crc kubenswrapper[4631]: I1128 13:21:34.367585 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Nov 28 13:21:34 crc kubenswrapper[4631]: I1128 13:21:34.368254 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 28 13:21:34 crc kubenswrapper[4631]: I1128 13:21:34.370639 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Nov 28 13:21:34 crc kubenswrapper[4631]: I1128 13:21:34.370823 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Nov 28 13:21:34 crc kubenswrapper[4631]: I1128 13:21:34.384052 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Nov 28 13:21:34 crc kubenswrapper[4631]: I1128 13:21:34.431219 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f99b964d-d26f-452f-8ce4-754c09023a6a-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"f99b964d-d26f-452f-8ce4-754c09023a6a\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 28 13:21:34 crc kubenswrapper[4631]: I1128 13:21:34.431267 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 13:21:34 crc kubenswrapper[4631]: I1128 13:21:34.431309 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 13:21:34 crc kubenswrapper[4631]: I1128 13:21:34.431343 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 13:21:34 crc kubenswrapper[4631]: I1128 13:21:34.431360 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/f99b964d-d26f-452f-8ce4-754c09023a6a-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"f99b964d-d26f-452f-8ce4-754c09023a6a\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 28 13:21:34 crc kubenswrapper[4631]: I1128 13:21:34.431387 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 13:21:34 crc kubenswrapper[4631]: I1128 13:21:34.438542 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 13:21:34 crc kubenswrapper[4631]: I1128 13:21:34.440145 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 13:21:34 crc kubenswrapper[4631]: I1128 13:21:34.445693 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 13:21:34 crc kubenswrapper[4631]: I1128 13:21:34.449839 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 13:21:34 crc kubenswrapper[4631]: I1128 13:21:34.533041 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/f99b964d-d26f-452f-8ce4-754c09023a6a-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"f99b964d-d26f-452f-8ce4-754c09023a6a\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 28 13:21:34 crc kubenswrapper[4631]: I1128 13:21:34.533151 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f99b964d-d26f-452f-8ce4-754c09023a6a-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"f99b964d-d26f-452f-8ce4-754c09023a6a\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 28 13:21:34 crc kubenswrapper[4631]: I1128 13:21:34.533164 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/f99b964d-d26f-452f-8ce4-754c09023a6a-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"f99b964d-d26f-452f-8ce4-754c09023a6a\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 28 13:21:34 crc kubenswrapper[4631]: I1128 13:21:34.565331 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f99b964d-d26f-452f-8ce4-754c09023a6a-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"f99b964d-d26f-452f-8ce4-754c09023a6a\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 28 13:21:34 crc kubenswrapper[4631]: I1128 13:21:34.610037 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-5v5xp"] Nov 28 13:21:34 crc kubenswrapper[4631]: I1128 13:21:34.647549 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 13:21:34 crc kubenswrapper[4631]: I1128 13:21:34.728535 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 28 13:21:34 crc kubenswrapper[4631]: I1128 13:21:34.813388 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8tbnv" Nov 28 13:21:34 crc kubenswrapper[4631]: I1128 13:21:34.813427 4631 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8tbnv" Nov 28 13:21:34 crc kubenswrapper[4631]: I1128 13:21:34.824453 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-jgx6f" Nov 28 13:21:34 crc kubenswrapper[4631]: I1128 13:21:34.825394 4631 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-jgx6f" Nov 28 13:21:34 crc kubenswrapper[4631]: I1128 13:21:34.834519 4631 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8tbnv" Nov 28 13:21:34 crc kubenswrapper[4631]: I1128 13:21:34.840887 4631 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-jgx6f" Nov 28 13:21:34 crc kubenswrapper[4631]: I1128 13:21:34.891383 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-p5v2z"] Nov 28 13:21:34 crc kubenswrapper[4631]: I1128 13:21:34.905626 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-p5v2z" Nov 28 13:21:34 crc kubenswrapper[4631]: I1128 13:21:34.908059 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Nov 28 13:21:34 crc kubenswrapper[4631]: I1128 13:21:34.917691 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-p5v2z"] Nov 28 13:21:34 crc kubenswrapper[4631]: I1128 13:21:34.937898 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/67afe4ac-28a7-42a2-8155-2c4bf496b834-utilities\") pod \"redhat-marketplace-p5v2z\" (UID: \"67afe4ac-28a7-42a2-8155-2c4bf496b834\") " pod="openshift-marketplace/redhat-marketplace-p5v2z" Nov 28 13:21:34 crc kubenswrapper[4631]: I1128 13:21:34.937975 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/67afe4ac-28a7-42a2-8155-2c4bf496b834-catalog-content\") pod \"redhat-marketplace-p5v2z\" (UID: \"67afe4ac-28a7-42a2-8155-2c4bf496b834\") " pod="openshift-marketplace/redhat-marketplace-p5v2z" Nov 28 13:21:34 crc kubenswrapper[4631]: I1128 13:21:34.938093 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9lkb4\" (UniqueName: \"kubernetes.io/projected/67afe4ac-28a7-42a2-8155-2c4bf496b834-kube-api-access-9lkb4\") pod \"redhat-marketplace-p5v2z\" (UID: \"67afe4ac-28a7-42a2-8155-2c4bf496b834\") " pod="openshift-marketplace/redhat-marketplace-p5v2z" Nov 28 13:21:34 crc kubenswrapper[4631]: I1128 13:21:34.982957 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 13:21:35 crc kubenswrapper[4631]: W1128 13:21:35.000202 4631 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9d751cbb_f2e2_430d_9754_c882a5e924a5.slice/crio-06cc7a7a96c0a1d15a6cae19743df14a1e7892661e3431a9c083654c84a1a5d9 WatchSource:0}: Error finding container 06cc7a7a96c0a1d15a6cae19743df14a1e7892661e3431a9c083654c84a1a5d9: Status 404 returned error can't find the container with id 06cc7a7a96c0a1d15a6cae19743df14a1e7892661e3431a9c083654c84a1a5d9 Nov 28 13:21:35 crc kubenswrapper[4631]: I1128 13:21:35.006947 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29405595-2v6fk" Nov 28 13:21:35 crc kubenswrapper[4631]: I1128 13:21:35.038725 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b54w6\" (UniqueName: \"kubernetes.io/projected/76a05ca6-a98c-4313-855c-fd6ffaceee4f-kube-api-access-b54w6\") pod \"76a05ca6-a98c-4313-855c-fd6ffaceee4f\" (UID: \"76a05ca6-a98c-4313-855c-fd6ffaceee4f\") " Nov 28 13:21:35 crc kubenswrapper[4631]: I1128 13:21:35.038790 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/76a05ca6-a98c-4313-855c-fd6ffaceee4f-secret-volume\") pod \"76a05ca6-a98c-4313-855c-fd6ffaceee4f\" (UID: \"76a05ca6-a98c-4313-855c-fd6ffaceee4f\") " Nov 28 13:21:35 crc kubenswrapper[4631]: I1128 13:21:35.038849 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/76a05ca6-a98c-4313-855c-fd6ffaceee4f-config-volume\") pod \"76a05ca6-a98c-4313-855c-fd6ffaceee4f\" (UID: \"76a05ca6-a98c-4313-855c-fd6ffaceee4f\") " Nov 28 13:21:35 crc kubenswrapper[4631]: I1128 13:21:35.038970 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/67afe4ac-28a7-42a2-8155-2c4bf496b834-utilities\") pod \"redhat-marketplace-p5v2z\" (UID: \"67afe4ac-28a7-42a2-8155-2c4bf496b834\") " pod="openshift-marketplace/redhat-marketplace-p5v2z" Nov 28 13:21:35 crc kubenswrapper[4631]: I1128 13:21:35.039003 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/67afe4ac-28a7-42a2-8155-2c4bf496b834-catalog-content\") pod \"redhat-marketplace-p5v2z\" (UID: \"67afe4ac-28a7-42a2-8155-2c4bf496b834\") " pod="openshift-marketplace/redhat-marketplace-p5v2z" Nov 28 13:21:35 crc kubenswrapper[4631]: I1128 13:21:35.039148 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9lkb4\" (UniqueName: \"kubernetes.io/projected/67afe4ac-28a7-42a2-8155-2c4bf496b834-kube-api-access-9lkb4\") pod \"redhat-marketplace-p5v2z\" (UID: \"67afe4ac-28a7-42a2-8155-2c4bf496b834\") " pod="openshift-marketplace/redhat-marketplace-p5v2z" Nov 28 13:21:35 crc kubenswrapper[4631]: I1128 13:21:35.040085 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/67afe4ac-28a7-42a2-8155-2c4bf496b834-utilities\") pod \"redhat-marketplace-p5v2z\" (UID: \"67afe4ac-28a7-42a2-8155-2c4bf496b834\") " pod="openshift-marketplace/redhat-marketplace-p5v2z" Nov 28 13:21:35 crc kubenswrapper[4631]: I1128 13:21:35.040247 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/76a05ca6-a98c-4313-855c-fd6ffaceee4f-config-volume" (OuterVolumeSpecName: "config-volume") pod "76a05ca6-a98c-4313-855c-fd6ffaceee4f" (UID: "76a05ca6-a98c-4313-855c-fd6ffaceee4f"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 13:21:35 crc kubenswrapper[4631]: I1128 13:21:35.040388 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/67afe4ac-28a7-42a2-8155-2c4bf496b834-catalog-content\") pod \"redhat-marketplace-p5v2z\" (UID: \"67afe4ac-28a7-42a2-8155-2c4bf496b834\") " pod="openshift-marketplace/redhat-marketplace-p5v2z" Nov 28 13:21:35 crc kubenswrapper[4631]: I1128 13:21:35.043748 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/76a05ca6-a98c-4313-855c-fd6ffaceee4f-kube-api-access-b54w6" (OuterVolumeSpecName: "kube-api-access-b54w6") pod "76a05ca6-a98c-4313-855c-fd6ffaceee4f" (UID: "76a05ca6-a98c-4313-855c-fd6ffaceee4f"). InnerVolumeSpecName "kube-api-access-b54w6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:21:35 crc kubenswrapper[4631]: I1128 13:21:35.055294 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9lkb4\" (UniqueName: \"kubernetes.io/projected/67afe4ac-28a7-42a2-8155-2c4bf496b834-kube-api-access-9lkb4\") pod \"redhat-marketplace-p5v2z\" (UID: \"67afe4ac-28a7-42a2-8155-2c4bf496b834\") " pod="openshift-marketplace/redhat-marketplace-p5v2z" Nov 28 13:21:35 crc kubenswrapper[4631]: I1128 13:21:35.057625 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/76a05ca6-a98c-4313-855c-fd6ffaceee4f-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "76a05ca6-a98c-4313-855c-fd6ffaceee4f" (UID: "76a05ca6-a98c-4313-855c-fd6ffaceee4f"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:21:35 crc kubenswrapper[4631]: I1128 13:21:35.086164 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"06cc7a7a96c0a1d15a6cae19743df14a1e7892661e3431a9c083654c84a1a5d9"} Nov 28 13:21:35 crc kubenswrapper[4631]: I1128 13:21:35.087690 4631 patch_prober.go:28] interesting pod/router-default-5444994796-jcnpq container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 28 13:21:35 crc kubenswrapper[4631]: [-]has-synced failed: reason withheld Nov 28 13:21:35 crc kubenswrapper[4631]: [+]process-running ok Nov 28 13:21:35 crc kubenswrapper[4631]: healthz check failed Nov 28 13:21:35 crc kubenswrapper[4631]: I1128 13:21:35.087738 4631 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-jcnpq" podUID="22010e1d-3ec1-4564-a362-bc963f691f1f" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 28 13:21:35 crc kubenswrapper[4631]: I1128 13:21:35.093458 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29405595-2v6fk" event={"ID":"76a05ca6-a98c-4313-855c-fd6ffaceee4f","Type":"ContainerDied","Data":"f75613555c90f9c4c2505bf8ec5be1a675aed747146be2b68c6f8ada4d4ba88c"} Nov 28 13:21:35 crc kubenswrapper[4631]: I1128 13:21:35.093491 4631 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f75613555c90f9c4c2505bf8ec5be1a675aed747146be2b68c6f8ada4d4ba88c" Nov 28 13:21:35 crc kubenswrapper[4631]: I1128 13:21:35.093561 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29405595-2v6fk" Nov 28 13:21:35 crc kubenswrapper[4631]: I1128 13:21:35.114834 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-5v5xp" event={"ID":"14717577-3fab-482b-bdbe-0e604d23d6d8","Type":"ContainerStarted","Data":"6f8e107523760a208fc044fa471cb14ffbde0fe03ac81ffabdf1604ae7e2582a"} Nov 28 13:21:35 crc kubenswrapper[4631]: I1128 13:21:35.122032 4631 generic.go:334] "Generic (PLEG): container finished" podID="72b60e37-5508-451f-8602-9dcebd55324e" containerID="e0cadea03b1d6a0341c5e2f626ac4163ccb8542fb464bf13aeffc51cc43f1ede" exitCode=0 Nov 28 13:21:35 crc kubenswrapper[4631]: I1128 13:21:35.124396 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6nrc5" event={"ID":"72b60e37-5508-451f-8602-9dcebd55324e","Type":"ContainerDied","Data":"e0cadea03b1d6a0341c5e2f626ac4163ccb8542fb464bf13aeffc51cc43f1ede"} Nov 28 13:21:35 crc kubenswrapper[4631]: I1128 13:21:35.131807 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-jgx6f" Nov 28 13:21:35 crc kubenswrapper[4631]: I1128 13:21:35.140006 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b54w6\" (UniqueName: \"kubernetes.io/projected/76a05ca6-a98c-4313-855c-fd6ffaceee4f-kube-api-access-b54w6\") on node \"crc\" DevicePath \"\"" Nov 28 13:21:35 crc kubenswrapper[4631]: I1128 13:21:35.140029 4631 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/76a05ca6-a98c-4313-855c-fd6ffaceee4f-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 28 13:21:35 crc kubenswrapper[4631]: I1128 13:21:35.140040 4631 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/76a05ca6-a98c-4313-855c-fd6ffaceee4f-config-volume\") on node \"crc\" DevicePath \"\"" Nov 28 13:21:35 crc kubenswrapper[4631]: I1128 13:21:35.180832 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8tbnv" Nov 28 13:21:35 crc kubenswrapper[4631]: I1128 13:21:35.234667 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-p5v2z" Nov 28 13:21:35 crc kubenswrapper[4631]: I1128 13:21:35.272953 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-bm9gn"] Nov 28 13:21:35 crc kubenswrapper[4631]: E1128 13:21:35.273447 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="76a05ca6-a98c-4313-855c-fd6ffaceee4f" containerName="collect-profiles" Nov 28 13:21:35 crc kubenswrapper[4631]: I1128 13:21:35.273462 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="76a05ca6-a98c-4313-855c-fd6ffaceee4f" containerName="collect-profiles" Nov 28 13:21:35 crc kubenswrapper[4631]: I1128 13:21:35.273573 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="76a05ca6-a98c-4313-855c-fd6ffaceee4f" containerName="collect-profiles" Nov 28 13:21:35 crc kubenswrapper[4631]: I1128 13:21:35.274260 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bm9gn" Nov 28 13:21:35 crc kubenswrapper[4631]: I1128 13:21:35.283886 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 13:21:35 crc kubenswrapper[4631]: I1128 13:21:35.293835 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-bm9gn"] Nov 28 13:21:35 crc kubenswrapper[4631]: I1128 13:21:35.358615 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dk86v\" (UniqueName: \"kubernetes.io/projected/68b72f7f-4534-480d-983e-04ea986e3b41-kube-api-access-dk86v\") pod \"redhat-marketplace-bm9gn\" (UID: \"68b72f7f-4534-480d-983e-04ea986e3b41\") " pod="openshift-marketplace/redhat-marketplace-bm9gn" Nov 28 13:21:35 crc kubenswrapper[4631]: I1128 13:21:35.358673 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/68b72f7f-4534-480d-983e-04ea986e3b41-utilities\") pod \"redhat-marketplace-bm9gn\" (UID: \"68b72f7f-4534-480d-983e-04ea986e3b41\") " pod="openshift-marketplace/redhat-marketplace-bm9gn" Nov 28 13:21:35 crc kubenswrapper[4631]: I1128 13:21:35.358724 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/68b72f7f-4534-480d-983e-04ea986e3b41-catalog-content\") pod \"redhat-marketplace-bm9gn\" (UID: \"68b72f7f-4534-480d-983e-04ea986e3b41\") " pod="openshift-marketplace/redhat-marketplace-bm9gn" Nov 28 13:21:35 crc kubenswrapper[4631]: I1128 13:21:35.459947 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/68b72f7f-4534-480d-983e-04ea986e3b41-catalog-content\") pod \"redhat-marketplace-bm9gn\" (UID: \"68b72f7f-4534-480d-983e-04ea986e3b41\") " pod="openshift-marketplace/redhat-marketplace-bm9gn" Nov 28 13:21:35 crc kubenswrapper[4631]: I1128 13:21:35.460023 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dk86v\" (UniqueName: \"kubernetes.io/projected/68b72f7f-4534-480d-983e-04ea986e3b41-kube-api-access-dk86v\") pod \"redhat-marketplace-bm9gn\" (UID: \"68b72f7f-4534-480d-983e-04ea986e3b41\") " pod="openshift-marketplace/redhat-marketplace-bm9gn" Nov 28 13:21:35 crc kubenswrapper[4631]: I1128 13:21:35.460077 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/68b72f7f-4534-480d-983e-04ea986e3b41-utilities\") pod \"redhat-marketplace-bm9gn\" (UID: \"68b72f7f-4534-480d-983e-04ea986e3b41\") " pod="openshift-marketplace/redhat-marketplace-bm9gn" Nov 28 13:21:35 crc kubenswrapper[4631]: I1128 13:21:35.460644 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/68b72f7f-4534-480d-983e-04ea986e3b41-utilities\") pod \"redhat-marketplace-bm9gn\" (UID: \"68b72f7f-4534-480d-983e-04ea986e3b41\") " pod="openshift-marketplace/redhat-marketplace-bm9gn" Nov 28 13:21:35 crc kubenswrapper[4631]: I1128 13:21:35.470103 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/68b72f7f-4534-480d-983e-04ea986e3b41-catalog-content\") pod \"redhat-marketplace-bm9gn\" (UID: \"68b72f7f-4534-480d-983e-04ea986e3b41\") " pod="openshift-marketplace/redhat-marketplace-bm9gn" Nov 28 13:21:35 crc kubenswrapper[4631]: I1128 13:21:35.518440 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dk86v\" (UniqueName: \"kubernetes.io/projected/68b72f7f-4534-480d-983e-04ea986e3b41-kube-api-access-dk86v\") pod \"redhat-marketplace-bm9gn\" (UID: \"68b72f7f-4534-480d-983e-04ea986e3b41\") " pod="openshift-marketplace/redhat-marketplace-bm9gn" Nov 28 13:21:35 crc kubenswrapper[4631]: I1128 13:21:35.530943 4631 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Nov 28 13:21:35 crc kubenswrapper[4631]: I1128 13:21:35.607554 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Nov 28 13:21:35 crc kubenswrapper[4631]: I1128 13:21:35.627325 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bm9gn" Nov 28 13:21:35 crc kubenswrapper[4631]: W1128 13:21:35.636687 4631 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5fe485a1_e14f_4c09_b5b9_f252bc42b7e8.slice/crio-d143278bfb2eaff4844feaf85ce3b54ae89bdd1f7c193bba5b6750f4a465512b WatchSource:0}: Error finding container d143278bfb2eaff4844feaf85ce3b54ae89bdd1f7c193bba5b6750f4a465512b: Status 404 returned error can't find the container with id d143278bfb2eaff4844feaf85ce3b54ae89bdd1f7c193bba5b6750f4a465512b Nov 28 13:21:35 crc kubenswrapper[4631]: I1128 13:21:35.687422 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-p5v2z"] Nov 28 13:21:35 crc kubenswrapper[4631]: W1128 13:21:35.700525 4631 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod67afe4ac_28a7_42a2_8155_2c4bf496b834.slice/crio-1db98735051371eafbcd998d22e47e1fcd13a954aad0552bb42a56fc5c2334f7 WatchSource:0}: Error finding container 1db98735051371eafbcd998d22e47e1fcd13a954aad0552bb42a56fc5c2334f7: Status 404 returned error can't find the container with id 1db98735051371eafbcd998d22e47e1fcd13a954aad0552bb42a56fc5c2334f7 Nov 28 13:21:35 crc kubenswrapper[4631]: I1128 13:21:35.876395 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-c7dxv"] Nov 28 13:21:35 crc kubenswrapper[4631]: I1128 13:21:35.878996 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-c7dxv" Nov 28 13:21:35 crc kubenswrapper[4631]: I1128 13:21:35.882176 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Nov 28 13:21:35 crc kubenswrapper[4631]: I1128 13:21:35.902528 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-c7dxv"] Nov 28 13:21:35 crc kubenswrapper[4631]: I1128 13:21:35.987065 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-bm9gn"] Nov 28 13:21:35 crc kubenswrapper[4631]: I1128 13:21:35.996717 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5xxgd\" (UniqueName: \"kubernetes.io/projected/45e56fbb-cf6f-4676-a78d-37c59b154fe7-kube-api-access-5xxgd\") pod \"redhat-operators-c7dxv\" (UID: \"45e56fbb-cf6f-4676-a78d-37c59b154fe7\") " pod="openshift-marketplace/redhat-operators-c7dxv" Nov 28 13:21:35 crc kubenswrapper[4631]: I1128 13:21:35.996767 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/45e56fbb-cf6f-4676-a78d-37c59b154fe7-utilities\") pod \"redhat-operators-c7dxv\" (UID: \"45e56fbb-cf6f-4676-a78d-37c59b154fe7\") " pod="openshift-marketplace/redhat-operators-c7dxv" Nov 28 13:21:35 crc kubenswrapper[4631]: I1128 13:21:35.996814 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/45e56fbb-cf6f-4676-a78d-37c59b154fe7-catalog-content\") pod \"redhat-operators-c7dxv\" (UID: \"45e56fbb-cf6f-4676-a78d-37c59b154fe7\") " pod="openshift-marketplace/redhat-operators-c7dxv" Nov 28 13:21:36 crc kubenswrapper[4631]: W1128 13:21:36.009109 4631 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod68b72f7f_4534_480d_983e_04ea986e3b41.slice/crio-783903b653dc550d97d798d90daa227b069f82b54c7bf0fdcde2cd65494cb53c WatchSource:0}: Error finding container 783903b653dc550d97d798d90daa227b069f82b54c7bf0fdcde2cd65494cb53c: Status 404 returned error can't find the container with id 783903b653dc550d97d798d90daa227b069f82b54c7bf0fdcde2cd65494cb53c Nov 28 13:21:36 crc kubenswrapper[4631]: I1128 13:21:36.084130 4631 patch_prober.go:28] interesting pod/router-default-5444994796-jcnpq container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 28 13:21:36 crc kubenswrapper[4631]: [+]has-synced ok Nov 28 13:21:36 crc kubenswrapper[4631]: [+]process-running ok Nov 28 13:21:36 crc kubenswrapper[4631]: healthz check failed Nov 28 13:21:36 crc kubenswrapper[4631]: I1128 13:21:36.084414 4631 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-jcnpq" podUID="22010e1d-3ec1-4564-a362-bc963f691f1f" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 28 13:21:36 crc kubenswrapper[4631]: I1128 13:21:36.098227 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5xxgd\" (UniqueName: \"kubernetes.io/projected/45e56fbb-cf6f-4676-a78d-37c59b154fe7-kube-api-access-5xxgd\") pod \"redhat-operators-c7dxv\" (UID: \"45e56fbb-cf6f-4676-a78d-37c59b154fe7\") " pod="openshift-marketplace/redhat-operators-c7dxv" Nov 28 13:21:36 crc kubenswrapper[4631]: I1128 13:21:36.098273 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/45e56fbb-cf6f-4676-a78d-37c59b154fe7-utilities\") pod \"redhat-operators-c7dxv\" (UID: \"45e56fbb-cf6f-4676-a78d-37c59b154fe7\") " pod="openshift-marketplace/redhat-operators-c7dxv" Nov 28 13:21:36 crc kubenswrapper[4631]: I1128 13:21:36.098333 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/45e56fbb-cf6f-4676-a78d-37c59b154fe7-catalog-content\") pod \"redhat-operators-c7dxv\" (UID: \"45e56fbb-cf6f-4676-a78d-37c59b154fe7\") " pod="openshift-marketplace/redhat-operators-c7dxv" Nov 28 13:21:36 crc kubenswrapper[4631]: I1128 13:21:36.098714 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/45e56fbb-cf6f-4676-a78d-37c59b154fe7-catalog-content\") pod \"redhat-operators-c7dxv\" (UID: \"45e56fbb-cf6f-4676-a78d-37c59b154fe7\") " pod="openshift-marketplace/redhat-operators-c7dxv" Nov 28 13:21:36 crc kubenswrapper[4631]: I1128 13:21:36.099105 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/45e56fbb-cf6f-4676-a78d-37c59b154fe7-utilities\") pod \"redhat-operators-c7dxv\" (UID: \"45e56fbb-cf6f-4676-a78d-37c59b154fe7\") " pod="openshift-marketplace/redhat-operators-c7dxv" Nov 28 13:21:36 crc kubenswrapper[4631]: I1128 13:21:36.115788 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5xxgd\" (UniqueName: \"kubernetes.io/projected/45e56fbb-cf6f-4676-a78d-37c59b154fe7-kube-api-access-5xxgd\") pod \"redhat-operators-c7dxv\" (UID: \"45e56fbb-cf6f-4676-a78d-37c59b154fe7\") " pod="openshift-marketplace/redhat-operators-c7dxv" Nov 28 13:21:36 crc kubenswrapper[4631]: I1128 13:21:36.132836 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-5v5xp" event={"ID":"14717577-3fab-482b-bdbe-0e604d23d6d8","Type":"ContainerStarted","Data":"a9523ebf1eca553f15218be6b889124d3e5f5d105019f43c00e66ba5e3270a43"} Nov 28 13:21:36 crc kubenswrapper[4631]: I1128 13:21:36.133799 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"f99b964d-d26f-452f-8ce4-754c09023a6a","Type":"ContainerStarted","Data":"b330e61df4775bfb66affa3b713331d2b82b183ecd8302271d48377a4fbdc53b"} Nov 28 13:21:36 crc kubenswrapper[4631]: I1128 13:21:36.135484 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"db6dbba7512b53af37d0b277840bb4bb7dc4acf0a22c2ba1e3114d901098ed04"} Nov 28 13:21:36 crc kubenswrapper[4631]: I1128 13:21:36.136253 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"d143278bfb2eaff4844feaf85ce3b54ae89bdd1f7c193bba5b6750f4a465512b"} Nov 28 13:21:36 crc kubenswrapper[4631]: I1128 13:21:36.136895 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-p5v2z" event={"ID":"67afe4ac-28a7-42a2-8155-2c4bf496b834","Type":"ContainerStarted","Data":"1db98735051371eafbcd998d22e47e1fcd13a954aad0552bb42a56fc5c2334f7"} Nov 28 13:21:36 crc kubenswrapper[4631]: I1128 13:21:36.138270 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bm9gn" event={"ID":"68b72f7f-4534-480d-983e-04ea986e3b41","Type":"ContainerStarted","Data":"783903b653dc550d97d798d90daa227b069f82b54c7bf0fdcde2cd65494cb53c"} Nov 28 13:21:36 crc kubenswrapper[4631]: I1128 13:21:36.210237 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-c7dxv" Nov 28 13:21:36 crc kubenswrapper[4631]: I1128 13:21:36.263720 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-44v44"] Nov 28 13:21:36 crc kubenswrapper[4631]: I1128 13:21:36.264858 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-44v44" Nov 28 13:21:36 crc kubenswrapper[4631]: I1128 13:21:36.279031 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-44v44"] Nov 28 13:21:36 crc kubenswrapper[4631]: I1128 13:21:36.402881 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5e49b00c-84e0-4d9c-9095-daaeb2817505-utilities\") pod \"redhat-operators-44v44\" (UID: \"5e49b00c-84e0-4d9c-9095-daaeb2817505\") " pod="openshift-marketplace/redhat-operators-44v44" Nov 28 13:21:36 crc kubenswrapper[4631]: I1128 13:21:36.402957 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5e49b00c-84e0-4d9c-9095-daaeb2817505-catalog-content\") pod \"redhat-operators-44v44\" (UID: \"5e49b00c-84e0-4d9c-9095-daaeb2817505\") " pod="openshift-marketplace/redhat-operators-44v44" Nov 28 13:21:36 crc kubenswrapper[4631]: I1128 13:21:36.403143 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c59wr\" (UniqueName: \"kubernetes.io/projected/5e49b00c-84e0-4d9c-9095-daaeb2817505-kube-api-access-c59wr\") pod \"redhat-operators-44v44\" (UID: \"5e49b00c-84e0-4d9c-9095-daaeb2817505\") " pod="openshift-marketplace/redhat-operators-44v44" Nov 28 13:21:36 crc kubenswrapper[4631]: I1128 13:21:36.426768 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-c7dxv"] Nov 28 13:21:36 crc kubenswrapper[4631]: W1128 13:21:36.433018 4631 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod45e56fbb_cf6f_4676_a78d_37c59b154fe7.slice/crio-d94683e58267a3244facde0e33a428fc34e8f3f8791a6bd1962a64379a31eeab WatchSource:0}: Error finding container d94683e58267a3244facde0e33a428fc34e8f3f8791a6bd1962a64379a31eeab: Status 404 returned error can't find the container with id d94683e58267a3244facde0e33a428fc34e8f3f8791a6bd1962a64379a31eeab Nov 28 13:21:36 crc kubenswrapper[4631]: I1128 13:21:36.504022 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c59wr\" (UniqueName: \"kubernetes.io/projected/5e49b00c-84e0-4d9c-9095-daaeb2817505-kube-api-access-c59wr\") pod \"redhat-operators-44v44\" (UID: \"5e49b00c-84e0-4d9c-9095-daaeb2817505\") " pod="openshift-marketplace/redhat-operators-44v44" Nov 28 13:21:36 crc kubenswrapper[4631]: I1128 13:21:36.504344 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5e49b00c-84e0-4d9c-9095-daaeb2817505-utilities\") pod \"redhat-operators-44v44\" (UID: \"5e49b00c-84e0-4d9c-9095-daaeb2817505\") " pod="openshift-marketplace/redhat-operators-44v44" Nov 28 13:21:36 crc kubenswrapper[4631]: I1128 13:21:36.504406 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5e49b00c-84e0-4d9c-9095-daaeb2817505-catalog-content\") pod \"redhat-operators-44v44\" (UID: \"5e49b00c-84e0-4d9c-9095-daaeb2817505\") " pod="openshift-marketplace/redhat-operators-44v44" Nov 28 13:21:36 crc kubenswrapper[4631]: I1128 13:21:36.504758 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5e49b00c-84e0-4d9c-9095-daaeb2817505-utilities\") pod \"redhat-operators-44v44\" (UID: \"5e49b00c-84e0-4d9c-9095-daaeb2817505\") " pod="openshift-marketplace/redhat-operators-44v44" Nov 28 13:21:36 crc kubenswrapper[4631]: I1128 13:21:36.504789 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5e49b00c-84e0-4d9c-9095-daaeb2817505-catalog-content\") pod \"redhat-operators-44v44\" (UID: \"5e49b00c-84e0-4d9c-9095-daaeb2817505\") " pod="openshift-marketplace/redhat-operators-44v44" Nov 28 13:21:36 crc kubenswrapper[4631]: I1128 13:21:36.520357 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c59wr\" (UniqueName: \"kubernetes.io/projected/5e49b00c-84e0-4d9c-9095-daaeb2817505-kube-api-access-c59wr\") pod \"redhat-operators-44v44\" (UID: \"5e49b00c-84e0-4d9c-9095-daaeb2817505\") " pod="openshift-marketplace/redhat-operators-44v44" Nov 28 13:21:36 crc kubenswrapper[4631]: I1128 13:21:36.579600 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-44v44" Nov 28 13:21:36 crc kubenswrapper[4631]: I1128 13:21:36.754504 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-44v44"] Nov 28 13:21:36 crc kubenswrapper[4631]: W1128 13:21:36.774127 4631 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5e49b00c_84e0_4d9c_9095_daaeb2817505.slice/crio-734d2c4c8e528cfc2f685e98e28714535180b837ee540f3b2c1a8167d41db697 WatchSource:0}: Error finding container 734d2c4c8e528cfc2f685e98e28714535180b837ee540f3b2c1a8167d41db697: Status 404 returned error can't find the container with id 734d2c4c8e528cfc2f685e98e28714535180b837ee540f3b2c1a8167d41db697 Nov 28 13:21:37 crc kubenswrapper[4631]: I1128 13:21:37.084357 4631 patch_prober.go:28] interesting pod/router-default-5444994796-jcnpq container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 28 13:21:37 crc kubenswrapper[4631]: [+]has-synced ok Nov 28 13:21:37 crc kubenswrapper[4631]: [+]process-running ok Nov 28 13:21:37 crc kubenswrapper[4631]: healthz check failed Nov 28 13:21:37 crc kubenswrapper[4631]: I1128 13:21:37.084423 4631 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-jcnpq" podUID="22010e1d-3ec1-4564-a362-bc963f691f1f" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 28 13:21:37 crc kubenswrapper[4631]: I1128 13:21:37.144127 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-44v44" event={"ID":"5e49b00c-84e0-4d9c-9095-daaeb2817505","Type":"ContainerStarted","Data":"734d2c4c8e528cfc2f685e98e28714535180b837ee540f3b2c1a8167d41db697"} Nov 28 13:21:37 crc kubenswrapper[4631]: I1128 13:21:37.145598 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-c7dxv" event={"ID":"45e56fbb-cf6f-4676-a78d-37c59b154fe7","Type":"ContainerStarted","Data":"d94683e58267a3244facde0e33a428fc34e8f3f8791a6bd1962a64379a31eeab"} Nov 28 13:21:37 crc kubenswrapper[4631]: I1128 13:21:37.520702 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 28 13:21:37 crc kubenswrapper[4631]: I1128 13:21:37.534877 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Nov 28 13:21:38 crc kubenswrapper[4631]: I1128 13:21:38.084619 4631 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-jcnpq" Nov 28 13:21:38 crc kubenswrapper[4631]: I1128 13:21:38.089812 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-jcnpq" Nov 28 13:21:38 crc kubenswrapper[4631]: I1128 13:21:38.178746 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-44v44" event={"ID":"5e49b00c-84e0-4d9c-9095-daaeb2817505","Type":"ContainerStarted","Data":"5ab3a50a40a68a1e65aa2b887cc200aee402962b646b1d888af154557c276444"} Nov 28 13:21:38 crc kubenswrapper[4631]: I1128 13:21:38.180170 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"147a69b315f4284b1b486b9ceb9be2b8e7efc3c3d557c50e63a9c77a14bc0de4"} Nov 28 13:21:38 crc kubenswrapper[4631]: I1128 13:21:38.180935 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 13:21:38 crc kubenswrapper[4631]: I1128 13:21:38.187166 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"4ca0250f02694a789e1b0b78d43e05415798ab4d569dad2343c7246562289542"} Nov 28 13:21:38 crc kubenswrapper[4631]: I1128 13:21:38.188886 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-p5v2z" event={"ID":"67afe4ac-28a7-42a2-8155-2c4bf496b834","Type":"ContainerStarted","Data":"cbda83ba8ca2ef82d49ef76e2f1fb00c3ea2e5bec77864f11e07a994b3d84fc8"} Nov 28 13:21:38 crc kubenswrapper[4631]: I1128 13:21:38.190026 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"832293b77cfd14dbea93d109b7c06eafaf6a3d2c726844cbc9fc066a953976e4"} Nov 28 13:21:38 crc kubenswrapper[4631]: I1128 13:21:38.198052 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=1.198026613 podStartE2EDuration="1.198026613s" podCreationTimestamp="2025-11-28 13:21:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 13:21:38.143958076 +0000 UTC m=+54.951261420" watchObservedRunningTime="2025-11-28 13:21:38.198026613 +0000 UTC m=+55.005329957" Nov 28 13:21:38 crc kubenswrapper[4631]: I1128 13:21:38.207438 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bm9gn" event={"ID":"68b72f7f-4534-480d-983e-04ea986e3b41","Type":"ContainerStarted","Data":"e0f5f5dc880cea08555d369e01e4490d9c81b57030e14ec6a4a3c934c8761bda"} Nov 28 13:21:38 crc kubenswrapper[4631]: I1128 13:21:38.227745 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"f99b964d-d26f-452f-8ce4-754c09023a6a","Type":"ContainerStarted","Data":"6425d6ac3e628b4b6372829b45f9aaf836d35f120f7b2d24563a455d8852e564"} Nov 28 13:21:38 crc kubenswrapper[4631]: I1128 13:21:38.236343 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-c7dxv" event={"ID":"45e56fbb-cf6f-4676-a78d-37c59b154fe7","Type":"ContainerStarted","Data":"08297c988d0d0bf16010b6eb5a860143dbfedbb495dc13c4f985424eb3ab5560"} Nov 28 13:21:38 crc kubenswrapper[4631]: I1128 13:21:38.236498 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-5v5xp" Nov 28 13:21:38 crc kubenswrapper[4631]: I1128 13:21:38.769651 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-5v5xp" podStartSLOduration=34.769635027 podStartE2EDuration="34.769635027s" podCreationTimestamp="2025-11-28 13:21:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 13:21:38.285964261 +0000 UTC m=+55.093267595" watchObservedRunningTime="2025-11-28 13:21:38.769635027 +0000 UTC m=+55.576938371" Nov 28 13:21:38 crc kubenswrapper[4631]: I1128 13:21:38.771430 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Nov 28 13:21:38 crc kubenswrapper[4631]: I1128 13:21:38.772024 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 28 13:21:38 crc kubenswrapper[4631]: I1128 13:21:38.777171 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Nov 28 13:21:38 crc kubenswrapper[4631]: I1128 13:21:38.777325 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Nov 28 13:21:38 crc kubenswrapper[4631]: I1128 13:21:38.785288 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Nov 28 13:21:38 crc kubenswrapper[4631]: I1128 13:21:38.838072 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/4ced586e-cae0-4c9a-b4a2-a9d67c5c38ad-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"4ced586e-cae0-4c9a-b4a2-a9d67c5c38ad\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 28 13:21:38 crc kubenswrapper[4631]: I1128 13:21:38.838120 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4ced586e-cae0-4c9a-b4a2-a9d67c5c38ad-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"4ced586e-cae0-4c9a-b4a2-a9d67c5c38ad\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 28 13:21:38 crc kubenswrapper[4631]: I1128 13:21:38.940059 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/4ced586e-cae0-4c9a-b4a2-a9d67c5c38ad-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"4ced586e-cae0-4c9a-b4a2-a9d67c5c38ad\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 28 13:21:38 crc kubenswrapper[4631]: I1128 13:21:38.940108 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4ced586e-cae0-4c9a-b4a2-a9d67c5c38ad-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"4ced586e-cae0-4c9a-b4a2-a9d67c5c38ad\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 28 13:21:38 crc kubenswrapper[4631]: I1128 13:21:38.940206 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/4ced586e-cae0-4c9a-b4a2-a9d67c5c38ad-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"4ced586e-cae0-4c9a-b4a2-a9d67c5c38ad\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 28 13:21:38 crc kubenswrapper[4631]: I1128 13:21:38.977857 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4ced586e-cae0-4c9a-b4a2-a9d67c5c38ad-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"4ced586e-cae0-4c9a-b4a2-a9d67c5c38ad\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 28 13:21:39 crc kubenswrapper[4631]: I1128 13:21:39.086414 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 28 13:21:39 crc kubenswrapper[4631]: I1128 13:21:39.249039 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bm9gn" event={"ID":"68b72f7f-4534-480d-983e-04ea986e3b41","Type":"ContainerDied","Data":"e0f5f5dc880cea08555d369e01e4490d9c81b57030e14ec6a4a3c934c8761bda"} Nov 28 13:21:39 crc kubenswrapper[4631]: I1128 13:21:39.248772 4631 generic.go:334] "Generic (PLEG): container finished" podID="68b72f7f-4534-480d-983e-04ea986e3b41" containerID="e0f5f5dc880cea08555d369e01e4490d9c81b57030e14ec6a4a3c934c8761bda" exitCode=0 Nov 28 13:21:39 crc kubenswrapper[4631]: I1128 13:21:39.260069 4631 generic.go:334] "Generic (PLEG): container finished" podID="45e56fbb-cf6f-4676-a78d-37c59b154fe7" containerID="08297c988d0d0bf16010b6eb5a860143dbfedbb495dc13c4f985424eb3ab5560" exitCode=0 Nov 28 13:21:39 crc kubenswrapper[4631]: I1128 13:21:39.260126 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-c7dxv" event={"ID":"45e56fbb-cf6f-4676-a78d-37c59b154fe7","Type":"ContainerDied","Data":"08297c988d0d0bf16010b6eb5a860143dbfedbb495dc13c4f985424eb3ab5560"} Nov 28 13:21:39 crc kubenswrapper[4631]: I1128 13:21:39.262717 4631 generic.go:334] "Generic (PLEG): container finished" podID="5e49b00c-84e0-4d9c-9095-daaeb2817505" containerID="5ab3a50a40a68a1e65aa2b887cc200aee402962b646b1d888af154557c276444" exitCode=0 Nov 28 13:21:39 crc kubenswrapper[4631]: I1128 13:21:39.262789 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-44v44" event={"ID":"5e49b00c-84e0-4d9c-9095-daaeb2817505","Type":"ContainerDied","Data":"5ab3a50a40a68a1e65aa2b887cc200aee402962b646b1d888af154557c276444"} Nov 28 13:21:39 crc kubenswrapper[4631]: I1128 13:21:39.269932 4631 generic.go:334] "Generic (PLEG): container finished" podID="67afe4ac-28a7-42a2-8155-2c4bf496b834" containerID="cbda83ba8ca2ef82d49ef76e2f1fb00c3ea2e5bec77864f11e07a994b3d84fc8" exitCode=0 Nov 28 13:21:39 crc kubenswrapper[4631]: I1128 13:21:39.273143 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-p5v2z" event={"ID":"67afe4ac-28a7-42a2-8155-2c4bf496b834","Type":"ContainerDied","Data":"cbda83ba8ca2ef82d49ef76e2f1fb00c3ea2e5bec77864f11e07a994b3d84fc8"} Nov 28 13:21:39 crc kubenswrapper[4631]: I1128 13:21:39.398320 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/revision-pruner-9-crc" podStartSLOduration=5.398289807 podStartE2EDuration="5.398289807s" podCreationTimestamp="2025-11-28 13:21:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 13:21:39.39666671 +0000 UTC m=+56.203970054" watchObservedRunningTime="2025-11-28 13:21:39.398289807 +0000 UTC m=+56.205593141" Nov 28 13:21:39 crc kubenswrapper[4631]: I1128 13:21:39.543094 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Nov 28 13:21:39 crc kubenswrapper[4631]: I1128 13:21:39.904022 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-c2rks" Nov 28 13:21:40 crc kubenswrapper[4631]: I1128 13:21:40.275321 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"4ced586e-cae0-4c9a-b4a2-a9d67c5c38ad","Type":"ContainerStarted","Data":"177a3bdd3590b2aa14b351969d6c083888e5c1c8f3cf424d1dc5fa214c374270"} Nov 28 13:21:40 crc kubenswrapper[4631]: I1128 13:21:40.277159 4631 generic.go:334] "Generic (PLEG): container finished" podID="f99b964d-d26f-452f-8ce4-754c09023a6a" containerID="6425d6ac3e628b4b6372829b45f9aaf836d35f120f7b2d24563a455d8852e564" exitCode=0 Nov 28 13:21:40 crc kubenswrapper[4631]: I1128 13:21:40.277858 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"f99b964d-d26f-452f-8ce4-754c09023a6a","Type":"ContainerDied","Data":"6425d6ac3e628b4b6372829b45f9aaf836d35f120f7b2d24563a455d8852e564"} Nov 28 13:21:40 crc kubenswrapper[4631]: I1128 13:21:40.326641 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-cg4zv" Nov 28 13:21:41 crc kubenswrapper[4631]: I1128 13:21:41.317763 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"4ced586e-cae0-4c9a-b4a2-a9d67c5c38ad","Type":"ContainerStarted","Data":"651e724a8d7f1e73ec5482bf464b31a17eda32e7f3e6c4ab7059b2cfca99ad16"} Nov 28 13:21:41 crc kubenswrapper[4631]: I1128 13:21:41.435592 4631 patch_prober.go:28] interesting pod/console-f9d7485db-659jg container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.15:8443/health\": dial tcp 10.217.0.15:8443: connect: connection refused" start-of-body= Nov 28 13:21:41 crc kubenswrapper[4631]: I1128 13:21:41.435647 4631 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-659jg" podUID="8e852fdc-5b82-4b51-b0e5-804a1b3be0ec" containerName="console" probeResult="failure" output="Get \"https://10.217.0.15:8443/health\": dial tcp 10.217.0.15:8443: connect: connection refused" Nov 28 13:21:41 crc kubenswrapper[4631]: I1128 13:21:41.656148 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 28 13:21:41 crc kubenswrapper[4631]: I1128 13:21:41.801982 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-2bc4b" Nov 28 13:21:41 crc kubenswrapper[4631]: I1128 13:21:41.834649 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f99b964d-d26f-452f-8ce4-754c09023a6a-kube-api-access\") pod \"f99b964d-d26f-452f-8ce4-754c09023a6a\" (UID: \"f99b964d-d26f-452f-8ce4-754c09023a6a\") " Nov 28 13:21:41 crc kubenswrapper[4631]: I1128 13:21:41.834691 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/f99b964d-d26f-452f-8ce4-754c09023a6a-kubelet-dir\") pod \"f99b964d-d26f-452f-8ce4-754c09023a6a\" (UID: \"f99b964d-d26f-452f-8ce4-754c09023a6a\") " Nov 28 13:21:41 crc kubenswrapper[4631]: I1128 13:21:41.835007 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f99b964d-d26f-452f-8ce4-754c09023a6a-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "f99b964d-d26f-452f-8ce4-754c09023a6a" (UID: "f99b964d-d26f-452f-8ce4-754c09023a6a"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 13:21:41 crc kubenswrapper[4631]: I1128 13:21:41.860596 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f99b964d-d26f-452f-8ce4-754c09023a6a-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "f99b964d-d26f-452f-8ce4-754c09023a6a" (UID: "f99b964d-d26f-452f-8ce4-754c09023a6a"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:21:41 crc kubenswrapper[4631]: I1128 13:21:41.936077 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f99b964d-d26f-452f-8ce4-754c09023a6a-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 28 13:21:41 crc kubenswrapper[4631]: I1128 13:21:41.937065 4631 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/f99b964d-d26f-452f-8ce4-754c09023a6a-kubelet-dir\") on node \"crc\" DevicePath \"\"" Nov 28 13:21:42 crc kubenswrapper[4631]: E1128 13:21:42.022708 4631 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="3b12a8f186ea08cfb45b04b2d0d67de069fb809dcbf835a4dd4e008e4e24a4ae" cmd=["/bin/bash","-c","test -f /ready/ready"] Nov 28 13:21:42 crc kubenswrapper[4631]: E1128 13:21:42.026518 4631 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="3b12a8f186ea08cfb45b04b2d0d67de069fb809dcbf835a4dd4e008e4e24a4ae" cmd=["/bin/bash","-c","test -f /ready/ready"] Nov 28 13:21:42 crc kubenswrapper[4631]: E1128 13:21:42.028604 4631 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="3b12a8f186ea08cfb45b04b2d0d67de069fb809dcbf835a4dd4e008e4e24a4ae" cmd=["/bin/bash","-c","test -f /ready/ready"] Nov 28 13:21:42 crc kubenswrapper[4631]: E1128 13:21:42.028724 4631 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openshift-multus/cni-sysctl-allowlist-ds-44pwx" podUID="9504d491-6c27-4387-b6da-fa0192cc7a05" containerName="kube-multus-additional-cni-plugins" Nov 28 13:21:42 crc kubenswrapper[4631]: I1128 13:21:42.334439 4631 generic.go:334] "Generic (PLEG): container finished" podID="4ced586e-cae0-4c9a-b4a2-a9d67c5c38ad" containerID="651e724a8d7f1e73ec5482bf464b31a17eda32e7f3e6c4ab7059b2cfca99ad16" exitCode=0 Nov 28 13:21:42 crc kubenswrapper[4631]: I1128 13:21:42.334511 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"4ced586e-cae0-4c9a-b4a2-a9d67c5c38ad","Type":"ContainerDied","Data":"651e724a8d7f1e73ec5482bf464b31a17eda32e7f3e6c4ab7059b2cfca99ad16"} Nov 28 13:21:42 crc kubenswrapper[4631]: I1128 13:21:42.338003 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"f99b964d-d26f-452f-8ce4-754c09023a6a","Type":"ContainerDied","Data":"b330e61df4775bfb66affa3b713331d2b82b183ecd8302271d48377a4fbdc53b"} Nov 28 13:21:42 crc kubenswrapper[4631]: I1128 13:21:42.338064 4631 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b330e61df4775bfb66affa3b713331d2b82b183ecd8302271d48377a4fbdc53b" Nov 28 13:21:42 crc kubenswrapper[4631]: I1128 13:21:42.338073 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 28 13:21:43 crc kubenswrapper[4631]: I1128 13:21:43.810527 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 28 13:21:43 crc kubenswrapper[4631]: I1128 13:21:43.897249 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4ced586e-cae0-4c9a-b4a2-a9d67c5c38ad-kube-api-access\") pod \"4ced586e-cae0-4c9a-b4a2-a9d67c5c38ad\" (UID: \"4ced586e-cae0-4c9a-b4a2-a9d67c5c38ad\") " Nov 28 13:21:43 crc kubenswrapper[4631]: I1128 13:21:43.897417 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/4ced586e-cae0-4c9a-b4a2-a9d67c5c38ad-kubelet-dir\") pod \"4ced586e-cae0-4c9a-b4a2-a9d67c5c38ad\" (UID: \"4ced586e-cae0-4c9a-b4a2-a9d67c5c38ad\") " Nov 28 13:21:43 crc kubenswrapper[4631]: I1128 13:21:43.897465 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4ced586e-cae0-4c9a-b4a2-a9d67c5c38ad-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "4ced586e-cae0-4c9a-b4a2-a9d67c5c38ad" (UID: "4ced586e-cae0-4c9a-b4a2-a9d67c5c38ad"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 13:21:43 crc kubenswrapper[4631]: I1128 13:21:43.897772 4631 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/4ced586e-cae0-4c9a-b4a2-a9d67c5c38ad-kubelet-dir\") on node \"crc\" DevicePath \"\"" Nov 28 13:21:43 crc kubenswrapper[4631]: I1128 13:21:43.903591 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4ced586e-cae0-4c9a-b4a2-a9d67c5c38ad-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "4ced586e-cae0-4c9a-b4a2-a9d67c5c38ad" (UID: "4ced586e-cae0-4c9a-b4a2-a9d67c5c38ad"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:21:43 crc kubenswrapper[4631]: I1128 13:21:43.999304 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4ced586e-cae0-4c9a-b4a2-a9d67c5c38ad-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 28 13:21:44 crc kubenswrapper[4631]: I1128 13:21:44.476615 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"4ced586e-cae0-4c9a-b4a2-a9d67c5c38ad","Type":"ContainerDied","Data":"177a3bdd3590b2aa14b351969d6c083888e5c1c8f3cf424d1dc5fa214c374270"} Nov 28 13:21:44 crc kubenswrapper[4631]: I1128 13:21:44.476657 4631 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="177a3bdd3590b2aa14b351969d6c083888e5c1c8f3cf424d1dc5fa214c374270" Nov 28 13:21:44 crc kubenswrapper[4631]: I1128 13:21:44.476716 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 28 13:21:51 crc kubenswrapper[4631]: I1128 13:21:51.437886 4631 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-659jg" Nov 28 13:21:51 crc kubenswrapper[4631]: I1128 13:21:51.444152 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-659jg" Nov 28 13:21:52 crc kubenswrapper[4631]: E1128 13:21:52.040171 4631 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="3b12a8f186ea08cfb45b04b2d0d67de069fb809dcbf835a4dd4e008e4e24a4ae" cmd=["/bin/bash","-c","test -f /ready/ready"] Nov 28 13:21:52 crc kubenswrapper[4631]: E1128 13:21:52.047112 4631 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="3b12a8f186ea08cfb45b04b2d0d67de069fb809dcbf835a4dd4e008e4e24a4ae" cmd=["/bin/bash","-c","test -f /ready/ready"] Nov 28 13:21:52 crc kubenswrapper[4631]: E1128 13:21:52.051453 4631 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="3b12a8f186ea08cfb45b04b2d0d67de069fb809dcbf835a4dd4e008e4e24a4ae" cmd=["/bin/bash","-c","test -f /ready/ready"] Nov 28 13:21:52 crc kubenswrapper[4631]: E1128 13:21:52.051568 4631 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openshift-multus/cni-sysctl-allowlist-ds-44pwx" podUID="9504d491-6c27-4387-b6da-fa0192cc7a05" containerName="kube-multus-additional-cni-plugins" Nov 28 13:21:54 crc kubenswrapper[4631]: I1128 13:21:54.054805 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-5v5xp" Nov 28 13:22:01 crc kubenswrapper[4631]: I1128 13:22:01.528939 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Nov 28 13:22:01 crc kubenswrapper[4631]: I1128 13:22:01.776060 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-mdws7" Nov 28 13:22:01 crc kubenswrapper[4631]: I1128 13:22:01.837276 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=0.83725164 podStartE2EDuration="837.25164ms" podCreationTimestamp="2025-11-28 13:22:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 13:22:01.833644336 +0000 UTC m=+78.640947700" watchObservedRunningTime="2025-11-28 13:22:01.83725164 +0000 UTC m=+78.644554994" Nov 28 13:22:02 crc kubenswrapper[4631]: E1128 13:22:02.013681 4631 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 3b12a8f186ea08cfb45b04b2d0d67de069fb809dcbf835a4dd4e008e4e24a4ae is running failed: container process not found" containerID="3b12a8f186ea08cfb45b04b2d0d67de069fb809dcbf835a4dd4e008e4e24a4ae" cmd=["/bin/bash","-c","test -f /ready/ready"] Nov 28 13:22:02 crc kubenswrapper[4631]: E1128 13:22:02.014406 4631 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 3b12a8f186ea08cfb45b04b2d0d67de069fb809dcbf835a4dd4e008e4e24a4ae is running failed: container process not found" containerID="3b12a8f186ea08cfb45b04b2d0d67de069fb809dcbf835a4dd4e008e4e24a4ae" cmd=["/bin/bash","-c","test -f /ready/ready"] Nov 28 13:22:02 crc kubenswrapper[4631]: E1128 13:22:02.014792 4631 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 3b12a8f186ea08cfb45b04b2d0d67de069fb809dcbf835a4dd4e008e4e24a4ae is running failed: container process not found" containerID="3b12a8f186ea08cfb45b04b2d0d67de069fb809dcbf835a4dd4e008e4e24a4ae" cmd=["/bin/bash","-c","test -f /ready/ready"] Nov 28 13:22:02 crc kubenswrapper[4631]: E1128 13:22:02.014906 4631 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 3b12a8f186ea08cfb45b04b2d0d67de069fb809dcbf835a4dd4e008e4e24a4ae is running failed: container process not found" probeType="Readiness" pod="openshift-multus/cni-sysctl-allowlist-ds-44pwx" podUID="9504d491-6c27-4387-b6da-fa0192cc7a05" containerName="kube-multus-additional-cni-plugins" Nov 28 13:22:02 crc kubenswrapper[4631]: I1128 13:22:02.727876 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_cni-sysctl-allowlist-ds-44pwx_9504d491-6c27-4387-b6da-fa0192cc7a05/kube-multus-additional-cni-plugins/0.log" Nov 28 13:22:02 crc kubenswrapper[4631]: I1128 13:22:02.727923 4631 generic.go:334] "Generic (PLEG): container finished" podID="9504d491-6c27-4387-b6da-fa0192cc7a05" containerID="3b12a8f186ea08cfb45b04b2d0d67de069fb809dcbf835a4dd4e008e4e24a4ae" exitCode=137 Nov 28 13:22:02 crc kubenswrapper[4631]: I1128 13:22:02.727951 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/cni-sysctl-allowlist-ds-44pwx" event={"ID":"9504d491-6c27-4387-b6da-fa0192cc7a05","Type":"ContainerDied","Data":"3b12a8f186ea08cfb45b04b2d0d67de069fb809dcbf835a4dd4e008e4e24a4ae"} Nov 28 13:22:09 crc kubenswrapper[4631]: I1128 13:22:09.545875 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Nov 28 13:22:12 crc kubenswrapper[4631]: E1128 13:22:12.014589 4631 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 3b12a8f186ea08cfb45b04b2d0d67de069fb809dcbf835a4dd4e008e4e24a4ae is running failed: container process not found" containerID="3b12a8f186ea08cfb45b04b2d0d67de069fb809dcbf835a4dd4e008e4e24a4ae" cmd=["/bin/bash","-c","test -f /ready/ready"] Nov 28 13:22:12 crc kubenswrapper[4631]: E1128 13:22:12.015831 4631 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 3b12a8f186ea08cfb45b04b2d0d67de069fb809dcbf835a4dd4e008e4e24a4ae is running failed: container process not found" containerID="3b12a8f186ea08cfb45b04b2d0d67de069fb809dcbf835a4dd4e008e4e24a4ae" cmd=["/bin/bash","-c","test -f /ready/ready"] Nov 28 13:22:12 crc kubenswrapper[4631]: E1128 13:22:12.016546 4631 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 3b12a8f186ea08cfb45b04b2d0d67de069fb809dcbf835a4dd4e008e4e24a4ae is running failed: container process not found" containerID="3b12a8f186ea08cfb45b04b2d0d67de069fb809dcbf835a4dd4e008e4e24a4ae" cmd=["/bin/bash","-c","test -f /ready/ready"] Nov 28 13:22:12 crc kubenswrapper[4631]: E1128 13:22:12.016719 4631 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 3b12a8f186ea08cfb45b04b2d0d67de069fb809dcbf835a4dd4e008e4e24a4ae is running failed: container process not found" probeType="Readiness" pod="openshift-multus/cni-sysctl-allowlist-ds-44pwx" podUID="9504d491-6c27-4387-b6da-fa0192cc7a05" containerName="kube-multus-additional-cni-plugins" Nov 28 13:22:12 crc kubenswrapper[4631]: I1128 13:22:12.175978 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Nov 28 13:22:12 crc kubenswrapper[4631]: E1128 13:22:12.176357 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ced586e-cae0-4c9a-b4a2-a9d67c5c38ad" containerName="pruner" Nov 28 13:22:12 crc kubenswrapper[4631]: I1128 13:22:12.176379 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ced586e-cae0-4c9a-b4a2-a9d67c5c38ad" containerName="pruner" Nov 28 13:22:12 crc kubenswrapper[4631]: E1128 13:22:12.176409 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f99b964d-d26f-452f-8ce4-754c09023a6a" containerName="pruner" Nov 28 13:22:12 crc kubenswrapper[4631]: I1128 13:22:12.176426 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="f99b964d-d26f-452f-8ce4-754c09023a6a" containerName="pruner" Nov 28 13:22:12 crc kubenswrapper[4631]: I1128 13:22:12.176619 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="4ced586e-cae0-4c9a-b4a2-a9d67c5c38ad" containerName="pruner" Nov 28 13:22:12 crc kubenswrapper[4631]: I1128 13:22:12.176653 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="f99b964d-d26f-452f-8ce4-754c09023a6a" containerName="pruner" Nov 28 13:22:12 crc kubenswrapper[4631]: I1128 13:22:12.177232 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 28 13:22:12 crc kubenswrapper[4631]: I1128 13:22:12.180426 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Nov 28 13:22:12 crc kubenswrapper[4631]: I1128 13:22:12.181545 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Nov 28 13:22:12 crc kubenswrapper[4631]: I1128 13:22:12.197796 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Nov 28 13:22:12 crc kubenswrapper[4631]: I1128 13:22:12.250794 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=3.2507669679999998 podStartE2EDuration="3.250766968s" podCreationTimestamp="2025-11-28 13:22:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 13:22:12.248200281 +0000 UTC m=+89.055503625" watchObservedRunningTime="2025-11-28 13:22:12.250766968 +0000 UTC m=+89.058070312" Nov 28 13:22:12 crc kubenswrapper[4631]: I1128 13:22:12.318471 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/7c934e9f-8e8c-48a4-924a-0d0c8e4e08c7-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"7c934e9f-8e8c-48a4-924a-0d0c8e4e08c7\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 28 13:22:12 crc kubenswrapper[4631]: I1128 13:22:12.318582 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7c934e9f-8e8c-48a4-924a-0d0c8e4e08c7-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"7c934e9f-8e8c-48a4-924a-0d0c8e4e08c7\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 28 13:22:12 crc kubenswrapper[4631]: I1128 13:22:12.420132 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/7c934e9f-8e8c-48a4-924a-0d0c8e4e08c7-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"7c934e9f-8e8c-48a4-924a-0d0c8e4e08c7\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 28 13:22:12 crc kubenswrapper[4631]: I1128 13:22:12.420439 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/7c934e9f-8e8c-48a4-924a-0d0c8e4e08c7-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"7c934e9f-8e8c-48a4-924a-0d0c8e4e08c7\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 28 13:22:12 crc kubenswrapper[4631]: I1128 13:22:12.420717 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7c934e9f-8e8c-48a4-924a-0d0c8e4e08c7-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"7c934e9f-8e8c-48a4-924a-0d0c8e4e08c7\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 28 13:22:12 crc kubenswrapper[4631]: I1128 13:22:12.451133 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7c934e9f-8e8c-48a4-924a-0d0c8e4e08c7-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"7c934e9f-8e8c-48a4-924a-0d0c8e4e08c7\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 28 13:22:12 crc kubenswrapper[4631]: I1128 13:22:12.502090 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 28 13:22:14 crc kubenswrapper[4631]: I1128 13:22:14.652211 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 13:22:16 crc kubenswrapper[4631]: I1128 13:22:16.104583 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_cni-sysctl-allowlist-ds-44pwx_9504d491-6c27-4387-b6da-fa0192cc7a05/kube-multus-additional-cni-plugins/0.log" Nov 28 13:22:16 crc kubenswrapper[4631]: I1128 13:22:16.105020 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-multus/cni-sysctl-allowlist-ds-44pwx" Nov 28 13:22:16 crc kubenswrapper[4631]: I1128 13:22:16.172236 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j9wzh\" (UniqueName: \"kubernetes.io/projected/9504d491-6c27-4387-b6da-fa0192cc7a05-kube-api-access-j9wzh\") pod \"9504d491-6c27-4387-b6da-fa0192cc7a05\" (UID: \"9504d491-6c27-4387-b6da-fa0192cc7a05\") " Nov 28 13:22:16 crc kubenswrapper[4631]: I1128 13:22:16.172356 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/9504d491-6c27-4387-b6da-fa0192cc7a05-tuning-conf-dir\") pod \"9504d491-6c27-4387-b6da-fa0192cc7a05\" (UID: \"9504d491-6c27-4387-b6da-fa0192cc7a05\") " Nov 28 13:22:16 crc kubenswrapper[4631]: I1128 13:22:16.172432 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/9504d491-6c27-4387-b6da-fa0192cc7a05-cni-sysctl-allowlist\") pod \"9504d491-6c27-4387-b6da-fa0192cc7a05\" (UID: \"9504d491-6c27-4387-b6da-fa0192cc7a05\") " Nov 28 13:22:16 crc kubenswrapper[4631]: I1128 13:22:16.172447 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9504d491-6c27-4387-b6da-fa0192cc7a05-tuning-conf-dir" (OuterVolumeSpecName: "tuning-conf-dir") pod "9504d491-6c27-4387-b6da-fa0192cc7a05" (UID: "9504d491-6c27-4387-b6da-fa0192cc7a05"). InnerVolumeSpecName "tuning-conf-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 13:22:16 crc kubenswrapper[4631]: I1128 13:22:16.172501 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ready\" (UniqueName: \"kubernetes.io/empty-dir/9504d491-6c27-4387-b6da-fa0192cc7a05-ready\") pod \"9504d491-6c27-4387-b6da-fa0192cc7a05\" (UID: \"9504d491-6c27-4387-b6da-fa0192cc7a05\") " Nov 28 13:22:16 crc kubenswrapper[4631]: I1128 13:22:16.172790 4631 reconciler_common.go:293] "Volume detached for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/9504d491-6c27-4387-b6da-fa0192cc7a05-tuning-conf-dir\") on node \"crc\" DevicePath \"\"" Nov 28 13:22:16 crc kubenswrapper[4631]: I1128 13:22:16.173326 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9504d491-6c27-4387-b6da-fa0192cc7a05-ready" (OuterVolumeSpecName: "ready") pod "9504d491-6c27-4387-b6da-fa0192cc7a05" (UID: "9504d491-6c27-4387-b6da-fa0192cc7a05"). InnerVolumeSpecName "ready". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 13:22:16 crc kubenswrapper[4631]: I1128 13:22:16.173576 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9504d491-6c27-4387-b6da-fa0192cc7a05-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "9504d491-6c27-4387-b6da-fa0192cc7a05" (UID: "9504d491-6c27-4387-b6da-fa0192cc7a05"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 13:22:16 crc kubenswrapper[4631]: I1128 13:22:16.194005 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9504d491-6c27-4387-b6da-fa0192cc7a05-kube-api-access-j9wzh" (OuterVolumeSpecName: "kube-api-access-j9wzh") pod "9504d491-6c27-4387-b6da-fa0192cc7a05" (UID: "9504d491-6c27-4387-b6da-fa0192cc7a05"). InnerVolumeSpecName "kube-api-access-j9wzh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:22:16 crc kubenswrapper[4631]: I1128 13:22:16.274558 4631 reconciler_common.go:293] "Volume detached for volume \"ready\" (UniqueName: \"kubernetes.io/empty-dir/9504d491-6c27-4387-b6da-fa0192cc7a05-ready\") on node \"crc\" DevicePath \"\"" Nov 28 13:22:16 crc kubenswrapper[4631]: I1128 13:22:16.274602 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j9wzh\" (UniqueName: \"kubernetes.io/projected/9504d491-6c27-4387-b6da-fa0192cc7a05-kube-api-access-j9wzh\") on node \"crc\" DevicePath \"\"" Nov 28 13:22:16 crc kubenswrapper[4631]: I1128 13:22:16.274619 4631 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/9504d491-6c27-4387-b6da-fa0192cc7a05-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Nov 28 13:22:16 crc kubenswrapper[4631]: I1128 13:22:16.823976 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_cni-sysctl-allowlist-ds-44pwx_9504d491-6c27-4387-b6da-fa0192cc7a05/kube-multus-additional-cni-plugins/0.log" Nov 28 13:22:16 crc kubenswrapper[4631]: I1128 13:22:16.824405 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/cni-sysctl-allowlist-ds-44pwx" event={"ID":"9504d491-6c27-4387-b6da-fa0192cc7a05","Type":"ContainerDied","Data":"944aa14a03b8001a1a5f339a895060dcf282c666e08145bda17fd13886b219d5"} Nov 28 13:22:16 crc kubenswrapper[4631]: I1128 13:22:16.824474 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-multus/cni-sysctl-allowlist-ds-44pwx" Nov 28 13:22:16 crc kubenswrapper[4631]: I1128 13:22:16.824546 4631 scope.go:117] "RemoveContainer" containerID="3b12a8f186ea08cfb45b04b2d0d67de069fb809dcbf835a4dd4e008e4e24a4ae" Nov 28 13:22:16 crc kubenswrapper[4631]: I1128 13:22:16.867198 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-multus/cni-sysctl-allowlist-ds-44pwx"] Nov 28 13:22:16 crc kubenswrapper[4631]: I1128 13:22:16.870135 4631 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-multus/cni-sysctl-allowlist-ds-44pwx"] Nov 28 13:22:17 crc kubenswrapper[4631]: I1128 13:22:17.522847 4631 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9504d491-6c27-4387-b6da-fa0192cc7a05" path="/var/lib/kubelet/pods/9504d491-6c27-4387-b6da-fa0192cc7a05/volumes" Nov 28 13:22:17 crc kubenswrapper[4631]: I1128 13:22:17.770054 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Nov 28 13:22:17 crc kubenswrapper[4631]: E1128 13:22:17.770465 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9504d491-6c27-4387-b6da-fa0192cc7a05" containerName="kube-multus-additional-cni-plugins" Nov 28 13:22:17 crc kubenswrapper[4631]: I1128 13:22:17.770483 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="9504d491-6c27-4387-b6da-fa0192cc7a05" containerName="kube-multus-additional-cni-plugins" Nov 28 13:22:17 crc kubenswrapper[4631]: I1128 13:22:17.770668 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="9504d491-6c27-4387-b6da-fa0192cc7a05" containerName="kube-multus-additional-cni-plugins" Nov 28 13:22:17 crc kubenswrapper[4631]: I1128 13:22:17.771147 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Nov 28 13:22:17 crc kubenswrapper[4631]: I1128 13:22:17.781103 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Nov 28 13:22:17 crc kubenswrapper[4631]: I1128 13:22:17.896950 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/3a23494b-5ae8-4ace-8763-b94ffe64c5a2-var-lock\") pod \"installer-9-crc\" (UID: \"3a23494b-5ae8-4ace-8763-b94ffe64c5a2\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 28 13:22:17 crc kubenswrapper[4631]: I1128 13:22:17.896999 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3a23494b-5ae8-4ace-8763-b94ffe64c5a2-kube-api-access\") pod \"installer-9-crc\" (UID: \"3a23494b-5ae8-4ace-8763-b94ffe64c5a2\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 28 13:22:17 crc kubenswrapper[4631]: I1128 13:22:17.897024 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/3a23494b-5ae8-4ace-8763-b94ffe64c5a2-kubelet-dir\") pod \"installer-9-crc\" (UID: \"3a23494b-5ae8-4ace-8763-b94ffe64c5a2\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 28 13:22:17 crc kubenswrapper[4631]: I1128 13:22:17.998726 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/3a23494b-5ae8-4ace-8763-b94ffe64c5a2-var-lock\") pod \"installer-9-crc\" (UID: \"3a23494b-5ae8-4ace-8763-b94ffe64c5a2\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 28 13:22:17 crc kubenswrapper[4631]: I1128 13:22:17.998786 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3a23494b-5ae8-4ace-8763-b94ffe64c5a2-kube-api-access\") pod \"installer-9-crc\" (UID: \"3a23494b-5ae8-4ace-8763-b94ffe64c5a2\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 28 13:22:17 crc kubenswrapper[4631]: I1128 13:22:17.998819 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/3a23494b-5ae8-4ace-8763-b94ffe64c5a2-kubelet-dir\") pod \"installer-9-crc\" (UID: \"3a23494b-5ae8-4ace-8763-b94ffe64c5a2\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 28 13:22:17 crc kubenswrapper[4631]: I1128 13:22:17.998866 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/3a23494b-5ae8-4ace-8763-b94ffe64c5a2-var-lock\") pod \"installer-9-crc\" (UID: \"3a23494b-5ae8-4ace-8763-b94ffe64c5a2\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 28 13:22:17 crc kubenswrapper[4631]: I1128 13:22:17.998917 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/3a23494b-5ae8-4ace-8763-b94ffe64c5a2-kubelet-dir\") pod \"installer-9-crc\" (UID: \"3a23494b-5ae8-4ace-8763-b94ffe64c5a2\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 28 13:22:18 crc kubenswrapper[4631]: I1128 13:22:18.034077 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3a23494b-5ae8-4ace-8763-b94ffe64c5a2-kube-api-access\") pod \"installer-9-crc\" (UID: \"3a23494b-5ae8-4ace-8763-b94ffe64c5a2\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 28 13:22:18 crc kubenswrapper[4631]: I1128 13:22:18.100324 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Nov 28 13:22:18 crc kubenswrapper[4631]: E1128 13:22:18.316525 4631 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Nov 28 13:22:18 crc kubenswrapper[4631]: E1128 13:22:18.316869 4631 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-d52xh,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-tpjm2_openshift-marketplace(2fdd01cb-8e67-4bc4-aa2b-e023882103ce): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 28 13:22:18 crc kubenswrapper[4631]: E1128 13:22:18.318313 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-tpjm2" podUID="2fdd01cb-8e67-4bc4-aa2b-e023882103ce" Nov 28 13:22:22 crc kubenswrapper[4631]: E1128 13:22:22.395198 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-tpjm2" podUID="2fdd01cb-8e67-4bc4-aa2b-e023882103ce" Nov 28 13:22:22 crc kubenswrapper[4631]: E1128 13:22:22.545964 4631 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Nov 28 13:22:22 crc kubenswrapper[4631]: E1128 13:22:22.546693 4631 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-5xxgd,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-c7dxv_openshift-marketplace(45e56fbb-cf6f-4676-a78d-37c59b154fe7): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 28 13:22:22 crc kubenswrapper[4631]: E1128 13:22:22.548202 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-c7dxv" podUID="45e56fbb-cf6f-4676-a78d-37c59b154fe7" Nov 28 13:22:22 crc kubenswrapper[4631]: E1128 13:22:22.576327 4631 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Nov 28 13:22:22 crc kubenswrapper[4631]: E1128 13:22:22.576558 4631 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-c59wr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-44v44_openshift-marketplace(5e49b00c-84e0-4d9c-9095-daaeb2817505): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 28 13:22:22 crc kubenswrapper[4631]: E1128 13:22:22.577835 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-44v44" podUID="5e49b00c-84e0-4d9c-9095-daaeb2817505" Nov 28 13:22:26 crc kubenswrapper[4631]: E1128 13:22:26.374389 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-44v44" podUID="5e49b00c-84e0-4d9c-9095-daaeb2817505" Nov 28 13:22:26 crc kubenswrapper[4631]: E1128 13:22:26.376436 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-c7dxv" podUID="45e56fbb-cf6f-4676-a78d-37c59b154fe7" Nov 28 13:22:26 crc kubenswrapper[4631]: E1128 13:22:26.456942 4631 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Nov 28 13:22:26 crc kubenswrapper[4631]: E1128 13:22:26.457297 4631 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-njf98,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-4trst_openshift-marketplace(f6675fed-3a90-4779-9b95-1acba5248fd2): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 28 13:22:26 crc kubenswrapper[4631]: E1128 13:22:26.458773 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-4trst" podUID="f6675fed-3a90-4779-9b95-1acba5248fd2" Nov 28 13:22:26 crc kubenswrapper[4631]: E1128 13:22:26.506448 4631 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Nov 28 13:22:26 crc kubenswrapper[4631]: E1128 13:22:26.506689 4631 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-tfshw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-zkczn_openshift-marketplace(10dbdccc-7009-407e-846d-8a2976a58fad): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 28 13:22:26 crc kubenswrapper[4631]: E1128 13:22:26.507855 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-zkczn" podUID="10dbdccc-7009-407e-846d-8a2976a58fad" Nov 28 13:22:27 crc kubenswrapper[4631]: E1128 13:22:27.695174 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-4trst" podUID="f6675fed-3a90-4779-9b95-1acba5248fd2" Nov 28 13:22:27 crc kubenswrapper[4631]: E1128 13:22:27.696052 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-zkczn" podUID="10dbdccc-7009-407e-846d-8a2976a58fad" Nov 28 13:22:27 crc kubenswrapper[4631]: E1128 13:22:27.835592 4631 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Nov 28 13:22:27 crc kubenswrapper[4631]: E1128 13:22:27.836363 4631 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-dk86v,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-bm9gn_openshift-marketplace(68b72f7f-4534-480d-983e-04ea986e3b41): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 28 13:22:27 crc kubenswrapper[4631]: E1128 13:22:27.840592 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-bm9gn" podUID="68b72f7f-4534-480d-983e-04ea986e3b41" Nov 28 13:22:27 crc kubenswrapper[4631]: E1128 13:22:27.898120 4631 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Nov 28 13:22:27 crc kubenswrapper[4631]: E1128 13:22:27.898345 4631 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-qpqxs,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-6nrc5_openshift-marketplace(72b60e37-5508-451f-8602-9dcebd55324e): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 28 13:22:27 crc kubenswrapper[4631]: E1128 13:22:27.899792 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-6nrc5" podUID="72b60e37-5508-451f-8602-9dcebd55324e" Nov 28 13:22:27 crc kubenswrapper[4631]: E1128 13:22:27.902025 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-bm9gn" podUID="68b72f7f-4534-480d-983e-04ea986e3b41" Nov 28 13:22:27 crc kubenswrapper[4631]: E1128 13:22:27.977983 4631 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Nov 28 13:22:27 crc kubenswrapper[4631]: E1128 13:22:27.978164 4631 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-9lkb4,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-p5v2z_openshift-marketplace(67afe4ac-28a7-42a2-8155-2c4bf496b834): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 28 13:22:27 crc kubenswrapper[4631]: E1128 13:22:27.979650 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-p5v2z" podUID="67afe4ac-28a7-42a2-8155-2c4bf496b834" Nov 28 13:22:28 crc kubenswrapper[4631]: I1128 13:22:28.149846 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Nov 28 13:22:28 crc kubenswrapper[4631]: I1128 13:22:28.251103 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Nov 28 13:22:28 crc kubenswrapper[4631]: I1128 13:22:28.907129 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"7c934e9f-8e8c-48a4-924a-0d0c8e4e08c7","Type":"ContainerStarted","Data":"71dc7fca12d8f93cc8c17e1f138eb75b5f8d234569de9a06836981d764da44da"} Nov 28 13:22:28 crc kubenswrapper[4631]: I1128 13:22:28.907458 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"7c934e9f-8e8c-48a4-924a-0d0c8e4e08c7","Type":"ContainerStarted","Data":"f49fcff6abded76ae83078acf6b5911ac587bed3298e89a7f52b5e2a3638db76"} Nov 28 13:22:28 crc kubenswrapper[4631]: I1128 13:22:28.909299 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"3a23494b-5ae8-4ace-8763-b94ffe64c5a2","Type":"ContainerStarted","Data":"93f973696558d2b35823f54d37651dffd0dca0b0a270471899448eeae23ca4d2"} Nov 28 13:22:28 crc kubenswrapper[4631]: I1128 13:22:28.909517 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"3a23494b-5ae8-4ace-8763-b94ffe64c5a2","Type":"ContainerStarted","Data":"6c5ff36ffbac32af2835a89b780efe3f57ac371f22108b849656f4e985416966"} Nov 28 13:22:28 crc kubenswrapper[4631]: E1128 13:22:28.911452 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-6nrc5" podUID="72b60e37-5508-451f-8602-9dcebd55324e" Nov 28 13:22:28 crc kubenswrapper[4631]: E1128 13:22:28.911467 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-p5v2z" podUID="67afe4ac-28a7-42a2-8155-2c4bf496b834" Nov 28 13:22:28 crc kubenswrapper[4631]: I1128 13:22:28.926497 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-9-crc" podStartSLOduration=16.926481891999998 podStartE2EDuration="16.926481892s" podCreationTimestamp="2025-11-28 13:22:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 13:22:28.925163988 +0000 UTC m=+105.732467342" watchObservedRunningTime="2025-11-28 13:22:28.926481892 +0000 UTC m=+105.733785236" Nov 28 13:22:29 crc kubenswrapper[4631]: I1128 13:22:29.002681 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/installer-9-crc" podStartSLOduration=12.002656142 podStartE2EDuration="12.002656142s" podCreationTimestamp="2025-11-28 13:22:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 13:22:28.979580925 +0000 UTC m=+105.786884289" watchObservedRunningTime="2025-11-28 13:22:29.002656142 +0000 UTC m=+105.809959486" Nov 28 13:22:30 crc kubenswrapper[4631]: I1128 13:22:30.921810 4631 generic.go:334] "Generic (PLEG): container finished" podID="7c934e9f-8e8c-48a4-924a-0d0c8e4e08c7" containerID="71dc7fca12d8f93cc8c17e1f138eb75b5f8d234569de9a06836981d764da44da" exitCode=0 Nov 28 13:22:30 crc kubenswrapper[4631]: I1128 13:22:30.922174 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"7c934e9f-8e8c-48a4-924a-0d0c8e4e08c7","Type":"ContainerDied","Data":"71dc7fca12d8f93cc8c17e1f138eb75b5f8d234569de9a06836981d764da44da"} Nov 28 13:22:32 crc kubenswrapper[4631]: I1128 13:22:32.153500 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 28 13:22:32 crc kubenswrapper[4631]: I1128 13:22:32.230447 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7c934e9f-8e8c-48a4-924a-0d0c8e4e08c7-kube-api-access\") pod \"7c934e9f-8e8c-48a4-924a-0d0c8e4e08c7\" (UID: \"7c934e9f-8e8c-48a4-924a-0d0c8e4e08c7\") " Nov 28 13:22:32 crc kubenswrapper[4631]: I1128 13:22:32.230580 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/7c934e9f-8e8c-48a4-924a-0d0c8e4e08c7-kubelet-dir\") pod \"7c934e9f-8e8c-48a4-924a-0d0c8e4e08c7\" (UID: \"7c934e9f-8e8c-48a4-924a-0d0c8e4e08c7\") " Nov 28 13:22:32 crc kubenswrapper[4631]: I1128 13:22:32.230746 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7c934e9f-8e8c-48a4-924a-0d0c8e4e08c7-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "7c934e9f-8e8c-48a4-924a-0d0c8e4e08c7" (UID: "7c934e9f-8e8c-48a4-924a-0d0c8e4e08c7"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 13:22:32 crc kubenswrapper[4631]: I1128 13:22:32.241900 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7c934e9f-8e8c-48a4-924a-0d0c8e4e08c7-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "7c934e9f-8e8c-48a4-924a-0d0c8e4e08c7" (UID: "7c934e9f-8e8c-48a4-924a-0d0c8e4e08c7"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:22:32 crc kubenswrapper[4631]: I1128 13:22:32.331979 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7c934e9f-8e8c-48a4-924a-0d0c8e4e08c7-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 28 13:22:32 crc kubenswrapper[4631]: I1128 13:22:32.332016 4631 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/7c934e9f-8e8c-48a4-924a-0d0c8e4e08c7-kubelet-dir\") on node \"crc\" DevicePath \"\"" Nov 28 13:22:32 crc kubenswrapper[4631]: I1128 13:22:32.936435 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"7c934e9f-8e8c-48a4-924a-0d0c8e4e08c7","Type":"ContainerDied","Data":"f49fcff6abded76ae83078acf6b5911ac587bed3298e89a7f52b5e2a3638db76"} Nov 28 13:22:32 crc kubenswrapper[4631]: I1128 13:22:32.936478 4631 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f49fcff6abded76ae83078acf6b5911ac587bed3298e89a7f52b5e2a3638db76" Nov 28 13:22:32 crc kubenswrapper[4631]: I1128 13:22:32.936512 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 28 13:22:38 crc kubenswrapper[4631]: I1128 13:22:38.967156 4631 generic.go:334] "Generic (PLEG): container finished" podID="2fdd01cb-8e67-4bc4-aa2b-e023882103ce" containerID="6768ec2f4e644e060c069b1397cccee1b4463bc5cdacc6f0f3e12822304061c7" exitCode=0 Nov 28 13:22:38 crc kubenswrapper[4631]: I1128 13:22:38.967439 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tpjm2" event={"ID":"2fdd01cb-8e67-4bc4-aa2b-e023882103ce","Type":"ContainerDied","Data":"6768ec2f4e644e060c069b1397cccee1b4463bc5cdacc6f0f3e12822304061c7"} Nov 28 13:22:39 crc kubenswrapper[4631]: I1128 13:22:39.054886 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-jhwlx"] Nov 28 13:22:39 crc kubenswrapper[4631]: E1128 13:22:39.055168 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7c934e9f-8e8c-48a4-924a-0d0c8e4e08c7" containerName="pruner" Nov 28 13:22:39 crc kubenswrapper[4631]: I1128 13:22:39.055189 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="7c934e9f-8e8c-48a4-924a-0d0c8e4e08c7" containerName="pruner" Nov 28 13:22:39 crc kubenswrapper[4631]: I1128 13:22:39.055331 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="7c934e9f-8e8c-48a4-924a-0d0c8e4e08c7" containerName="pruner" Nov 28 13:22:39 crc kubenswrapper[4631]: I1128 13:22:39.055815 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-jhwlx" Nov 28 13:22:39 crc kubenswrapper[4631]: I1128 13:22:39.078935 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-jhwlx"] Nov 28 13:22:39 crc kubenswrapper[4631]: I1128 13:22:39.129432 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/c6cb07da-9136-445f-b855-19e16f66b46f-installation-pull-secrets\") pod \"image-registry-66df7c8f76-jhwlx\" (UID: \"c6cb07da-9136-445f-b855-19e16f66b46f\") " pod="openshift-image-registry/image-registry-66df7c8f76-jhwlx" Nov 28 13:22:39 crc kubenswrapper[4631]: I1128 13:22:39.129833 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c6cb07da-9136-445f-b855-19e16f66b46f-trusted-ca\") pod \"image-registry-66df7c8f76-jhwlx\" (UID: \"c6cb07da-9136-445f-b855-19e16f66b46f\") " pod="openshift-image-registry/image-registry-66df7c8f76-jhwlx" Nov 28 13:22:39 crc kubenswrapper[4631]: I1128 13:22:39.129860 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-jhwlx\" (UID: \"c6cb07da-9136-445f-b855-19e16f66b46f\") " pod="openshift-image-registry/image-registry-66df7c8f76-jhwlx" Nov 28 13:22:39 crc kubenswrapper[4631]: I1128 13:22:39.129883 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bsd5d\" (UniqueName: \"kubernetes.io/projected/c6cb07da-9136-445f-b855-19e16f66b46f-kube-api-access-bsd5d\") pod \"image-registry-66df7c8f76-jhwlx\" (UID: \"c6cb07da-9136-445f-b855-19e16f66b46f\") " pod="openshift-image-registry/image-registry-66df7c8f76-jhwlx" Nov 28 13:22:39 crc kubenswrapper[4631]: I1128 13:22:39.129959 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/c6cb07da-9136-445f-b855-19e16f66b46f-ca-trust-extracted\") pod \"image-registry-66df7c8f76-jhwlx\" (UID: \"c6cb07da-9136-445f-b855-19e16f66b46f\") " pod="openshift-image-registry/image-registry-66df7c8f76-jhwlx" Nov 28 13:22:39 crc kubenswrapper[4631]: I1128 13:22:39.130018 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/c6cb07da-9136-445f-b855-19e16f66b46f-registry-certificates\") pod \"image-registry-66df7c8f76-jhwlx\" (UID: \"c6cb07da-9136-445f-b855-19e16f66b46f\") " pod="openshift-image-registry/image-registry-66df7c8f76-jhwlx" Nov 28 13:22:39 crc kubenswrapper[4631]: I1128 13:22:39.130112 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/c6cb07da-9136-445f-b855-19e16f66b46f-bound-sa-token\") pod \"image-registry-66df7c8f76-jhwlx\" (UID: \"c6cb07da-9136-445f-b855-19e16f66b46f\") " pod="openshift-image-registry/image-registry-66df7c8f76-jhwlx" Nov 28 13:22:39 crc kubenswrapper[4631]: I1128 13:22:39.130139 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/c6cb07da-9136-445f-b855-19e16f66b46f-registry-tls\") pod \"image-registry-66df7c8f76-jhwlx\" (UID: \"c6cb07da-9136-445f-b855-19e16f66b46f\") " pod="openshift-image-registry/image-registry-66df7c8f76-jhwlx" Nov 28 13:22:39 crc kubenswrapper[4631]: I1128 13:22:39.157897 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-jhwlx\" (UID: \"c6cb07da-9136-445f-b855-19e16f66b46f\") " pod="openshift-image-registry/image-registry-66df7c8f76-jhwlx" Nov 28 13:22:39 crc kubenswrapper[4631]: I1128 13:22:39.231663 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/c6cb07da-9136-445f-b855-19e16f66b46f-installation-pull-secrets\") pod \"image-registry-66df7c8f76-jhwlx\" (UID: \"c6cb07da-9136-445f-b855-19e16f66b46f\") " pod="openshift-image-registry/image-registry-66df7c8f76-jhwlx" Nov 28 13:22:39 crc kubenswrapper[4631]: I1128 13:22:39.231715 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c6cb07da-9136-445f-b855-19e16f66b46f-trusted-ca\") pod \"image-registry-66df7c8f76-jhwlx\" (UID: \"c6cb07da-9136-445f-b855-19e16f66b46f\") " pod="openshift-image-registry/image-registry-66df7c8f76-jhwlx" Nov 28 13:22:39 crc kubenswrapper[4631]: I1128 13:22:39.231736 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bsd5d\" (UniqueName: \"kubernetes.io/projected/c6cb07da-9136-445f-b855-19e16f66b46f-kube-api-access-bsd5d\") pod \"image-registry-66df7c8f76-jhwlx\" (UID: \"c6cb07da-9136-445f-b855-19e16f66b46f\") " pod="openshift-image-registry/image-registry-66df7c8f76-jhwlx" Nov 28 13:22:39 crc kubenswrapper[4631]: I1128 13:22:39.231769 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/c6cb07da-9136-445f-b855-19e16f66b46f-ca-trust-extracted\") pod \"image-registry-66df7c8f76-jhwlx\" (UID: \"c6cb07da-9136-445f-b855-19e16f66b46f\") " pod="openshift-image-registry/image-registry-66df7c8f76-jhwlx" Nov 28 13:22:39 crc kubenswrapper[4631]: I1128 13:22:39.231818 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/c6cb07da-9136-445f-b855-19e16f66b46f-registry-certificates\") pod \"image-registry-66df7c8f76-jhwlx\" (UID: \"c6cb07da-9136-445f-b855-19e16f66b46f\") " pod="openshift-image-registry/image-registry-66df7c8f76-jhwlx" Nov 28 13:22:39 crc kubenswrapper[4631]: I1128 13:22:39.231847 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/c6cb07da-9136-445f-b855-19e16f66b46f-bound-sa-token\") pod \"image-registry-66df7c8f76-jhwlx\" (UID: \"c6cb07da-9136-445f-b855-19e16f66b46f\") " pod="openshift-image-registry/image-registry-66df7c8f76-jhwlx" Nov 28 13:22:39 crc kubenswrapper[4631]: I1128 13:22:39.231871 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/c6cb07da-9136-445f-b855-19e16f66b46f-registry-tls\") pod \"image-registry-66df7c8f76-jhwlx\" (UID: \"c6cb07da-9136-445f-b855-19e16f66b46f\") " pod="openshift-image-registry/image-registry-66df7c8f76-jhwlx" Nov 28 13:22:39 crc kubenswrapper[4631]: I1128 13:22:39.233130 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/c6cb07da-9136-445f-b855-19e16f66b46f-ca-trust-extracted\") pod \"image-registry-66df7c8f76-jhwlx\" (UID: \"c6cb07da-9136-445f-b855-19e16f66b46f\") " pod="openshift-image-registry/image-registry-66df7c8f76-jhwlx" Nov 28 13:22:39 crc kubenswrapper[4631]: I1128 13:22:39.233263 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c6cb07da-9136-445f-b855-19e16f66b46f-trusted-ca\") pod \"image-registry-66df7c8f76-jhwlx\" (UID: \"c6cb07da-9136-445f-b855-19e16f66b46f\") " pod="openshift-image-registry/image-registry-66df7c8f76-jhwlx" Nov 28 13:22:39 crc kubenswrapper[4631]: I1128 13:22:39.234090 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/c6cb07da-9136-445f-b855-19e16f66b46f-registry-certificates\") pod \"image-registry-66df7c8f76-jhwlx\" (UID: \"c6cb07da-9136-445f-b855-19e16f66b46f\") " pod="openshift-image-registry/image-registry-66df7c8f76-jhwlx" Nov 28 13:22:39 crc kubenswrapper[4631]: I1128 13:22:39.239401 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/c6cb07da-9136-445f-b855-19e16f66b46f-installation-pull-secrets\") pod \"image-registry-66df7c8f76-jhwlx\" (UID: \"c6cb07da-9136-445f-b855-19e16f66b46f\") " pod="openshift-image-registry/image-registry-66df7c8f76-jhwlx" Nov 28 13:22:39 crc kubenswrapper[4631]: I1128 13:22:39.241966 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/c6cb07da-9136-445f-b855-19e16f66b46f-registry-tls\") pod \"image-registry-66df7c8f76-jhwlx\" (UID: \"c6cb07da-9136-445f-b855-19e16f66b46f\") " pod="openshift-image-registry/image-registry-66df7c8f76-jhwlx" Nov 28 13:22:39 crc kubenswrapper[4631]: I1128 13:22:39.251111 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bsd5d\" (UniqueName: \"kubernetes.io/projected/c6cb07da-9136-445f-b855-19e16f66b46f-kube-api-access-bsd5d\") pod \"image-registry-66df7c8f76-jhwlx\" (UID: \"c6cb07da-9136-445f-b855-19e16f66b46f\") " pod="openshift-image-registry/image-registry-66df7c8f76-jhwlx" Nov 28 13:22:39 crc kubenswrapper[4631]: I1128 13:22:39.253018 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/c6cb07da-9136-445f-b855-19e16f66b46f-bound-sa-token\") pod \"image-registry-66df7c8f76-jhwlx\" (UID: \"c6cb07da-9136-445f-b855-19e16f66b46f\") " pod="openshift-image-registry/image-registry-66df7c8f76-jhwlx" Nov 28 13:22:39 crc kubenswrapper[4631]: I1128 13:22:39.368595 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-jhwlx" Nov 28 13:22:39 crc kubenswrapper[4631]: I1128 13:22:39.831008 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-jhwlx"] Nov 28 13:22:39 crc kubenswrapper[4631]: I1128 13:22:39.975226 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-jhwlx" event={"ID":"c6cb07da-9136-445f-b855-19e16f66b46f","Type":"ContainerStarted","Data":"66ab0c07273471e2bad5eebb49df1aebda3d3d8bbdf5439061ba347de8e64285"} Nov 28 13:22:40 crc kubenswrapper[4631]: I1128 13:22:40.986199 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tpjm2" event={"ID":"2fdd01cb-8e67-4bc4-aa2b-e023882103ce","Type":"ContainerStarted","Data":"33640802287f0cb9a012e6e16d2d519b5b585e7a36f6eed0417001361b83fffd"} Nov 28 13:22:40 crc kubenswrapper[4631]: I1128 13:22:40.989120 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-jhwlx" event={"ID":"c6cb07da-9136-445f-b855-19e16f66b46f","Type":"ContainerStarted","Data":"6ed78f69f83206e9e5225f1fd7c00cd97b904786a8e1ea6e889265e22ab48ef4"} Nov 28 13:22:40 crc kubenswrapper[4631]: I1128 13:22:40.989589 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-jhwlx" Nov 28 13:22:40 crc kubenswrapper[4631]: I1128 13:22:40.992124 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-c7dxv" event={"ID":"45e56fbb-cf6f-4676-a78d-37c59b154fe7","Type":"ContainerStarted","Data":"f1cce648b1c8f5ddfc84196fe92c332a6f6065ef54ab21417926937379fc26b4"} Nov 28 13:22:41 crc kubenswrapper[4631]: I1128 13:22:41.007125 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-tpjm2" podStartSLOduration=2.926310184 podStartE2EDuration="1m9.00710755s" podCreationTimestamp="2025-11-28 13:21:32 +0000 UTC" firstStartedPulling="2025-11-28 13:21:34.061626326 +0000 UTC m=+50.868929670" lastFinishedPulling="2025-11-28 13:22:40.142423692 +0000 UTC m=+116.949727036" observedRunningTime="2025-11-28 13:22:41.004725869 +0000 UTC m=+117.812029213" watchObservedRunningTime="2025-11-28 13:22:41.00710755 +0000 UTC m=+117.814410894" Nov 28 13:22:41 crc kubenswrapper[4631]: I1128 13:22:41.035833 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-jhwlx" podStartSLOduration=2.035811172 podStartE2EDuration="2.035811172s" podCreationTimestamp="2025-11-28 13:22:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 13:22:41.030002032 +0000 UTC m=+117.837305366" watchObservedRunningTime="2025-11-28 13:22:41.035811172 +0000 UTC m=+117.843114516" Nov 28 13:22:41 crc kubenswrapper[4631]: I1128 13:22:41.999521 4631 generic.go:334] "Generic (PLEG): container finished" podID="45e56fbb-cf6f-4676-a78d-37c59b154fe7" containerID="f1cce648b1c8f5ddfc84196fe92c332a6f6065ef54ab21417926937379fc26b4" exitCode=0 Nov 28 13:22:42 crc kubenswrapper[4631]: I1128 13:22:41.999714 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-c7dxv" event={"ID":"45e56fbb-cf6f-4676-a78d-37c59b154fe7","Type":"ContainerDied","Data":"f1cce648b1c8f5ddfc84196fe92c332a6f6065ef54ab21417926937379fc26b4"} Nov 28 13:22:43 crc kubenswrapper[4631]: I1128 13:22:43.019795 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-44v44" event={"ID":"5e49b00c-84e0-4d9c-9095-daaeb2817505","Type":"ContainerStarted","Data":"0658d484b316b0053beed4c248f90d5559d74314f1a55840f9cb998691d366f4"} Nov 28 13:22:43 crc kubenswrapper[4631]: I1128 13:22:43.024228 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-tpjm2" Nov 28 13:22:43 crc kubenswrapper[4631]: I1128 13:22:43.025124 4631 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-tpjm2" Nov 28 13:22:43 crc kubenswrapper[4631]: I1128 13:22:43.025451 4631 generic.go:334] "Generic (PLEG): container finished" podID="10dbdccc-7009-407e-846d-8a2976a58fad" containerID="503d4797bf5ffd1dcd5d9c15a9b8319e91b159e157b945d7d78caa6c21c89da4" exitCode=0 Nov 28 13:22:43 crc kubenswrapper[4631]: I1128 13:22:43.025499 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zkczn" event={"ID":"10dbdccc-7009-407e-846d-8a2976a58fad","Type":"ContainerDied","Data":"503d4797bf5ffd1dcd5d9c15a9b8319e91b159e157b945d7d78caa6c21c89da4"} Nov 28 13:22:43 crc kubenswrapper[4631]: I1128 13:22:43.028708 4631 generic.go:334] "Generic (PLEG): container finished" podID="68b72f7f-4534-480d-983e-04ea986e3b41" containerID="2fd8d8bd8d2e176ad906537a003ed1da101fffb8ff378cc8904b9d2a15d47d06" exitCode=0 Nov 28 13:22:43 crc kubenswrapper[4631]: I1128 13:22:43.028921 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bm9gn" event={"ID":"68b72f7f-4534-480d-983e-04ea986e3b41","Type":"ContainerDied","Data":"2fd8d8bd8d2e176ad906537a003ed1da101fffb8ff378cc8904b9d2a15d47d06"} Nov 28 13:22:43 crc kubenswrapper[4631]: I1128 13:22:43.115112 4631 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-tpjm2" Nov 28 13:22:44 crc kubenswrapper[4631]: I1128 13:22:44.036697 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-c7dxv" event={"ID":"45e56fbb-cf6f-4676-a78d-37c59b154fe7","Type":"ContainerStarted","Data":"c5aa3d11b88805706a6e474810030bc9826180026b5ec33fff46bb58cbc58e71"} Nov 28 13:22:44 crc kubenswrapper[4631]: I1128 13:22:44.038837 4631 generic.go:334] "Generic (PLEG): container finished" podID="5e49b00c-84e0-4d9c-9095-daaeb2817505" containerID="0658d484b316b0053beed4c248f90d5559d74314f1a55840f9cb998691d366f4" exitCode=0 Nov 28 13:22:44 crc kubenswrapper[4631]: I1128 13:22:44.038894 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-44v44" event={"ID":"5e49b00c-84e0-4d9c-9095-daaeb2817505","Type":"ContainerDied","Data":"0658d484b316b0053beed4c248f90d5559d74314f1a55840f9cb998691d366f4"} Nov 28 13:22:44 crc kubenswrapper[4631]: I1128 13:22:44.040929 4631 generic.go:334] "Generic (PLEG): container finished" podID="67afe4ac-28a7-42a2-8155-2c4bf496b834" containerID="dee2a84c2481360ae63392e497b8f84b25b55abe81bf3ce3655f99641cd8b6b9" exitCode=0 Nov 28 13:22:44 crc kubenswrapper[4631]: I1128 13:22:44.041105 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-p5v2z" event={"ID":"67afe4ac-28a7-42a2-8155-2c4bf496b834","Type":"ContainerDied","Data":"dee2a84c2481360ae63392e497b8f84b25b55abe81bf3ce3655f99641cd8b6b9"} Nov 28 13:22:44 crc kubenswrapper[4631]: I1128 13:22:44.075836 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-c7dxv" podStartSLOduration=5.241352108 podStartE2EDuration="1m9.075811865s" podCreationTimestamp="2025-11-28 13:21:35 +0000 UTC" firstStartedPulling="2025-11-28 13:21:39.261738958 +0000 UTC m=+56.069042302" lastFinishedPulling="2025-11-28 13:22:43.096198725 +0000 UTC m=+119.903502059" observedRunningTime="2025-11-28 13:22:44.056612378 +0000 UTC m=+120.863915722" watchObservedRunningTime="2025-11-28 13:22:44.075811865 +0000 UTC m=+120.883115209" Nov 28 13:22:45 crc kubenswrapper[4631]: I1128 13:22:45.088447 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-tpjm2" Nov 28 13:22:46 crc kubenswrapper[4631]: I1128 13:22:46.210740 4631 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-c7dxv" Nov 28 13:22:46 crc kubenswrapper[4631]: I1128 13:22:46.211521 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-c7dxv" Nov 28 13:22:47 crc kubenswrapper[4631]: I1128 13:22:47.057162 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zkczn" event={"ID":"10dbdccc-7009-407e-846d-8a2976a58fad","Type":"ContainerStarted","Data":"57c6a81a1ccc4d4be112631d8e61f7ecd93cf2f19dc6c42ceb51c1fb1ea4c0b4"} Nov 28 13:22:47 crc kubenswrapper[4631]: I1128 13:22:47.058915 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bm9gn" event={"ID":"68b72f7f-4534-480d-983e-04ea986e3b41","Type":"ContainerStarted","Data":"6874003d194b999f1822c5253b067abe1a2df13cd079b9f57b230e036928603e"} Nov 28 13:22:47 crc kubenswrapper[4631]: I1128 13:22:47.077973 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-zkczn" podStartSLOduration=1.617903026 podStartE2EDuration="1m14.077957769s" podCreationTimestamp="2025-11-28 13:21:33 +0000 UTC" firstStartedPulling="2025-11-28 13:21:34.057343128 +0000 UTC m=+50.864646462" lastFinishedPulling="2025-11-28 13:22:46.517397861 +0000 UTC m=+123.324701205" observedRunningTime="2025-11-28 13:22:47.074584492 +0000 UTC m=+123.881887846" watchObservedRunningTime="2025-11-28 13:22:47.077957769 +0000 UTC m=+123.885261103" Nov 28 13:22:47 crc kubenswrapper[4631]: I1128 13:22:47.093149 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-bm9gn" podStartSLOduration=5.107312236 podStartE2EDuration="1m12.093132941s" podCreationTimestamp="2025-11-28 13:21:35 +0000 UTC" firstStartedPulling="2025-11-28 13:21:39.253522248 +0000 UTC m=+56.060825602" lastFinishedPulling="2025-11-28 13:22:46.239342963 +0000 UTC m=+123.046646307" observedRunningTime="2025-11-28 13:22:47.092643549 +0000 UTC m=+123.899946893" watchObservedRunningTime="2025-11-28 13:22:47.093132941 +0000 UTC m=+123.900436285" Nov 28 13:22:47 crc kubenswrapper[4631]: I1128 13:22:47.313858 4631 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-c7dxv" podUID="45e56fbb-cf6f-4676-a78d-37c59b154fe7" containerName="registry-server" probeResult="failure" output=< Nov 28 13:22:47 crc kubenswrapper[4631]: timeout: failed to connect service ":50051" within 1s Nov 28 13:22:47 crc kubenswrapper[4631]: > Nov 28 13:22:48 crc kubenswrapper[4631]: I1128 13:22:48.751604 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-hgsfk"] Nov 28 13:22:49 crc kubenswrapper[4631]: I1128 13:22:49.267045 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-tpjm2"] Nov 28 13:22:49 crc kubenswrapper[4631]: I1128 13:22:49.267418 4631 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-tpjm2" podUID="2fdd01cb-8e67-4bc4-aa2b-e023882103ce" containerName="registry-server" containerID="cri-o://33640802287f0cb9a012e6e16d2d519b5b585e7a36f6eed0417001361b83fffd" gracePeriod=30 Nov 28 13:22:49 crc kubenswrapper[4631]: I1128 13:22:49.282240 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-zkczn"] Nov 28 13:22:49 crc kubenswrapper[4631]: I1128 13:22:49.283361 4631 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-zkczn" podUID="10dbdccc-7009-407e-846d-8a2976a58fad" containerName="registry-server" containerID="cri-o://57c6a81a1ccc4d4be112631d8e61f7ecd93cf2f19dc6c42ceb51c1fb1ea4c0b4" gracePeriod=30 Nov 28 13:22:49 crc kubenswrapper[4631]: I1128 13:22:49.299988 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-4trst"] Nov 28 13:22:49 crc kubenswrapper[4631]: I1128 13:22:49.317991 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-6nrc5"] Nov 28 13:22:49 crc kubenswrapper[4631]: I1128 13:22:49.322813 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-2bc4b"] Nov 28 13:22:49 crc kubenswrapper[4631]: I1128 13:22:49.323018 4631 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-2bc4b" podUID="9eb418ce-65b6-4da3-b282-831dd62916cf" containerName="marketplace-operator" containerID="cri-o://48793a8f9c2488bf82d5c7596352e34ede9baff8e45779b3bb05fecc4a16daef" gracePeriod=30 Nov 28 13:22:49 crc kubenswrapper[4631]: I1128 13:22:49.336995 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-bm9gn"] Nov 28 13:22:49 crc kubenswrapper[4631]: I1128 13:22:49.337401 4631 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-bm9gn" podUID="68b72f7f-4534-480d-983e-04ea986e3b41" containerName="registry-server" containerID="cri-o://6874003d194b999f1822c5253b067abe1a2df13cd079b9f57b230e036928603e" gracePeriod=30 Nov 28 13:22:49 crc kubenswrapper[4631]: I1128 13:22:49.348723 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-p5v2z"] Nov 28 13:22:49 crc kubenswrapper[4631]: I1128 13:22:49.354355 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-gsrzv"] Nov 28 13:22:49 crc kubenswrapper[4631]: I1128 13:22:49.355334 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-gsrzv" Nov 28 13:22:49 crc kubenswrapper[4631]: I1128 13:22:49.363251 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-44v44"] Nov 28 13:22:49 crc kubenswrapper[4631]: I1128 13:22:49.373221 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-c7dxv"] Nov 28 13:22:49 crc kubenswrapper[4631]: I1128 13:22:49.378765 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-gsrzv"] Nov 28 13:22:49 crc kubenswrapper[4631]: I1128 13:22:49.379052 4631 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-c7dxv" podUID="45e56fbb-cf6f-4676-a78d-37c59b154fe7" containerName="registry-server" containerID="cri-o://c5aa3d11b88805706a6e474810030bc9826180026b5ec33fff46bb58cbc58e71" gracePeriod=30 Nov 28 13:22:49 crc kubenswrapper[4631]: I1128 13:22:49.392492 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/4eb86ab5-b771-49dc-8bfb-49ce46c0c712-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-gsrzv\" (UID: \"4eb86ab5-b771-49dc-8bfb-49ce46c0c712\") " pod="openshift-marketplace/marketplace-operator-79b997595-gsrzv" Nov 28 13:22:49 crc kubenswrapper[4631]: I1128 13:22:49.392593 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c7wbf\" (UniqueName: \"kubernetes.io/projected/4eb86ab5-b771-49dc-8bfb-49ce46c0c712-kube-api-access-c7wbf\") pod \"marketplace-operator-79b997595-gsrzv\" (UID: \"4eb86ab5-b771-49dc-8bfb-49ce46c0c712\") " pod="openshift-marketplace/marketplace-operator-79b997595-gsrzv" Nov 28 13:22:49 crc kubenswrapper[4631]: I1128 13:22:49.392636 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/4eb86ab5-b771-49dc-8bfb-49ce46c0c712-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-gsrzv\" (UID: \"4eb86ab5-b771-49dc-8bfb-49ce46c0c712\") " pod="openshift-marketplace/marketplace-operator-79b997595-gsrzv" Nov 28 13:22:49 crc kubenswrapper[4631]: I1128 13:22:49.494698 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/4eb86ab5-b771-49dc-8bfb-49ce46c0c712-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-gsrzv\" (UID: \"4eb86ab5-b771-49dc-8bfb-49ce46c0c712\") " pod="openshift-marketplace/marketplace-operator-79b997595-gsrzv" Nov 28 13:22:49 crc kubenswrapper[4631]: I1128 13:22:49.494849 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/4eb86ab5-b771-49dc-8bfb-49ce46c0c712-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-gsrzv\" (UID: \"4eb86ab5-b771-49dc-8bfb-49ce46c0c712\") " pod="openshift-marketplace/marketplace-operator-79b997595-gsrzv" Nov 28 13:22:49 crc kubenswrapper[4631]: I1128 13:22:49.495062 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c7wbf\" (UniqueName: \"kubernetes.io/projected/4eb86ab5-b771-49dc-8bfb-49ce46c0c712-kube-api-access-c7wbf\") pod \"marketplace-operator-79b997595-gsrzv\" (UID: \"4eb86ab5-b771-49dc-8bfb-49ce46c0c712\") " pod="openshift-marketplace/marketplace-operator-79b997595-gsrzv" Nov 28 13:22:49 crc kubenswrapper[4631]: I1128 13:22:49.498672 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/4eb86ab5-b771-49dc-8bfb-49ce46c0c712-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-gsrzv\" (UID: \"4eb86ab5-b771-49dc-8bfb-49ce46c0c712\") " pod="openshift-marketplace/marketplace-operator-79b997595-gsrzv" Nov 28 13:22:49 crc kubenswrapper[4631]: I1128 13:22:49.503951 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/4eb86ab5-b771-49dc-8bfb-49ce46c0c712-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-gsrzv\" (UID: \"4eb86ab5-b771-49dc-8bfb-49ce46c0c712\") " pod="openshift-marketplace/marketplace-operator-79b997595-gsrzv" Nov 28 13:22:49 crc kubenswrapper[4631]: I1128 13:22:49.528945 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c7wbf\" (UniqueName: \"kubernetes.io/projected/4eb86ab5-b771-49dc-8bfb-49ce46c0c712-kube-api-access-c7wbf\") pod \"marketplace-operator-79b997595-gsrzv\" (UID: \"4eb86ab5-b771-49dc-8bfb-49ce46c0c712\") " pod="openshift-marketplace/marketplace-operator-79b997595-gsrzv" Nov 28 13:22:49 crc kubenswrapper[4631]: I1128 13:22:49.679334 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-gsrzv" Nov 28 13:22:50 crc kubenswrapper[4631]: I1128 13:22:50.482230 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-gsrzv"] Nov 28 13:22:50 crc kubenswrapper[4631]: W1128 13:22:50.489066 4631 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4eb86ab5_b771_49dc_8bfb_49ce46c0c712.slice/crio-34a261810da738a765c9d5cbc78ba57f068312e7bf3a9ff0625554da5e1359ed WatchSource:0}: Error finding container 34a261810da738a765c9d5cbc78ba57f068312e7bf3a9ff0625554da5e1359ed: Status 404 returned error can't find the container with id 34a261810da738a765c9d5cbc78ba57f068312e7bf3a9ff0625554da5e1359ed Nov 28 13:22:51 crc kubenswrapper[4631]: I1128 13:22:51.153915 4631 generic.go:334] "Generic (PLEG): container finished" podID="10dbdccc-7009-407e-846d-8a2976a58fad" containerID="57c6a81a1ccc4d4be112631d8e61f7ecd93cf2f19dc6c42ceb51c1fb1ea4c0b4" exitCode=0 Nov 28 13:22:51 crc kubenswrapper[4631]: I1128 13:22:51.153996 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zkczn" event={"ID":"10dbdccc-7009-407e-846d-8a2976a58fad","Type":"ContainerDied","Data":"57c6a81a1ccc4d4be112631d8e61f7ecd93cf2f19dc6c42ceb51c1fb1ea4c0b4"} Nov 28 13:22:51 crc kubenswrapper[4631]: I1128 13:22:51.161188 4631 generic.go:334] "Generic (PLEG): container finished" podID="68b72f7f-4534-480d-983e-04ea986e3b41" containerID="6874003d194b999f1822c5253b067abe1a2df13cd079b9f57b230e036928603e" exitCode=0 Nov 28 13:22:51 crc kubenswrapper[4631]: I1128 13:22:51.161254 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bm9gn" event={"ID":"68b72f7f-4534-480d-983e-04ea986e3b41","Type":"ContainerDied","Data":"6874003d194b999f1822c5253b067abe1a2df13cd079b9f57b230e036928603e"} Nov 28 13:22:51 crc kubenswrapper[4631]: I1128 13:22:51.163272 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6nrc5" event={"ID":"72b60e37-5508-451f-8602-9dcebd55324e","Type":"ContainerStarted","Data":"a93f839d3b8e8c46ac42d69047324d2174189284ad3257b716d7d644c9c703a3"} Nov 28 13:22:51 crc kubenswrapper[4631]: I1128 13:22:51.163478 4631 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-6nrc5" podUID="72b60e37-5508-451f-8602-9dcebd55324e" containerName="extract-content" containerID="cri-o://a93f839d3b8e8c46ac42d69047324d2174189284ad3257b716d7d644c9c703a3" gracePeriod=30 Nov 28 13:22:51 crc kubenswrapper[4631]: I1128 13:22:51.173197 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-gsrzv" event={"ID":"4eb86ab5-b771-49dc-8bfb-49ce46c0c712","Type":"ContainerStarted","Data":"2078a2f0667051b98511c38e5302c75844d471029b97f73b4434d5b5664f536c"} Nov 28 13:22:51 crc kubenswrapper[4631]: I1128 13:22:51.173247 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-gsrzv" event={"ID":"4eb86ab5-b771-49dc-8bfb-49ce46c0c712","Type":"ContainerStarted","Data":"34a261810da738a765c9d5cbc78ba57f068312e7bf3a9ff0625554da5e1359ed"} Nov 28 13:22:51 crc kubenswrapper[4631]: I1128 13:22:51.173629 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-gsrzv" Nov 28 13:22:51 crc kubenswrapper[4631]: I1128 13:22:51.175491 4631 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-gsrzv container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.57:8080/healthz\": dial tcp 10.217.0.57:8080: connect: connection refused" start-of-body= Nov 28 13:22:51 crc kubenswrapper[4631]: I1128 13:22:51.175547 4631 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-gsrzv" podUID="4eb86ab5-b771-49dc-8bfb-49ce46c0c712" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.57:8080/healthz\": dial tcp 10.217.0.57:8080: connect: connection refused" Nov 28 13:22:51 crc kubenswrapper[4631]: I1128 13:22:51.184687 4631 generic.go:334] "Generic (PLEG): container finished" podID="45e56fbb-cf6f-4676-a78d-37c59b154fe7" containerID="c5aa3d11b88805706a6e474810030bc9826180026b5ec33fff46bb58cbc58e71" exitCode=0 Nov 28 13:22:51 crc kubenswrapper[4631]: I1128 13:22:51.184844 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-c7dxv" event={"ID":"45e56fbb-cf6f-4676-a78d-37c59b154fe7","Type":"ContainerDied","Data":"c5aa3d11b88805706a6e474810030bc9826180026b5ec33fff46bb58cbc58e71"} Nov 28 13:22:51 crc kubenswrapper[4631]: I1128 13:22:51.190062 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4trst" event={"ID":"f6675fed-3a90-4779-9b95-1acba5248fd2","Type":"ContainerStarted","Data":"ed26ab582f1984f67baaad6fb180ac3491c3237d286bfdf88f4503175b8079b3"} Nov 28 13:22:51 crc kubenswrapper[4631]: I1128 13:22:51.190324 4631 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-4trst" podUID="f6675fed-3a90-4779-9b95-1acba5248fd2" containerName="extract-content" containerID="cri-o://ed26ab582f1984f67baaad6fb180ac3491c3237d286bfdf88f4503175b8079b3" gracePeriod=30 Nov 28 13:22:51 crc kubenswrapper[4631]: I1128 13:22:51.207377 4631 generic.go:334] "Generic (PLEG): container finished" podID="2fdd01cb-8e67-4bc4-aa2b-e023882103ce" containerID="33640802287f0cb9a012e6e16d2d519b5b585e7a36f6eed0417001361b83fffd" exitCode=0 Nov 28 13:22:51 crc kubenswrapper[4631]: I1128 13:22:51.207456 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tpjm2" event={"ID":"2fdd01cb-8e67-4bc4-aa2b-e023882103ce","Type":"ContainerDied","Data":"33640802287f0cb9a012e6e16d2d519b5b585e7a36f6eed0417001361b83fffd"} Nov 28 13:22:51 crc kubenswrapper[4631]: I1128 13:22:51.244242 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-44v44" event={"ID":"5e49b00c-84e0-4d9c-9095-daaeb2817505","Type":"ContainerStarted","Data":"20a58c60bdb60307ad0e09e4dde48caf094b28dc5fcdf9d45cc2b407486b3463"} Nov 28 13:22:51 crc kubenswrapper[4631]: I1128 13:22:51.244461 4631 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-44v44" podUID="5e49b00c-84e0-4d9c-9095-daaeb2817505" containerName="registry-server" containerID="cri-o://20a58c60bdb60307ad0e09e4dde48caf094b28dc5fcdf9d45cc2b407486b3463" gracePeriod=30 Nov 28 13:22:51 crc kubenswrapper[4631]: I1128 13:22:51.251999 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-p5v2z" event={"ID":"67afe4ac-28a7-42a2-8155-2c4bf496b834","Type":"ContainerStarted","Data":"14e17a95f26adc182db616f80169b29b7bfc52ce3c1aa51772bbdac6228a6bb1"} Nov 28 13:22:51 crc kubenswrapper[4631]: I1128 13:22:51.252346 4631 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-p5v2z" podUID="67afe4ac-28a7-42a2-8155-2c4bf496b834" containerName="registry-server" containerID="cri-o://14e17a95f26adc182db616f80169b29b7bfc52ce3c1aa51772bbdac6228a6bb1" gracePeriod=30 Nov 28 13:22:51 crc kubenswrapper[4631]: I1128 13:22:51.270788 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-gsrzv" podStartSLOduration=2.270765677 podStartE2EDuration="2.270765677s" podCreationTimestamp="2025-11-28 13:22:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 13:22:51.255000849 +0000 UTC m=+128.062304203" watchObservedRunningTime="2025-11-28 13:22:51.270765677 +0000 UTC m=+128.078069021" Nov 28 13:22:51 crc kubenswrapper[4631]: I1128 13:22:51.289185 4631 generic.go:334] "Generic (PLEG): container finished" podID="9eb418ce-65b6-4da3-b282-831dd62916cf" containerID="48793a8f9c2488bf82d5c7596352e34ede9baff8e45779b3bb05fecc4a16daef" exitCode=0 Nov 28 13:22:51 crc kubenswrapper[4631]: I1128 13:22:51.289256 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-2bc4b" event={"ID":"9eb418ce-65b6-4da3-b282-831dd62916cf","Type":"ContainerDied","Data":"48793a8f9c2488bf82d5c7596352e34ede9baff8e45779b3bb05fecc4a16daef"} Nov 28 13:22:51 crc kubenswrapper[4631]: I1128 13:22:51.322440 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zkczn" Nov 28 13:22:51 crc kubenswrapper[4631]: I1128 13:22:51.326744 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-44v44" podStartSLOduration=6.976547056 podStartE2EDuration="1m15.326722654s" podCreationTimestamp="2025-11-28 13:21:36 +0000 UTC" firstStartedPulling="2025-11-28 13:21:39.267633014 +0000 UTC m=+56.074936368" lastFinishedPulling="2025-11-28 13:22:47.617808622 +0000 UTC m=+124.425111966" observedRunningTime="2025-11-28 13:22:51.31844091 +0000 UTC m=+128.125744254" watchObservedRunningTime="2025-11-28 13:22:51.326722654 +0000 UTC m=+128.134025998" Nov 28 13:22:51 crc kubenswrapper[4631]: I1128 13:22:51.327252 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-p5v2z" podStartSLOduration=6.7565277120000005 podStartE2EDuration="1m17.327246947s" podCreationTimestamp="2025-11-28 13:21:34 +0000 UTC" firstStartedPulling="2025-11-28 13:21:39.276717124 +0000 UTC m=+56.084020468" lastFinishedPulling="2025-11-28 13:22:49.847436359 +0000 UTC m=+126.654739703" observedRunningTime="2025-11-28 13:22:51.292396306 +0000 UTC m=+128.099699650" watchObservedRunningTime="2025-11-28 13:22:51.327246947 +0000 UTC m=+128.134550291" Nov 28 13:22:51 crc kubenswrapper[4631]: I1128 13:22:51.420899 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tfshw\" (UniqueName: \"kubernetes.io/projected/10dbdccc-7009-407e-846d-8a2976a58fad-kube-api-access-tfshw\") pod \"10dbdccc-7009-407e-846d-8a2976a58fad\" (UID: \"10dbdccc-7009-407e-846d-8a2976a58fad\") " Nov 28 13:22:51 crc kubenswrapper[4631]: I1128 13:22:51.420962 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/10dbdccc-7009-407e-846d-8a2976a58fad-catalog-content\") pod \"10dbdccc-7009-407e-846d-8a2976a58fad\" (UID: \"10dbdccc-7009-407e-846d-8a2976a58fad\") " Nov 28 13:22:51 crc kubenswrapper[4631]: I1128 13:22:51.420991 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/10dbdccc-7009-407e-846d-8a2976a58fad-utilities\") pod \"10dbdccc-7009-407e-846d-8a2976a58fad\" (UID: \"10dbdccc-7009-407e-846d-8a2976a58fad\") " Nov 28 13:22:51 crc kubenswrapper[4631]: I1128 13:22:51.422114 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/10dbdccc-7009-407e-846d-8a2976a58fad-utilities" (OuterVolumeSpecName: "utilities") pod "10dbdccc-7009-407e-846d-8a2976a58fad" (UID: "10dbdccc-7009-407e-846d-8a2976a58fad"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 13:22:51 crc kubenswrapper[4631]: I1128 13:22:51.427480 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/10dbdccc-7009-407e-846d-8a2976a58fad-kube-api-access-tfshw" (OuterVolumeSpecName: "kube-api-access-tfshw") pod "10dbdccc-7009-407e-846d-8a2976a58fad" (UID: "10dbdccc-7009-407e-846d-8a2976a58fad"). InnerVolumeSpecName "kube-api-access-tfshw". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:22:51 crc kubenswrapper[4631]: I1128 13:22:51.517976 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/10dbdccc-7009-407e-846d-8a2976a58fad-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "10dbdccc-7009-407e-846d-8a2976a58fad" (UID: "10dbdccc-7009-407e-846d-8a2976a58fad"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 13:22:51 crc kubenswrapper[4631]: I1128 13:22:51.523447 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tfshw\" (UniqueName: \"kubernetes.io/projected/10dbdccc-7009-407e-846d-8a2976a58fad-kube-api-access-tfshw\") on node \"crc\" DevicePath \"\"" Nov 28 13:22:51 crc kubenswrapper[4631]: I1128 13:22:51.523499 4631 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/10dbdccc-7009-407e-846d-8a2976a58fad-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 28 13:22:51 crc kubenswrapper[4631]: I1128 13:22:51.523512 4631 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/10dbdccc-7009-407e-846d-8a2976a58fad-utilities\") on node \"crc\" DevicePath \"\"" Nov 28 13:22:51 crc kubenswrapper[4631]: I1128 13:22:51.549698 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bm9gn" Nov 28 13:22:51 crc kubenswrapper[4631]: I1128 13:22:51.627235 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dk86v\" (UniqueName: \"kubernetes.io/projected/68b72f7f-4534-480d-983e-04ea986e3b41-kube-api-access-dk86v\") pod \"68b72f7f-4534-480d-983e-04ea986e3b41\" (UID: \"68b72f7f-4534-480d-983e-04ea986e3b41\") " Nov 28 13:22:51 crc kubenswrapper[4631]: I1128 13:22:51.627322 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/68b72f7f-4534-480d-983e-04ea986e3b41-utilities\") pod \"68b72f7f-4534-480d-983e-04ea986e3b41\" (UID: \"68b72f7f-4534-480d-983e-04ea986e3b41\") " Nov 28 13:22:51 crc kubenswrapper[4631]: I1128 13:22:51.627354 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/68b72f7f-4534-480d-983e-04ea986e3b41-catalog-content\") pod \"68b72f7f-4534-480d-983e-04ea986e3b41\" (UID: \"68b72f7f-4534-480d-983e-04ea986e3b41\") " Nov 28 13:22:51 crc kubenswrapper[4631]: I1128 13:22:51.631599 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/68b72f7f-4534-480d-983e-04ea986e3b41-utilities" (OuterVolumeSpecName: "utilities") pod "68b72f7f-4534-480d-983e-04ea986e3b41" (UID: "68b72f7f-4534-480d-983e-04ea986e3b41"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 13:22:51 crc kubenswrapper[4631]: I1128 13:22:51.637427 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/68b72f7f-4534-480d-983e-04ea986e3b41-kube-api-access-dk86v" (OuterVolumeSpecName: "kube-api-access-dk86v") pod "68b72f7f-4534-480d-983e-04ea986e3b41" (UID: "68b72f7f-4534-480d-983e-04ea986e3b41"). InnerVolumeSpecName "kube-api-access-dk86v". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:22:51 crc kubenswrapper[4631]: I1128 13:22:51.651962 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/68b72f7f-4534-480d-983e-04ea986e3b41-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "68b72f7f-4534-480d-983e-04ea986e3b41" (UID: "68b72f7f-4534-480d-983e-04ea986e3b41"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 13:22:51 crc kubenswrapper[4631]: I1128 13:22:51.673497 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-tpjm2" Nov 28 13:22:51 crc kubenswrapper[4631]: I1128 13:22:51.729823 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2fdd01cb-8e67-4bc4-aa2b-e023882103ce-catalog-content\") pod \"2fdd01cb-8e67-4bc4-aa2b-e023882103ce\" (UID: \"2fdd01cb-8e67-4bc4-aa2b-e023882103ce\") " Nov 28 13:22:51 crc kubenswrapper[4631]: I1128 13:22:51.730223 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d52xh\" (UniqueName: \"kubernetes.io/projected/2fdd01cb-8e67-4bc4-aa2b-e023882103ce-kube-api-access-d52xh\") pod \"2fdd01cb-8e67-4bc4-aa2b-e023882103ce\" (UID: \"2fdd01cb-8e67-4bc4-aa2b-e023882103ce\") " Nov 28 13:22:51 crc kubenswrapper[4631]: I1128 13:22:51.730399 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2fdd01cb-8e67-4bc4-aa2b-e023882103ce-utilities\") pod \"2fdd01cb-8e67-4bc4-aa2b-e023882103ce\" (UID: \"2fdd01cb-8e67-4bc4-aa2b-e023882103ce\") " Nov 28 13:22:51 crc kubenswrapper[4631]: I1128 13:22:51.731652 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dk86v\" (UniqueName: \"kubernetes.io/projected/68b72f7f-4534-480d-983e-04ea986e3b41-kube-api-access-dk86v\") on node \"crc\" DevicePath \"\"" Nov 28 13:22:51 crc kubenswrapper[4631]: I1128 13:22:51.731705 4631 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/68b72f7f-4534-480d-983e-04ea986e3b41-utilities\") on node \"crc\" DevicePath \"\"" Nov 28 13:22:51 crc kubenswrapper[4631]: I1128 13:22:51.731717 4631 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/68b72f7f-4534-480d-983e-04ea986e3b41-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 28 13:22:51 crc kubenswrapper[4631]: I1128 13:22:51.733408 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2fdd01cb-8e67-4bc4-aa2b-e023882103ce-utilities" (OuterVolumeSpecName: "utilities") pod "2fdd01cb-8e67-4bc4-aa2b-e023882103ce" (UID: "2fdd01cb-8e67-4bc4-aa2b-e023882103ce"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 13:22:51 crc kubenswrapper[4631]: I1128 13:22:51.737605 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2fdd01cb-8e67-4bc4-aa2b-e023882103ce-kube-api-access-d52xh" (OuterVolumeSpecName: "kube-api-access-d52xh") pod "2fdd01cb-8e67-4bc4-aa2b-e023882103ce" (UID: "2fdd01cb-8e67-4bc4-aa2b-e023882103ce"). InnerVolumeSpecName "kube-api-access-d52xh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:22:51 crc kubenswrapper[4631]: I1128 13:22:51.826405 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2fdd01cb-8e67-4bc4-aa2b-e023882103ce-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2fdd01cb-8e67-4bc4-aa2b-e023882103ce" (UID: "2fdd01cb-8e67-4bc4-aa2b-e023882103ce"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 13:22:51 crc kubenswrapper[4631]: I1128 13:22:51.833733 4631 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2fdd01cb-8e67-4bc4-aa2b-e023882103ce-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 28 13:22:51 crc kubenswrapper[4631]: I1128 13:22:51.833770 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d52xh\" (UniqueName: \"kubernetes.io/projected/2fdd01cb-8e67-4bc4-aa2b-e023882103ce-kube-api-access-d52xh\") on node \"crc\" DevicePath \"\"" Nov 28 13:22:51 crc kubenswrapper[4631]: I1128 13:22:51.833786 4631 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2fdd01cb-8e67-4bc4-aa2b-e023882103ce-utilities\") on node \"crc\" DevicePath \"\"" Nov 28 13:22:51 crc kubenswrapper[4631]: I1128 13:22:51.844811 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-2bc4b" Nov 28 13:22:51 crc kubenswrapper[4631]: I1128 13:22:51.935191 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/9eb418ce-65b6-4da3-b282-831dd62916cf-marketplace-operator-metrics\") pod \"9eb418ce-65b6-4da3-b282-831dd62916cf\" (UID: \"9eb418ce-65b6-4da3-b282-831dd62916cf\") " Nov 28 13:22:51 crc kubenswrapper[4631]: I1128 13:22:51.935343 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bgtn5\" (UniqueName: \"kubernetes.io/projected/9eb418ce-65b6-4da3-b282-831dd62916cf-kube-api-access-bgtn5\") pod \"9eb418ce-65b6-4da3-b282-831dd62916cf\" (UID: \"9eb418ce-65b6-4da3-b282-831dd62916cf\") " Nov 28 13:22:51 crc kubenswrapper[4631]: I1128 13:22:51.935406 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9eb418ce-65b6-4da3-b282-831dd62916cf-marketplace-trusted-ca\") pod \"9eb418ce-65b6-4da3-b282-831dd62916cf\" (UID: \"9eb418ce-65b6-4da3-b282-831dd62916cf\") " Nov 28 13:22:51 crc kubenswrapper[4631]: I1128 13:22:51.936219 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9eb418ce-65b6-4da3-b282-831dd62916cf-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "9eb418ce-65b6-4da3-b282-831dd62916cf" (UID: "9eb418ce-65b6-4da3-b282-831dd62916cf"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 13:22:51 crc kubenswrapper[4631]: I1128 13:22:51.938548 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9eb418ce-65b6-4da3-b282-831dd62916cf-kube-api-access-bgtn5" (OuterVolumeSpecName: "kube-api-access-bgtn5") pod "9eb418ce-65b6-4da3-b282-831dd62916cf" (UID: "9eb418ce-65b6-4da3-b282-831dd62916cf"). InnerVolumeSpecName "kube-api-access-bgtn5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:22:51 crc kubenswrapper[4631]: I1128 13:22:51.941735 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9eb418ce-65b6-4da3-b282-831dd62916cf-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "9eb418ce-65b6-4da3-b282-831dd62916cf" (UID: "9eb418ce-65b6-4da3-b282-831dd62916cf"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:22:51 crc kubenswrapper[4631]: I1128 13:22:51.969223 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4trst" Nov 28 13:22:52 crc kubenswrapper[4631]: I1128 13:22:52.037095 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f6675fed-3a90-4779-9b95-1acba5248fd2-utilities\") pod \"f6675fed-3a90-4779-9b95-1acba5248fd2\" (UID: \"f6675fed-3a90-4779-9b95-1acba5248fd2\") " Nov 28 13:22:52 crc kubenswrapper[4631]: I1128 13:22:52.037244 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f6675fed-3a90-4779-9b95-1acba5248fd2-catalog-content\") pod \"f6675fed-3a90-4779-9b95-1acba5248fd2\" (UID: \"f6675fed-3a90-4779-9b95-1acba5248fd2\") " Nov 28 13:22:52 crc kubenswrapper[4631]: I1128 13:22:52.037347 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-njf98\" (UniqueName: \"kubernetes.io/projected/f6675fed-3a90-4779-9b95-1acba5248fd2-kube-api-access-njf98\") pod \"f6675fed-3a90-4779-9b95-1acba5248fd2\" (UID: \"f6675fed-3a90-4779-9b95-1acba5248fd2\") " Nov 28 13:22:52 crc kubenswrapper[4631]: I1128 13:22:52.037908 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f6675fed-3a90-4779-9b95-1acba5248fd2-utilities" (OuterVolumeSpecName: "utilities") pod "f6675fed-3a90-4779-9b95-1acba5248fd2" (UID: "f6675fed-3a90-4779-9b95-1acba5248fd2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 13:22:52 crc kubenswrapper[4631]: I1128 13:22:52.039712 4631 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f6675fed-3a90-4779-9b95-1acba5248fd2-utilities\") on node \"crc\" DevicePath \"\"" Nov 28 13:22:52 crc kubenswrapper[4631]: I1128 13:22:52.039732 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bgtn5\" (UniqueName: \"kubernetes.io/projected/9eb418ce-65b6-4da3-b282-831dd62916cf-kube-api-access-bgtn5\") on node \"crc\" DevicePath \"\"" Nov 28 13:22:52 crc kubenswrapper[4631]: I1128 13:22:52.039744 4631 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9eb418ce-65b6-4da3-b282-831dd62916cf-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 28 13:22:52 crc kubenswrapper[4631]: I1128 13:22:52.039753 4631 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/9eb418ce-65b6-4da3-b282-831dd62916cf-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Nov 28 13:22:52 crc kubenswrapper[4631]: I1128 13:22:52.040056 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f6675fed-3a90-4779-9b95-1acba5248fd2-kube-api-access-njf98" (OuterVolumeSpecName: "kube-api-access-njf98") pod "f6675fed-3a90-4779-9b95-1acba5248fd2" (UID: "f6675fed-3a90-4779-9b95-1acba5248fd2"). InnerVolumeSpecName "kube-api-access-njf98". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:22:52 crc kubenswrapper[4631]: I1128 13:22:52.092197 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f6675fed-3a90-4779-9b95-1acba5248fd2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f6675fed-3a90-4779-9b95-1acba5248fd2" (UID: "f6675fed-3a90-4779-9b95-1acba5248fd2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 13:22:52 crc kubenswrapper[4631]: I1128 13:22:52.141520 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-njf98\" (UniqueName: \"kubernetes.io/projected/f6675fed-3a90-4779-9b95-1acba5248fd2-kube-api-access-njf98\") on node \"crc\" DevicePath \"\"" Nov 28 13:22:52 crc kubenswrapper[4631]: I1128 13:22:52.141560 4631 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f6675fed-3a90-4779-9b95-1acba5248fd2-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 28 13:22:52 crc kubenswrapper[4631]: I1128 13:22:52.303961 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zkczn" event={"ID":"10dbdccc-7009-407e-846d-8a2976a58fad","Type":"ContainerDied","Data":"f4d5d88ce6700a920b115de3d73ffb17405cef6dae06bfcdb52097b5ab21fe95"} Nov 28 13:22:52 crc kubenswrapper[4631]: I1128 13:22:52.304041 4631 scope.go:117] "RemoveContainer" containerID="57c6a81a1ccc4d4be112631d8e61f7ecd93cf2f19dc6c42ceb51c1fb1ea4c0b4" Nov 28 13:22:52 crc kubenswrapper[4631]: I1128 13:22:52.304246 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zkczn" Nov 28 13:22:52 crc kubenswrapper[4631]: I1128 13:22:52.309860 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bm9gn" event={"ID":"68b72f7f-4534-480d-983e-04ea986e3b41","Type":"ContainerDied","Data":"783903b653dc550d97d798d90daa227b069f82b54c7bf0fdcde2cd65494cb53c"} Nov 28 13:22:52 crc kubenswrapper[4631]: I1128 13:22:52.309960 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bm9gn" Nov 28 13:22:52 crc kubenswrapper[4631]: I1128 13:22:52.315838 4631 generic.go:334] "Generic (PLEG): container finished" podID="f6675fed-3a90-4779-9b95-1acba5248fd2" containerID="ed26ab582f1984f67baaad6fb180ac3491c3237d286bfdf88f4503175b8079b3" exitCode=0 Nov 28 13:22:52 crc kubenswrapper[4631]: I1128 13:22:52.315941 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4trst" event={"ID":"f6675fed-3a90-4779-9b95-1acba5248fd2","Type":"ContainerDied","Data":"ed26ab582f1984f67baaad6fb180ac3491c3237d286bfdf88f4503175b8079b3"} Nov 28 13:22:52 crc kubenswrapper[4631]: I1128 13:22:52.315981 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4trst" event={"ID":"f6675fed-3a90-4779-9b95-1acba5248fd2","Type":"ContainerDied","Data":"70d6a2a429d0a2a96315f02b1a28caede0861023f7293afcc5c58f2927d41a1c"} Nov 28 13:22:52 crc kubenswrapper[4631]: I1128 13:22:52.316082 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4trst" Nov 28 13:22:52 crc kubenswrapper[4631]: I1128 13:22:52.318699 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tpjm2" event={"ID":"2fdd01cb-8e67-4bc4-aa2b-e023882103ce","Type":"ContainerDied","Data":"380a08fd5f1ebb2bde91ea8aa5fea36fbeac716fb80ea119fefec1ea3997cea3"} Nov 28 13:22:52 crc kubenswrapper[4631]: I1128 13:22:52.318748 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-tpjm2" Nov 28 13:22:52 crc kubenswrapper[4631]: I1128 13:22:52.320247 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-6nrc5_72b60e37-5508-451f-8602-9dcebd55324e/extract-content/0.log" Nov 28 13:22:52 crc kubenswrapper[4631]: I1128 13:22:52.321018 4631 generic.go:334] "Generic (PLEG): container finished" podID="72b60e37-5508-451f-8602-9dcebd55324e" containerID="a93f839d3b8e8c46ac42d69047324d2174189284ad3257b716d7d644c9c703a3" exitCode=2 Nov 28 13:22:52 crc kubenswrapper[4631]: I1128 13:22:52.321163 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6nrc5" event={"ID":"72b60e37-5508-451f-8602-9dcebd55324e","Type":"ContainerDied","Data":"a93f839d3b8e8c46ac42d69047324d2174189284ad3257b716d7d644c9c703a3"} Nov 28 13:22:52 crc kubenswrapper[4631]: I1128 13:22:52.323010 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-p5v2z_67afe4ac-28a7-42a2-8155-2c4bf496b834/registry-server/0.log" Nov 28 13:22:52 crc kubenswrapper[4631]: I1128 13:22:52.323577 4631 generic.go:334] "Generic (PLEG): container finished" podID="67afe4ac-28a7-42a2-8155-2c4bf496b834" containerID="14e17a95f26adc182db616f80169b29b7bfc52ce3c1aa51772bbdac6228a6bb1" exitCode=1 Nov 28 13:22:52 crc kubenswrapper[4631]: I1128 13:22:52.323615 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-p5v2z" event={"ID":"67afe4ac-28a7-42a2-8155-2c4bf496b834","Type":"ContainerDied","Data":"14e17a95f26adc182db616f80169b29b7bfc52ce3c1aa51772bbdac6228a6bb1"} Nov 28 13:22:52 crc kubenswrapper[4631]: I1128 13:22:52.328437 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-2bc4b" Nov 28 13:22:52 crc kubenswrapper[4631]: I1128 13:22:52.329744 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-2bc4b" event={"ID":"9eb418ce-65b6-4da3-b282-831dd62916cf","Type":"ContainerDied","Data":"da066ebca14c9bee0f45b1f08ff7c3dc5c1383c4e540419c9c9b754a8dac7dcb"} Nov 28 13:22:52 crc kubenswrapper[4631]: I1128 13:22:52.329842 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-zkczn"] Nov 28 13:22:52 crc kubenswrapper[4631]: I1128 13:22:52.333546 4631 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-zkczn"] Nov 28 13:22:52 crc kubenswrapper[4631]: I1128 13:22:52.336319 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-gsrzv" Nov 28 13:22:52 crc kubenswrapper[4631]: I1128 13:22:52.371429 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-bm9gn"] Nov 28 13:22:52 crc kubenswrapper[4631]: I1128 13:22:52.374091 4631 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-bm9gn"] Nov 28 13:22:52 crc kubenswrapper[4631]: I1128 13:22:52.436737 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-tpjm2"] Nov 28 13:22:52 crc kubenswrapper[4631]: I1128 13:22:52.441338 4631 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-tpjm2"] Nov 28 13:22:52 crc kubenswrapper[4631]: I1128 13:22:52.445059 4631 scope.go:117] "RemoveContainer" containerID="503d4797bf5ffd1dcd5d9c15a9b8319e91b159e157b945d7d78caa6c21c89da4" Nov 28 13:22:52 crc kubenswrapper[4631]: I1128 13:22:52.482149 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-4trst"] Nov 28 13:22:52 crc kubenswrapper[4631]: I1128 13:22:52.482206 4631 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-4trst"] Nov 28 13:22:52 crc kubenswrapper[4631]: I1128 13:22:52.497895 4631 scope.go:117] "RemoveContainer" containerID="a844c003570d7e42551cdfb77ab5afad3faade0c2383aaa0bba7016962d0d765" Nov 28 13:22:52 crc kubenswrapper[4631]: I1128 13:22:52.498069 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-2bc4b"] Nov 28 13:22:52 crc kubenswrapper[4631]: I1128 13:22:52.503886 4631 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-2bc4b"] Nov 28 13:22:52 crc kubenswrapper[4631]: I1128 13:22:52.549630 4631 scope.go:117] "RemoveContainer" containerID="6874003d194b999f1822c5253b067abe1a2df13cd079b9f57b230e036928603e" Nov 28 13:22:52 crc kubenswrapper[4631]: I1128 13:22:52.570771 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-c7dxv" Nov 28 13:22:52 crc kubenswrapper[4631]: I1128 13:22:52.576443 4631 scope.go:117] "RemoveContainer" containerID="2fd8d8bd8d2e176ad906537a003ed1da101fffb8ff378cc8904b9d2a15d47d06" Nov 28 13:22:52 crc kubenswrapper[4631]: I1128 13:22:52.577738 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-6nrc5_72b60e37-5508-451f-8602-9dcebd55324e/extract-content/0.log" Nov 28 13:22:52 crc kubenswrapper[4631]: I1128 13:22:52.578410 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6nrc5" Nov 28 13:22:52 crc kubenswrapper[4631]: I1128 13:22:52.602933 4631 scope.go:117] "RemoveContainer" containerID="e0f5f5dc880cea08555d369e01e4490d9c81b57030e14ec6a4a3c934c8761bda" Nov 28 13:22:52 crc kubenswrapper[4631]: I1128 13:22:52.630983 4631 scope.go:117] "RemoveContainer" containerID="ed26ab582f1984f67baaad6fb180ac3491c3237d286bfdf88f4503175b8079b3" Nov 28 13:22:52 crc kubenswrapper[4631]: I1128 13:22:52.647380 4631 scope.go:117] "RemoveContainer" containerID="0a6ee6c8cd381cbe0f8320e8ebb7c0d4d336a59006118461ae8a6458498aa3d6" Nov 28 13:22:52 crc kubenswrapper[4631]: I1128 13:22:52.652275 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qpqxs\" (UniqueName: \"kubernetes.io/projected/72b60e37-5508-451f-8602-9dcebd55324e-kube-api-access-qpqxs\") pod \"72b60e37-5508-451f-8602-9dcebd55324e\" (UID: \"72b60e37-5508-451f-8602-9dcebd55324e\") " Nov 28 13:22:52 crc kubenswrapper[4631]: I1128 13:22:52.652334 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5xxgd\" (UniqueName: \"kubernetes.io/projected/45e56fbb-cf6f-4676-a78d-37c59b154fe7-kube-api-access-5xxgd\") pod \"45e56fbb-cf6f-4676-a78d-37c59b154fe7\" (UID: \"45e56fbb-cf6f-4676-a78d-37c59b154fe7\") " Nov 28 13:22:52 crc kubenswrapper[4631]: I1128 13:22:52.652392 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/45e56fbb-cf6f-4676-a78d-37c59b154fe7-catalog-content\") pod \"45e56fbb-cf6f-4676-a78d-37c59b154fe7\" (UID: \"45e56fbb-cf6f-4676-a78d-37c59b154fe7\") " Nov 28 13:22:52 crc kubenswrapper[4631]: I1128 13:22:52.652451 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/72b60e37-5508-451f-8602-9dcebd55324e-utilities\") pod \"72b60e37-5508-451f-8602-9dcebd55324e\" (UID: \"72b60e37-5508-451f-8602-9dcebd55324e\") " Nov 28 13:22:52 crc kubenswrapper[4631]: I1128 13:22:52.652491 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/45e56fbb-cf6f-4676-a78d-37c59b154fe7-utilities\") pod \"45e56fbb-cf6f-4676-a78d-37c59b154fe7\" (UID: \"45e56fbb-cf6f-4676-a78d-37c59b154fe7\") " Nov 28 13:22:52 crc kubenswrapper[4631]: I1128 13:22:52.652511 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/72b60e37-5508-451f-8602-9dcebd55324e-catalog-content\") pod \"72b60e37-5508-451f-8602-9dcebd55324e\" (UID: \"72b60e37-5508-451f-8602-9dcebd55324e\") " Nov 28 13:22:52 crc kubenswrapper[4631]: I1128 13:22:52.654533 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/72b60e37-5508-451f-8602-9dcebd55324e-utilities" (OuterVolumeSpecName: "utilities") pod "72b60e37-5508-451f-8602-9dcebd55324e" (UID: "72b60e37-5508-451f-8602-9dcebd55324e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 13:22:52 crc kubenswrapper[4631]: I1128 13:22:52.654746 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/45e56fbb-cf6f-4676-a78d-37c59b154fe7-utilities" (OuterVolumeSpecName: "utilities") pod "45e56fbb-cf6f-4676-a78d-37c59b154fe7" (UID: "45e56fbb-cf6f-4676-a78d-37c59b154fe7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 13:22:52 crc kubenswrapper[4631]: I1128 13:22:52.656021 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/72b60e37-5508-451f-8602-9dcebd55324e-kube-api-access-qpqxs" (OuterVolumeSpecName: "kube-api-access-qpqxs") pod "72b60e37-5508-451f-8602-9dcebd55324e" (UID: "72b60e37-5508-451f-8602-9dcebd55324e"). InnerVolumeSpecName "kube-api-access-qpqxs". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:22:52 crc kubenswrapper[4631]: I1128 13:22:52.657865 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/45e56fbb-cf6f-4676-a78d-37c59b154fe7-kube-api-access-5xxgd" (OuterVolumeSpecName: "kube-api-access-5xxgd") pod "45e56fbb-cf6f-4676-a78d-37c59b154fe7" (UID: "45e56fbb-cf6f-4676-a78d-37c59b154fe7"). InnerVolumeSpecName "kube-api-access-5xxgd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:22:52 crc kubenswrapper[4631]: I1128 13:22:52.667148 4631 scope.go:117] "RemoveContainer" containerID="ed26ab582f1984f67baaad6fb180ac3491c3237d286bfdf88f4503175b8079b3" Nov 28 13:22:52 crc kubenswrapper[4631]: E1128 13:22:52.668259 4631 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ed26ab582f1984f67baaad6fb180ac3491c3237d286bfdf88f4503175b8079b3\": container with ID starting with ed26ab582f1984f67baaad6fb180ac3491c3237d286bfdf88f4503175b8079b3 not found: ID does not exist" containerID="ed26ab582f1984f67baaad6fb180ac3491c3237d286bfdf88f4503175b8079b3" Nov 28 13:22:52 crc kubenswrapper[4631]: I1128 13:22:52.668324 4631 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ed26ab582f1984f67baaad6fb180ac3491c3237d286bfdf88f4503175b8079b3"} err="failed to get container status \"ed26ab582f1984f67baaad6fb180ac3491c3237d286bfdf88f4503175b8079b3\": rpc error: code = NotFound desc = could not find container \"ed26ab582f1984f67baaad6fb180ac3491c3237d286bfdf88f4503175b8079b3\": container with ID starting with ed26ab582f1984f67baaad6fb180ac3491c3237d286bfdf88f4503175b8079b3 not found: ID does not exist" Nov 28 13:22:52 crc kubenswrapper[4631]: I1128 13:22:52.668374 4631 scope.go:117] "RemoveContainer" containerID="0a6ee6c8cd381cbe0f8320e8ebb7c0d4d336a59006118461ae8a6458498aa3d6" Nov 28 13:22:52 crc kubenswrapper[4631]: E1128 13:22:52.668805 4631 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0a6ee6c8cd381cbe0f8320e8ebb7c0d4d336a59006118461ae8a6458498aa3d6\": container with ID starting with 0a6ee6c8cd381cbe0f8320e8ebb7c0d4d336a59006118461ae8a6458498aa3d6 not found: ID does not exist" containerID="0a6ee6c8cd381cbe0f8320e8ebb7c0d4d336a59006118461ae8a6458498aa3d6" Nov 28 13:22:52 crc kubenswrapper[4631]: I1128 13:22:52.668847 4631 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0a6ee6c8cd381cbe0f8320e8ebb7c0d4d336a59006118461ae8a6458498aa3d6"} err="failed to get container status \"0a6ee6c8cd381cbe0f8320e8ebb7c0d4d336a59006118461ae8a6458498aa3d6\": rpc error: code = NotFound desc = could not find container \"0a6ee6c8cd381cbe0f8320e8ebb7c0d4d336a59006118461ae8a6458498aa3d6\": container with ID starting with 0a6ee6c8cd381cbe0f8320e8ebb7c0d4d336a59006118461ae8a6458498aa3d6 not found: ID does not exist" Nov 28 13:22:52 crc kubenswrapper[4631]: I1128 13:22:52.668881 4631 scope.go:117] "RemoveContainer" containerID="33640802287f0cb9a012e6e16d2d519b5b585e7a36f6eed0417001361b83fffd" Nov 28 13:22:52 crc kubenswrapper[4631]: I1128 13:22:52.682156 4631 scope.go:117] "RemoveContainer" containerID="6768ec2f4e644e060c069b1397cccee1b4463bc5cdacc6f0f3e12822304061c7" Nov 28 13:22:52 crc kubenswrapper[4631]: I1128 13:22:52.693749 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/72b60e37-5508-451f-8602-9dcebd55324e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "72b60e37-5508-451f-8602-9dcebd55324e" (UID: "72b60e37-5508-451f-8602-9dcebd55324e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 13:22:52 crc kubenswrapper[4631]: I1128 13:22:52.697825 4631 scope.go:117] "RemoveContainer" containerID="b402075f9cf569ae47999d890b945c40cafe4f81f7a434335ab9de307a1e0b90" Nov 28 13:22:52 crc kubenswrapper[4631]: I1128 13:22:52.711508 4631 scope.go:117] "RemoveContainer" containerID="48793a8f9c2488bf82d5c7596352e34ede9baff8e45779b3bb05fecc4a16daef" Nov 28 13:22:52 crc kubenswrapper[4631]: I1128 13:22:52.754281 4631 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/45e56fbb-cf6f-4676-a78d-37c59b154fe7-utilities\") on node \"crc\" DevicePath \"\"" Nov 28 13:22:52 crc kubenswrapper[4631]: I1128 13:22:52.754337 4631 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/72b60e37-5508-451f-8602-9dcebd55324e-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 28 13:22:52 crc kubenswrapper[4631]: I1128 13:22:52.754348 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qpqxs\" (UniqueName: \"kubernetes.io/projected/72b60e37-5508-451f-8602-9dcebd55324e-kube-api-access-qpqxs\") on node \"crc\" DevicePath \"\"" Nov 28 13:22:52 crc kubenswrapper[4631]: I1128 13:22:52.754359 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5xxgd\" (UniqueName: \"kubernetes.io/projected/45e56fbb-cf6f-4676-a78d-37c59b154fe7-kube-api-access-5xxgd\") on node \"crc\" DevicePath \"\"" Nov 28 13:22:52 crc kubenswrapper[4631]: I1128 13:22:52.754369 4631 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/72b60e37-5508-451f-8602-9dcebd55324e-utilities\") on node \"crc\" DevicePath \"\"" Nov 28 13:22:52 crc kubenswrapper[4631]: I1128 13:22:52.794917 4631 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-2bc4b container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.42:8080/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Nov 28 13:22:52 crc kubenswrapper[4631]: I1128 13:22:52.795504 4631 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-2bc4b" podUID="9eb418ce-65b6-4da3-b282-831dd62916cf" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.42:8080/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 28 13:22:52 crc kubenswrapper[4631]: I1128 13:22:52.846804 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/45e56fbb-cf6f-4676-a78d-37c59b154fe7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "45e56fbb-cf6f-4676-a78d-37c59b154fe7" (UID: "45e56fbb-cf6f-4676-a78d-37c59b154fe7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 13:22:52 crc kubenswrapper[4631]: I1128 13:22:52.855649 4631 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/45e56fbb-cf6f-4676-a78d-37c59b154fe7-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 28 13:22:53 crc kubenswrapper[4631]: I1128 13:22:53.344869 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-c7dxv" event={"ID":"45e56fbb-cf6f-4676-a78d-37c59b154fe7","Type":"ContainerDied","Data":"d94683e58267a3244facde0e33a428fc34e8f3f8791a6bd1962a64379a31eeab"} Nov 28 13:22:53 crc kubenswrapper[4631]: I1128 13:22:53.344966 4631 scope.go:117] "RemoveContainer" containerID="c5aa3d11b88805706a6e474810030bc9826180026b5ec33fff46bb58cbc58e71" Nov 28 13:22:53 crc kubenswrapper[4631]: I1128 13:22:53.345680 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-c7dxv" Nov 28 13:22:53 crc kubenswrapper[4631]: I1128 13:22:53.352850 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-44v44_5e49b00c-84e0-4d9c-9095-daaeb2817505/registry-server/0.log" Nov 28 13:22:53 crc kubenswrapper[4631]: I1128 13:22:53.353498 4631 generic.go:334] "Generic (PLEG): container finished" podID="5e49b00c-84e0-4d9c-9095-daaeb2817505" containerID="20a58c60bdb60307ad0e09e4dde48caf094b28dc5fcdf9d45cc2b407486b3463" exitCode=1 Nov 28 13:22:53 crc kubenswrapper[4631]: I1128 13:22:53.353554 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-44v44" event={"ID":"5e49b00c-84e0-4d9c-9095-daaeb2817505","Type":"ContainerDied","Data":"20a58c60bdb60307ad0e09e4dde48caf094b28dc5fcdf9d45cc2b407486b3463"} Nov 28 13:22:53 crc kubenswrapper[4631]: I1128 13:22:53.355107 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-6nrc5_72b60e37-5508-451f-8602-9dcebd55324e/extract-content/0.log" Nov 28 13:22:53 crc kubenswrapper[4631]: I1128 13:22:53.357883 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6nrc5" Nov 28 13:22:53 crc kubenswrapper[4631]: I1128 13:22:53.358618 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6nrc5" event={"ID":"72b60e37-5508-451f-8602-9dcebd55324e","Type":"ContainerDied","Data":"98e79530ae258d5c9d111947af766c51808513037656c4bc1cb0801f7a69bbb3"} Nov 28 13:22:53 crc kubenswrapper[4631]: I1128 13:22:53.366701 4631 scope.go:117] "RemoveContainer" containerID="f1cce648b1c8f5ddfc84196fe92c332a6f6065ef54ab21417926937379fc26b4" Nov 28 13:22:53 crc kubenswrapper[4631]: I1128 13:22:53.391944 4631 scope.go:117] "RemoveContainer" containerID="08297c988d0d0bf16010b6eb5a860143dbfedbb495dc13c4f985424eb3ab5560" Nov 28 13:22:53 crc kubenswrapper[4631]: I1128 13:22:53.426375 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-6nrc5"] Nov 28 13:22:53 crc kubenswrapper[4631]: I1128 13:22:53.429971 4631 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-6nrc5"] Nov 28 13:22:53 crc kubenswrapper[4631]: I1128 13:22:53.432826 4631 scope.go:117] "RemoveContainer" containerID="a93f839d3b8e8c46ac42d69047324d2174189284ad3257b716d7d644c9c703a3" Nov 28 13:22:53 crc kubenswrapper[4631]: I1128 13:22:53.461681 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-c7dxv"] Nov 28 13:22:53 crc kubenswrapper[4631]: I1128 13:22:53.463594 4631 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-c7dxv"] Nov 28 13:22:53 crc kubenswrapper[4631]: I1128 13:22:53.482261 4631 scope.go:117] "RemoveContainer" containerID="e0cadea03b1d6a0341c5e2f626ac4163ccb8542fb464bf13aeffc51cc43f1ede" Nov 28 13:22:53 crc kubenswrapper[4631]: I1128 13:22:53.530177 4631 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="10dbdccc-7009-407e-846d-8a2976a58fad" path="/var/lib/kubelet/pods/10dbdccc-7009-407e-846d-8a2976a58fad/volumes" Nov 28 13:22:53 crc kubenswrapper[4631]: I1128 13:22:53.530952 4631 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2fdd01cb-8e67-4bc4-aa2b-e023882103ce" path="/var/lib/kubelet/pods/2fdd01cb-8e67-4bc4-aa2b-e023882103ce/volumes" Nov 28 13:22:53 crc kubenswrapper[4631]: I1128 13:22:53.532050 4631 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="45e56fbb-cf6f-4676-a78d-37c59b154fe7" path="/var/lib/kubelet/pods/45e56fbb-cf6f-4676-a78d-37c59b154fe7/volumes" Nov 28 13:22:53 crc kubenswrapper[4631]: I1128 13:22:53.534489 4631 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="68b72f7f-4534-480d-983e-04ea986e3b41" path="/var/lib/kubelet/pods/68b72f7f-4534-480d-983e-04ea986e3b41/volumes" Nov 28 13:22:53 crc kubenswrapper[4631]: I1128 13:22:53.535409 4631 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="72b60e37-5508-451f-8602-9dcebd55324e" path="/var/lib/kubelet/pods/72b60e37-5508-451f-8602-9dcebd55324e/volumes" Nov 28 13:22:53 crc kubenswrapper[4631]: I1128 13:22:53.540737 4631 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9eb418ce-65b6-4da3-b282-831dd62916cf" path="/var/lib/kubelet/pods/9eb418ce-65b6-4da3-b282-831dd62916cf/volumes" Nov 28 13:22:53 crc kubenswrapper[4631]: I1128 13:22:53.541435 4631 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f6675fed-3a90-4779-9b95-1acba5248fd2" path="/var/lib/kubelet/pods/f6675fed-3a90-4779-9b95-1acba5248fd2/volumes" Nov 28 13:22:53 crc kubenswrapper[4631]: I1128 13:22:53.660161 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-44v44_5e49b00c-84e0-4d9c-9095-daaeb2817505/registry-server/0.log" Nov 28 13:22:53 crc kubenswrapper[4631]: I1128 13:22:53.660867 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-44v44" Nov 28 13:22:53 crc kubenswrapper[4631]: I1128 13:22:53.699329 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-p5v2z_67afe4ac-28a7-42a2-8155-2c4bf496b834/registry-server/0.log" Nov 28 13:22:53 crc kubenswrapper[4631]: I1128 13:22:53.700824 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-p5v2z" Nov 28 13:22:53 crc kubenswrapper[4631]: I1128 13:22:53.768932 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9lkb4\" (UniqueName: \"kubernetes.io/projected/67afe4ac-28a7-42a2-8155-2c4bf496b834-kube-api-access-9lkb4\") pod \"67afe4ac-28a7-42a2-8155-2c4bf496b834\" (UID: \"67afe4ac-28a7-42a2-8155-2c4bf496b834\") " Nov 28 13:22:53 crc kubenswrapper[4631]: I1128 13:22:53.769185 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c59wr\" (UniqueName: \"kubernetes.io/projected/5e49b00c-84e0-4d9c-9095-daaeb2817505-kube-api-access-c59wr\") pod \"5e49b00c-84e0-4d9c-9095-daaeb2817505\" (UID: \"5e49b00c-84e0-4d9c-9095-daaeb2817505\") " Nov 28 13:22:53 crc kubenswrapper[4631]: I1128 13:22:53.769216 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5e49b00c-84e0-4d9c-9095-daaeb2817505-catalog-content\") pod \"5e49b00c-84e0-4d9c-9095-daaeb2817505\" (UID: \"5e49b00c-84e0-4d9c-9095-daaeb2817505\") " Nov 28 13:22:53 crc kubenswrapper[4631]: I1128 13:22:53.769255 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/67afe4ac-28a7-42a2-8155-2c4bf496b834-catalog-content\") pod \"67afe4ac-28a7-42a2-8155-2c4bf496b834\" (UID: \"67afe4ac-28a7-42a2-8155-2c4bf496b834\") " Nov 28 13:22:53 crc kubenswrapper[4631]: I1128 13:22:53.769297 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/67afe4ac-28a7-42a2-8155-2c4bf496b834-utilities\") pod \"67afe4ac-28a7-42a2-8155-2c4bf496b834\" (UID: \"67afe4ac-28a7-42a2-8155-2c4bf496b834\") " Nov 28 13:22:53 crc kubenswrapper[4631]: I1128 13:22:53.769336 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5e49b00c-84e0-4d9c-9095-daaeb2817505-utilities\") pod \"5e49b00c-84e0-4d9c-9095-daaeb2817505\" (UID: \"5e49b00c-84e0-4d9c-9095-daaeb2817505\") " Nov 28 13:22:53 crc kubenswrapper[4631]: I1128 13:22:53.770975 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5e49b00c-84e0-4d9c-9095-daaeb2817505-utilities" (OuterVolumeSpecName: "utilities") pod "5e49b00c-84e0-4d9c-9095-daaeb2817505" (UID: "5e49b00c-84e0-4d9c-9095-daaeb2817505"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 13:22:53 crc kubenswrapper[4631]: I1128 13:22:53.770989 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/67afe4ac-28a7-42a2-8155-2c4bf496b834-utilities" (OuterVolumeSpecName: "utilities") pod "67afe4ac-28a7-42a2-8155-2c4bf496b834" (UID: "67afe4ac-28a7-42a2-8155-2c4bf496b834"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 13:22:53 crc kubenswrapper[4631]: I1128 13:22:53.776474 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5e49b00c-84e0-4d9c-9095-daaeb2817505-kube-api-access-c59wr" (OuterVolumeSpecName: "kube-api-access-c59wr") pod "5e49b00c-84e0-4d9c-9095-daaeb2817505" (UID: "5e49b00c-84e0-4d9c-9095-daaeb2817505"). InnerVolumeSpecName "kube-api-access-c59wr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:22:53 crc kubenswrapper[4631]: I1128 13:22:53.777673 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/67afe4ac-28a7-42a2-8155-2c4bf496b834-kube-api-access-9lkb4" (OuterVolumeSpecName: "kube-api-access-9lkb4") pod "67afe4ac-28a7-42a2-8155-2c4bf496b834" (UID: "67afe4ac-28a7-42a2-8155-2c4bf496b834"). InnerVolumeSpecName "kube-api-access-9lkb4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:22:53 crc kubenswrapper[4631]: I1128 13:22:53.795797 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/67afe4ac-28a7-42a2-8155-2c4bf496b834-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "67afe4ac-28a7-42a2-8155-2c4bf496b834" (UID: "67afe4ac-28a7-42a2-8155-2c4bf496b834"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 13:22:53 crc kubenswrapper[4631]: I1128 13:22:53.871130 4631 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/67afe4ac-28a7-42a2-8155-2c4bf496b834-utilities\") on node \"crc\" DevicePath \"\"" Nov 28 13:22:53 crc kubenswrapper[4631]: I1128 13:22:53.871554 4631 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5e49b00c-84e0-4d9c-9095-daaeb2817505-utilities\") on node \"crc\" DevicePath \"\"" Nov 28 13:22:53 crc kubenswrapper[4631]: I1128 13:22:53.871656 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9lkb4\" (UniqueName: \"kubernetes.io/projected/67afe4ac-28a7-42a2-8155-2c4bf496b834-kube-api-access-9lkb4\") on node \"crc\" DevicePath \"\"" Nov 28 13:22:53 crc kubenswrapper[4631]: I1128 13:22:53.871784 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c59wr\" (UniqueName: \"kubernetes.io/projected/5e49b00c-84e0-4d9c-9095-daaeb2817505-kube-api-access-c59wr\") on node \"crc\" DevicePath \"\"" Nov 28 13:22:53 crc kubenswrapper[4631]: I1128 13:22:53.871873 4631 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/67afe4ac-28a7-42a2-8155-2c4bf496b834-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 28 13:22:53 crc kubenswrapper[4631]: I1128 13:22:53.911497 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5e49b00c-84e0-4d9c-9095-daaeb2817505-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5e49b00c-84e0-4d9c-9095-daaeb2817505" (UID: "5e49b00c-84e0-4d9c-9095-daaeb2817505"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 13:22:53 crc kubenswrapper[4631]: I1128 13:22:53.973349 4631 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5e49b00c-84e0-4d9c-9095-daaeb2817505-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 28 13:22:54 crc kubenswrapper[4631]: I1128 13:22:54.371887 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-44v44_5e49b00c-84e0-4d9c-9095-daaeb2817505/registry-server/0.log" Nov 28 13:22:54 crc kubenswrapper[4631]: I1128 13:22:54.375834 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-44v44" event={"ID":"5e49b00c-84e0-4d9c-9095-daaeb2817505","Type":"ContainerDied","Data":"734d2c4c8e528cfc2f685e98e28714535180b837ee540f3b2c1a8167d41db697"} Nov 28 13:22:54 crc kubenswrapper[4631]: I1128 13:22:54.375971 4631 scope.go:117] "RemoveContainer" containerID="20a58c60bdb60307ad0e09e4dde48caf094b28dc5fcdf9d45cc2b407486b3463" Nov 28 13:22:54 crc kubenswrapper[4631]: I1128 13:22:54.376229 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-44v44" Nov 28 13:22:54 crc kubenswrapper[4631]: I1128 13:22:54.384085 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-p5v2z_67afe4ac-28a7-42a2-8155-2c4bf496b834/registry-server/0.log" Nov 28 13:22:54 crc kubenswrapper[4631]: I1128 13:22:54.385824 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-p5v2z" event={"ID":"67afe4ac-28a7-42a2-8155-2c4bf496b834","Type":"ContainerDied","Data":"1db98735051371eafbcd998d22e47e1fcd13a954aad0552bb42a56fc5c2334f7"} Nov 28 13:22:54 crc kubenswrapper[4631]: I1128 13:22:54.386020 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-p5v2z" Nov 28 13:22:54 crc kubenswrapper[4631]: I1128 13:22:54.409281 4631 scope.go:117] "RemoveContainer" containerID="0658d484b316b0053beed4c248f90d5559d74314f1a55840f9cb998691d366f4" Nov 28 13:22:54 crc kubenswrapper[4631]: I1128 13:22:54.445845 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-44v44"] Nov 28 13:22:54 crc kubenswrapper[4631]: I1128 13:22:54.460685 4631 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-44v44"] Nov 28 13:22:54 crc kubenswrapper[4631]: I1128 13:22:54.467961 4631 scope.go:117] "RemoveContainer" containerID="5ab3a50a40a68a1e65aa2b887cc200aee402962b646b1d888af154557c276444" Nov 28 13:22:54 crc kubenswrapper[4631]: I1128 13:22:54.468655 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-p5v2z"] Nov 28 13:22:54 crc kubenswrapper[4631]: I1128 13:22:54.473952 4631 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-p5v2z"] Nov 28 13:22:54 crc kubenswrapper[4631]: I1128 13:22:54.506189 4631 scope.go:117] "RemoveContainer" containerID="14e17a95f26adc182db616f80169b29b7bfc52ce3c1aa51772bbdac6228a6bb1" Nov 28 13:22:54 crc kubenswrapper[4631]: I1128 13:22:54.524064 4631 scope.go:117] "RemoveContainer" containerID="dee2a84c2481360ae63392e497b8f84b25b55abe81bf3ce3655f99641cd8b6b9" Nov 28 13:22:54 crc kubenswrapper[4631]: I1128 13:22:54.546331 4631 scope.go:117] "RemoveContainer" containerID="cbda83ba8ca2ef82d49ef76e2f1fb00c3ea2e5bec77864f11e07a994b3d84fc8" Nov 28 13:22:55 crc kubenswrapper[4631]: I1128 13:22:55.356688 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-j6d86"] Nov 28 13:22:55 crc kubenswrapper[4631]: E1128 13:22:55.356922 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2fdd01cb-8e67-4bc4-aa2b-e023882103ce" containerName="extract-content" Nov 28 13:22:55 crc kubenswrapper[4631]: I1128 13:22:55.356938 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="2fdd01cb-8e67-4bc4-aa2b-e023882103ce" containerName="extract-content" Nov 28 13:22:55 crc kubenswrapper[4631]: E1128 13:22:55.356972 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="45e56fbb-cf6f-4676-a78d-37c59b154fe7" containerName="registry-server" Nov 28 13:22:55 crc kubenswrapper[4631]: I1128 13:22:55.356981 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="45e56fbb-cf6f-4676-a78d-37c59b154fe7" containerName="registry-server" Nov 28 13:22:55 crc kubenswrapper[4631]: E1128 13:22:55.356993 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2fdd01cb-8e67-4bc4-aa2b-e023882103ce" containerName="registry-server" Nov 28 13:22:55 crc kubenswrapper[4631]: I1128 13:22:55.357005 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="2fdd01cb-8e67-4bc4-aa2b-e023882103ce" containerName="registry-server" Nov 28 13:22:55 crc kubenswrapper[4631]: E1128 13:22:55.357018 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="45e56fbb-cf6f-4676-a78d-37c59b154fe7" containerName="extract-content" Nov 28 13:22:55 crc kubenswrapper[4631]: I1128 13:22:55.357026 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="45e56fbb-cf6f-4676-a78d-37c59b154fe7" containerName="extract-content" Nov 28 13:22:55 crc kubenswrapper[4631]: E1128 13:22:55.357038 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="67afe4ac-28a7-42a2-8155-2c4bf496b834" containerName="extract-content" Nov 28 13:22:55 crc kubenswrapper[4631]: I1128 13:22:55.357045 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="67afe4ac-28a7-42a2-8155-2c4bf496b834" containerName="extract-content" Nov 28 13:22:55 crc kubenswrapper[4631]: E1128 13:22:55.357056 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9eb418ce-65b6-4da3-b282-831dd62916cf" containerName="marketplace-operator" Nov 28 13:22:55 crc kubenswrapper[4631]: I1128 13:22:55.357065 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="9eb418ce-65b6-4da3-b282-831dd62916cf" containerName="marketplace-operator" Nov 28 13:22:55 crc kubenswrapper[4631]: E1128 13:22:55.357074 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="68b72f7f-4534-480d-983e-04ea986e3b41" containerName="registry-server" Nov 28 13:22:55 crc kubenswrapper[4631]: I1128 13:22:55.357081 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="68b72f7f-4534-480d-983e-04ea986e3b41" containerName="registry-server" Nov 28 13:22:55 crc kubenswrapper[4631]: E1128 13:22:55.357092 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="10dbdccc-7009-407e-846d-8a2976a58fad" containerName="extract-utilities" Nov 28 13:22:55 crc kubenswrapper[4631]: I1128 13:22:55.357100 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="10dbdccc-7009-407e-846d-8a2976a58fad" containerName="extract-utilities" Nov 28 13:22:55 crc kubenswrapper[4631]: E1128 13:22:55.357110 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="45e56fbb-cf6f-4676-a78d-37c59b154fe7" containerName="extract-utilities" Nov 28 13:22:55 crc kubenswrapper[4631]: I1128 13:22:55.357118 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="45e56fbb-cf6f-4676-a78d-37c59b154fe7" containerName="extract-utilities" Nov 28 13:22:55 crc kubenswrapper[4631]: E1128 13:22:55.357129 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="68b72f7f-4534-480d-983e-04ea986e3b41" containerName="extract-utilities" Nov 28 13:22:55 crc kubenswrapper[4631]: I1128 13:22:55.357138 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="68b72f7f-4534-480d-983e-04ea986e3b41" containerName="extract-utilities" Nov 28 13:22:55 crc kubenswrapper[4631]: E1128 13:22:55.357151 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="72b60e37-5508-451f-8602-9dcebd55324e" containerName="extract-utilities" Nov 28 13:22:55 crc kubenswrapper[4631]: I1128 13:22:55.357159 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="72b60e37-5508-451f-8602-9dcebd55324e" containerName="extract-utilities" Nov 28 13:22:55 crc kubenswrapper[4631]: E1128 13:22:55.357168 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="72b60e37-5508-451f-8602-9dcebd55324e" containerName="extract-content" Nov 28 13:22:55 crc kubenswrapper[4631]: I1128 13:22:55.357176 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="72b60e37-5508-451f-8602-9dcebd55324e" containerName="extract-content" Nov 28 13:22:55 crc kubenswrapper[4631]: E1128 13:22:55.357188 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="10dbdccc-7009-407e-846d-8a2976a58fad" containerName="registry-server" Nov 28 13:22:55 crc kubenswrapper[4631]: I1128 13:22:55.357195 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="10dbdccc-7009-407e-846d-8a2976a58fad" containerName="registry-server" Nov 28 13:22:55 crc kubenswrapper[4631]: E1128 13:22:55.357203 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5e49b00c-84e0-4d9c-9095-daaeb2817505" containerName="extract-content" Nov 28 13:22:55 crc kubenswrapper[4631]: I1128 13:22:55.357210 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="5e49b00c-84e0-4d9c-9095-daaeb2817505" containerName="extract-content" Nov 28 13:22:55 crc kubenswrapper[4631]: E1128 13:22:55.357219 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f6675fed-3a90-4779-9b95-1acba5248fd2" containerName="extract-content" Nov 28 13:22:55 crc kubenswrapper[4631]: I1128 13:22:55.357228 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="f6675fed-3a90-4779-9b95-1acba5248fd2" containerName="extract-content" Nov 28 13:22:55 crc kubenswrapper[4631]: E1128 13:22:55.357237 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5e49b00c-84e0-4d9c-9095-daaeb2817505" containerName="registry-server" Nov 28 13:22:55 crc kubenswrapper[4631]: I1128 13:22:55.357246 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="5e49b00c-84e0-4d9c-9095-daaeb2817505" containerName="registry-server" Nov 28 13:22:55 crc kubenswrapper[4631]: E1128 13:22:55.357257 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="67afe4ac-28a7-42a2-8155-2c4bf496b834" containerName="registry-server" Nov 28 13:22:55 crc kubenswrapper[4631]: I1128 13:22:55.357264 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="67afe4ac-28a7-42a2-8155-2c4bf496b834" containerName="registry-server" Nov 28 13:22:55 crc kubenswrapper[4631]: E1128 13:22:55.357274 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="68b72f7f-4534-480d-983e-04ea986e3b41" containerName="extract-content" Nov 28 13:22:55 crc kubenswrapper[4631]: I1128 13:22:55.357280 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="68b72f7f-4534-480d-983e-04ea986e3b41" containerName="extract-content" Nov 28 13:22:55 crc kubenswrapper[4631]: E1128 13:22:55.357310 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2fdd01cb-8e67-4bc4-aa2b-e023882103ce" containerName="extract-utilities" Nov 28 13:22:55 crc kubenswrapper[4631]: I1128 13:22:55.357317 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="2fdd01cb-8e67-4bc4-aa2b-e023882103ce" containerName="extract-utilities" Nov 28 13:22:55 crc kubenswrapper[4631]: E1128 13:22:55.357324 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="10dbdccc-7009-407e-846d-8a2976a58fad" containerName="extract-content" Nov 28 13:22:55 crc kubenswrapper[4631]: I1128 13:22:55.357331 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="10dbdccc-7009-407e-846d-8a2976a58fad" containerName="extract-content" Nov 28 13:22:55 crc kubenswrapper[4631]: E1128 13:22:55.357342 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5e49b00c-84e0-4d9c-9095-daaeb2817505" containerName="extract-utilities" Nov 28 13:22:55 crc kubenswrapper[4631]: I1128 13:22:55.357349 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="5e49b00c-84e0-4d9c-9095-daaeb2817505" containerName="extract-utilities" Nov 28 13:22:55 crc kubenswrapper[4631]: E1128 13:22:55.357360 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f6675fed-3a90-4779-9b95-1acba5248fd2" containerName="extract-utilities" Nov 28 13:22:55 crc kubenswrapper[4631]: I1128 13:22:55.357367 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="f6675fed-3a90-4779-9b95-1acba5248fd2" containerName="extract-utilities" Nov 28 13:22:55 crc kubenswrapper[4631]: E1128 13:22:55.357376 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="67afe4ac-28a7-42a2-8155-2c4bf496b834" containerName="extract-utilities" Nov 28 13:22:55 crc kubenswrapper[4631]: I1128 13:22:55.357383 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="67afe4ac-28a7-42a2-8155-2c4bf496b834" containerName="extract-utilities" Nov 28 13:22:55 crc kubenswrapper[4631]: I1128 13:22:55.357487 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="67afe4ac-28a7-42a2-8155-2c4bf496b834" containerName="registry-server" Nov 28 13:22:55 crc kubenswrapper[4631]: I1128 13:22:55.357496 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="5e49b00c-84e0-4d9c-9095-daaeb2817505" containerName="registry-server" Nov 28 13:22:55 crc kubenswrapper[4631]: I1128 13:22:55.357506 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="f6675fed-3a90-4779-9b95-1acba5248fd2" containerName="extract-content" Nov 28 13:22:55 crc kubenswrapper[4631]: I1128 13:22:55.357518 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="68b72f7f-4534-480d-983e-04ea986e3b41" containerName="registry-server" Nov 28 13:22:55 crc kubenswrapper[4631]: I1128 13:22:55.357527 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="10dbdccc-7009-407e-846d-8a2976a58fad" containerName="registry-server" Nov 28 13:22:55 crc kubenswrapper[4631]: I1128 13:22:55.357536 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="9eb418ce-65b6-4da3-b282-831dd62916cf" containerName="marketplace-operator" Nov 28 13:22:55 crc kubenswrapper[4631]: I1128 13:22:55.357545 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="72b60e37-5508-451f-8602-9dcebd55324e" containerName="extract-content" Nov 28 13:22:55 crc kubenswrapper[4631]: I1128 13:22:55.357556 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="2fdd01cb-8e67-4bc4-aa2b-e023882103ce" containerName="registry-server" Nov 28 13:22:55 crc kubenswrapper[4631]: I1128 13:22:55.357566 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="45e56fbb-cf6f-4676-a78d-37c59b154fe7" containerName="registry-server" Nov 28 13:22:55 crc kubenswrapper[4631]: I1128 13:22:55.358474 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-j6d86" Nov 28 13:22:55 crc kubenswrapper[4631]: I1128 13:22:55.361736 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Nov 28 13:22:55 crc kubenswrapper[4631]: I1128 13:22:55.373216 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-j6d86"] Nov 28 13:22:55 crc kubenswrapper[4631]: I1128 13:22:55.402006 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e55d89ce-6920-4cb5-951f-5b36cb73b6df-utilities\") pod \"certified-operators-j6d86\" (UID: \"e55d89ce-6920-4cb5-951f-5b36cb73b6df\") " pod="openshift-marketplace/certified-operators-j6d86" Nov 28 13:22:55 crc kubenswrapper[4631]: I1128 13:22:55.402396 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e55d89ce-6920-4cb5-951f-5b36cb73b6df-catalog-content\") pod \"certified-operators-j6d86\" (UID: \"e55d89ce-6920-4cb5-951f-5b36cb73b6df\") " pod="openshift-marketplace/certified-operators-j6d86" Nov 28 13:22:55 crc kubenswrapper[4631]: I1128 13:22:55.402448 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w8wm8\" (UniqueName: \"kubernetes.io/projected/e55d89ce-6920-4cb5-951f-5b36cb73b6df-kube-api-access-w8wm8\") pod \"certified-operators-j6d86\" (UID: \"e55d89ce-6920-4cb5-951f-5b36cb73b6df\") " pod="openshift-marketplace/certified-operators-j6d86" Nov 28 13:22:55 crc kubenswrapper[4631]: I1128 13:22:55.503698 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w8wm8\" (UniqueName: \"kubernetes.io/projected/e55d89ce-6920-4cb5-951f-5b36cb73b6df-kube-api-access-w8wm8\") pod \"certified-operators-j6d86\" (UID: \"e55d89ce-6920-4cb5-951f-5b36cb73b6df\") " pod="openshift-marketplace/certified-operators-j6d86" Nov 28 13:22:55 crc kubenswrapper[4631]: I1128 13:22:55.503834 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e55d89ce-6920-4cb5-951f-5b36cb73b6df-utilities\") pod \"certified-operators-j6d86\" (UID: \"e55d89ce-6920-4cb5-951f-5b36cb73b6df\") " pod="openshift-marketplace/certified-operators-j6d86" Nov 28 13:22:55 crc kubenswrapper[4631]: I1128 13:22:55.503862 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e55d89ce-6920-4cb5-951f-5b36cb73b6df-catalog-content\") pod \"certified-operators-j6d86\" (UID: \"e55d89ce-6920-4cb5-951f-5b36cb73b6df\") " pod="openshift-marketplace/certified-operators-j6d86" Nov 28 13:22:55 crc kubenswrapper[4631]: I1128 13:22:55.504308 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e55d89ce-6920-4cb5-951f-5b36cb73b6df-utilities\") pod \"certified-operators-j6d86\" (UID: \"e55d89ce-6920-4cb5-951f-5b36cb73b6df\") " pod="openshift-marketplace/certified-operators-j6d86" Nov 28 13:22:55 crc kubenswrapper[4631]: I1128 13:22:55.504489 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e55d89ce-6920-4cb5-951f-5b36cb73b6df-catalog-content\") pod \"certified-operators-j6d86\" (UID: \"e55d89ce-6920-4cb5-951f-5b36cb73b6df\") " pod="openshift-marketplace/certified-operators-j6d86" Nov 28 13:22:55 crc kubenswrapper[4631]: I1128 13:22:55.522773 4631 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5e49b00c-84e0-4d9c-9095-daaeb2817505" path="/var/lib/kubelet/pods/5e49b00c-84e0-4d9c-9095-daaeb2817505/volumes" Nov 28 13:22:55 crc kubenswrapper[4631]: I1128 13:22:55.523378 4631 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="67afe4ac-28a7-42a2-8155-2c4bf496b834" path="/var/lib/kubelet/pods/67afe4ac-28a7-42a2-8155-2c4bf496b834/volumes" Nov 28 13:22:55 crc kubenswrapper[4631]: I1128 13:22:55.529382 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w8wm8\" (UniqueName: \"kubernetes.io/projected/e55d89ce-6920-4cb5-951f-5b36cb73b6df-kube-api-access-w8wm8\") pod \"certified-operators-j6d86\" (UID: \"e55d89ce-6920-4cb5-951f-5b36cb73b6df\") " pod="openshift-marketplace/certified-operators-j6d86" Nov 28 13:22:55 crc kubenswrapper[4631]: I1128 13:22:55.683430 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-j6d86" Nov 28 13:22:56 crc kubenswrapper[4631]: I1128 13:22:56.130115 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-j6d86"] Nov 28 13:22:56 crc kubenswrapper[4631]: I1128 13:22:56.358275 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-7hvgh"] Nov 28 13:22:56 crc kubenswrapper[4631]: I1128 13:22:56.359763 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7hvgh" Nov 28 13:22:56 crc kubenswrapper[4631]: I1128 13:22:56.363507 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Nov 28 13:22:56 crc kubenswrapper[4631]: I1128 13:22:56.381357 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-7hvgh"] Nov 28 13:22:56 crc kubenswrapper[4631]: I1128 13:22:56.407805 4631 generic.go:334] "Generic (PLEG): container finished" podID="e55d89ce-6920-4cb5-951f-5b36cb73b6df" containerID="28e1b5bfd08f3699b9ccea95f82293a45eb357d4f5f079759f7d9317b0087988" exitCode=0 Nov 28 13:22:56 crc kubenswrapper[4631]: I1128 13:22:56.408096 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-j6d86" event={"ID":"e55d89ce-6920-4cb5-951f-5b36cb73b6df","Type":"ContainerDied","Data":"28e1b5bfd08f3699b9ccea95f82293a45eb357d4f5f079759f7d9317b0087988"} Nov 28 13:22:56 crc kubenswrapper[4631]: I1128 13:22:56.408336 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-j6d86" event={"ID":"e55d89ce-6920-4cb5-951f-5b36cb73b6df","Type":"ContainerStarted","Data":"89b23b1cf431c44dba24f085d59c592553c3429350295eed27045c7376f90c3a"} Nov 28 13:22:56 crc kubenswrapper[4631]: I1128 13:22:56.428973 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bbd0356c-2316-430e-81d3-c3c505c86e9e-utilities\") pod \"community-operators-7hvgh\" (UID: \"bbd0356c-2316-430e-81d3-c3c505c86e9e\") " pod="openshift-marketplace/community-operators-7hvgh" Nov 28 13:22:56 crc kubenswrapper[4631]: I1128 13:22:56.429045 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zmqxb\" (UniqueName: \"kubernetes.io/projected/bbd0356c-2316-430e-81d3-c3c505c86e9e-kube-api-access-zmqxb\") pod \"community-operators-7hvgh\" (UID: \"bbd0356c-2316-430e-81d3-c3c505c86e9e\") " pod="openshift-marketplace/community-operators-7hvgh" Nov 28 13:22:56 crc kubenswrapper[4631]: I1128 13:22:56.429076 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bbd0356c-2316-430e-81d3-c3c505c86e9e-catalog-content\") pod \"community-operators-7hvgh\" (UID: \"bbd0356c-2316-430e-81d3-c3c505c86e9e\") " pod="openshift-marketplace/community-operators-7hvgh" Nov 28 13:22:56 crc kubenswrapper[4631]: I1128 13:22:56.530626 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zmqxb\" (UniqueName: \"kubernetes.io/projected/bbd0356c-2316-430e-81d3-c3c505c86e9e-kube-api-access-zmqxb\") pod \"community-operators-7hvgh\" (UID: \"bbd0356c-2316-430e-81d3-c3c505c86e9e\") " pod="openshift-marketplace/community-operators-7hvgh" Nov 28 13:22:56 crc kubenswrapper[4631]: I1128 13:22:56.530932 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bbd0356c-2316-430e-81d3-c3c505c86e9e-catalog-content\") pod \"community-operators-7hvgh\" (UID: \"bbd0356c-2316-430e-81d3-c3c505c86e9e\") " pod="openshift-marketplace/community-operators-7hvgh" Nov 28 13:22:56 crc kubenswrapper[4631]: I1128 13:22:56.531140 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bbd0356c-2316-430e-81d3-c3c505c86e9e-utilities\") pod \"community-operators-7hvgh\" (UID: \"bbd0356c-2316-430e-81d3-c3c505c86e9e\") " pod="openshift-marketplace/community-operators-7hvgh" Nov 28 13:22:56 crc kubenswrapper[4631]: I1128 13:22:56.531657 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bbd0356c-2316-430e-81d3-c3c505c86e9e-utilities\") pod \"community-operators-7hvgh\" (UID: \"bbd0356c-2316-430e-81d3-c3c505c86e9e\") " pod="openshift-marketplace/community-operators-7hvgh" Nov 28 13:22:56 crc kubenswrapper[4631]: I1128 13:22:56.531914 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bbd0356c-2316-430e-81d3-c3c505c86e9e-catalog-content\") pod \"community-operators-7hvgh\" (UID: \"bbd0356c-2316-430e-81d3-c3c505c86e9e\") " pod="openshift-marketplace/community-operators-7hvgh" Nov 28 13:22:56 crc kubenswrapper[4631]: I1128 13:22:56.562265 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zmqxb\" (UniqueName: \"kubernetes.io/projected/bbd0356c-2316-430e-81d3-c3c505c86e9e-kube-api-access-zmqxb\") pod \"community-operators-7hvgh\" (UID: \"bbd0356c-2316-430e-81d3-c3c505c86e9e\") " pod="openshift-marketplace/community-operators-7hvgh" Nov 28 13:22:56 crc kubenswrapper[4631]: I1128 13:22:56.685374 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7hvgh" Nov 28 13:22:56 crc kubenswrapper[4631]: I1128 13:22:56.925695 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-7hvgh"] Nov 28 13:22:56 crc kubenswrapper[4631]: W1128 13:22:56.941595 4631 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbbd0356c_2316_430e_81d3_c3c505c86e9e.slice/crio-707e9f79fc9b462eba3187280a4765de1f42af1aecf49a608974c7bca1dc943b WatchSource:0}: Error finding container 707e9f79fc9b462eba3187280a4765de1f42af1aecf49a608974c7bca1dc943b: Status 404 returned error can't find the container with id 707e9f79fc9b462eba3187280a4765de1f42af1aecf49a608974c7bca1dc943b Nov 28 13:22:57 crc kubenswrapper[4631]: I1128 13:22:57.425984 4631 generic.go:334] "Generic (PLEG): container finished" podID="bbd0356c-2316-430e-81d3-c3c505c86e9e" containerID="2c42d4cb9403952c4857a675b3970db49a55658500e7987386a7449ad8f5d1ca" exitCode=0 Nov 28 13:22:57 crc kubenswrapper[4631]: I1128 13:22:57.426129 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7hvgh" event={"ID":"bbd0356c-2316-430e-81d3-c3c505c86e9e","Type":"ContainerDied","Data":"2c42d4cb9403952c4857a675b3970db49a55658500e7987386a7449ad8f5d1ca"} Nov 28 13:22:57 crc kubenswrapper[4631]: I1128 13:22:57.426475 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7hvgh" event={"ID":"bbd0356c-2316-430e-81d3-c3c505c86e9e","Type":"ContainerStarted","Data":"707e9f79fc9b462eba3187280a4765de1f42af1aecf49a608974c7bca1dc943b"} Nov 28 13:22:57 crc kubenswrapper[4631]: I1128 13:22:57.756199 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-mzsr9"] Nov 28 13:22:57 crc kubenswrapper[4631]: I1128 13:22:57.757241 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mzsr9" Nov 28 13:22:57 crc kubenswrapper[4631]: I1128 13:22:57.761632 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Nov 28 13:22:57 crc kubenswrapper[4631]: I1128 13:22:57.775698 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-mzsr9"] Nov 28 13:22:57 crc kubenswrapper[4631]: I1128 13:22:57.851344 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d2115a7c-dc8e-4f02-bded-25e080c7de51-catalog-content\") pod \"redhat-operators-mzsr9\" (UID: \"d2115a7c-dc8e-4f02-bded-25e080c7de51\") " pod="openshift-marketplace/redhat-operators-mzsr9" Nov 28 13:22:57 crc kubenswrapper[4631]: I1128 13:22:57.851412 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5t2c4\" (UniqueName: \"kubernetes.io/projected/d2115a7c-dc8e-4f02-bded-25e080c7de51-kube-api-access-5t2c4\") pod \"redhat-operators-mzsr9\" (UID: \"d2115a7c-dc8e-4f02-bded-25e080c7de51\") " pod="openshift-marketplace/redhat-operators-mzsr9" Nov 28 13:22:57 crc kubenswrapper[4631]: I1128 13:22:57.851454 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d2115a7c-dc8e-4f02-bded-25e080c7de51-utilities\") pod \"redhat-operators-mzsr9\" (UID: \"d2115a7c-dc8e-4f02-bded-25e080c7de51\") " pod="openshift-marketplace/redhat-operators-mzsr9" Nov 28 13:22:57 crc kubenswrapper[4631]: I1128 13:22:57.952667 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d2115a7c-dc8e-4f02-bded-25e080c7de51-catalog-content\") pod \"redhat-operators-mzsr9\" (UID: \"d2115a7c-dc8e-4f02-bded-25e080c7de51\") " pod="openshift-marketplace/redhat-operators-mzsr9" Nov 28 13:22:57 crc kubenswrapper[4631]: I1128 13:22:57.953143 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5t2c4\" (UniqueName: \"kubernetes.io/projected/d2115a7c-dc8e-4f02-bded-25e080c7de51-kube-api-access-5t2c4\") pod \"redhat-operators-mzsr9\" (UID: \"d2115a7c-dc8e-4f02-bded-25e080c7de51\") " pod="openshift-marketplace/redhat-operators-mzsr9" Nov 28 13:22:57 crc kubenswrapper[4631]: I1128 13:22:57.953170 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d2115a7c-dc8e-4f02-bded-25e080c7de51-utilities\") pod \"redhat-operators-mzsr9\" (UID: \"d2115a7c-dc8e-4f02-bded-25e080c7de51\") " pod="openshift-marketplace/redhat-operators-mzsr9" Nov 28 13:22:57 crc kubenswrapper[4631]: I1128 13:22:57.953716 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d2115a7c-dc8e-4f02-bded-25e080c7de51-utilities\") pod \"redhat-operators-mzsr9\" (UID: \"d2115a7c-dc8e-4f02-bded-25e080c7de51\") " pod="openshift-marketplace/redhat-operators-mzsr9" Nov 28 13:22:57 crc kubenswrapper[4631]: I1128 13:22:57.953969 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d2115a7c-dc8e-4f02-bded-25e080c7de51-catalog-content\") pod \"redhat-operators-mzsr9\" (UID: \"d2115a7c-dc8e-4f02-bded-25e080c7de51\") " pod="openshift-marketplace/redhat-operators-mzsr9" Nov 28 13:22:57 crc kubenswrapper[4631]: I1128 13:22:57.974483 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5t2c4\" (UniqueName: \"kubernetes.io/projected/d2115a7c-dc8e-4f02-bded-25e080c7de51-kube-api-access-5t2c4\") pod \"redhat-operators-mzsr9\" (UID: \"d2115a7c-dc8e-4f02-bded-25e080c7de51\") " pod="openshift-marketplace/redhat-operators-mzsr9" Nov 28 13:22:58 crc kubenswrapper[4631]: I1128 13:22:58.081692 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mzsr9" Nov 28 13:22:58 crc kubenswrapper[4631]: I1128 13:22:58.441094 4631 generic.go:334] "Generic (PLEG): container finished" podID="e55d89ce-6920-4cb5-951f-5b36cb73b6df" containerID="7e682c5ae6d59cdc8567819f7f55ff90c05902f477a5998ed7b2fe6480af3f31" exitCode=0 Nov 28 13:22:58 crc kubenswrapper[4631]: I1128 13:22:58.441200 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-j6d86" event={"ID":"e55d89ce-6920-4cb5-951f-5b36cb73b6df","Type":"ContainerDied","Data":"7e682c5ae6d59cdc8567819f7f55ff90c05902f477a5998ed7b2fe6480af3f31"} Nov 28 13:22:58 crc kubenswrapper[4631]: I1128 13:22:58.479183 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-mzsr9"] Nov 28 13:22:58 crc kubenswrapper[4631]: W1128 13:22:58.490419 4631 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd2115a7c_dc8e_4f02_bded_25e080c7de51.slice/crio-9475c8df03a1175e6b229f7a743f0cc6c8d9ff9a723a506fcd955ca5ea26a55c WatchSource:0}: Error finding container 9475c8df03a1175e6b229f7a743f0cc6c8d9ff9a723a506fcd955ca5ea26a55c: Status 404 returned error can't find the container with id 9475c8df03a1175e6b229f7a743f0cc6c8d9ff9a723a506fcd955ca5ea26a55c Nov 28 13:22:58 crc kubenswrapper[4631]: I1128 13:22:58.757341 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-fsz47"] Nov 28 13:22:58 crc kubenswrapper[4631]: I1128 13:22:58.760367 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fsz47" Nov 28 13:22:58 crc kubenswrapper[4631]: I1128 13:22:58.768595 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Nov 28 13:22:58 crc kubenswrapper[4631]: I1128 13:22:58.769389 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-fsz47"] Nov 28 13:22:58 crc kubenswrapper[4631]: I1128 13:22:58.873879 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4nwg8\" (UniqueName: \"kubernetes.io/projected/abdc286b-6b18-47b2-b974-410a0f9614b9-kube-api-access-4nwg8\") pod \"redhat-marketplace-fsz47\" (UID: \"abdc286b-6b18-47b2-b974-410a0f9614b9\") " pod="openshift-marketplace/redhat-marketplace-fsz47" Nov 28 13:22:58 crc kubenswrapper[4631]: I1128 13:22:58.874108 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/abdc286b-6b18-47b2-b974-410a0f9614b9-catalog-content\") pod \"redhat-marketplace-fsz47\" (UID: \"abdc286b-6b18-47b2-b974-410a0f9614b9\") " pod="openshift-marketplace/redhat-marketplace-fsz47" Nov 28 13:22:58 crc kubenswrapper[4631]: I1128 13:22:58.874166 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/abdc286b-6b18-47b2-b974-410a0f9614b9-utilities\") pod \"redhat-marketplace-fsz47\" (UID: \"abdc286b-6b18-47b2-b974-410a0f9614b9\") " pod="openshift-marketplace/redhat-marketplace-fsz47" Nov 28 13:22:58 crc kubenswrapper[4631]: I1128 13:22:58.974969 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/abdc286b-6b18-47b2-b974-410a0f9614b9-catalog-content\") pod \"redhat-marketplace-fsz47\" (UID: \"abdc286b-6b18-47b2-b974-410a0f9614b9\") " pod="openshift-marketplace/redhat-marketplace-fsz47" Nov 28 13:22:58 crc kubenswrapper[4631]: I1128 13:22:58.975026 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/abdc286b-6b18-47b2-b974-410a0f9614b9-utilities\") pod \"redhat-marketplace-fsz47\" (UID: \"abdc286b-6b18-47b2-b974-410a0f9614b9\") " pod="openshift-marketplace/redhat-marketplace-fsz47" Nov 28 13:22:58 crc kubenswrapper[4631]: I1128 13:22:58.975064 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4nwg8\" (UniqueName: \"kubernetes.io/projected/abdc286b-6b18-47b2-b974-410a0f9614b9-kube-api-access-4nwg8\") pod \"redhat-marketplace-fsz47\" (UID: \"abdc286b-6b18-47b2-b974-410a0f9614b9\") " pod="openshift-marketplace/redhat-marketplace-fsz47" Nov 28 13:22:58 crc kubenswrapper[4631]: I1128 13:22:58.975676 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/abdc286b-6b18-47b2-b974-410a0f9614b9-catalog-content\") pod \"redhat-marketplace-fsz47\" (UID: \"abdc286b-6b18-47b2-b974-410a0f9614b9\") " pod="openshift-marketplace/redhat-marketplace-fsz47" Nov 28 13:22:58 crc kubenswrapper[4631]: I1128 13:22:58.975876 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/abdc286b-6b18-47b2-b974-410a0f9614b9-utilities\") pod \"redhat-marketplace-fsz47\" (UID: \"abdc286b-6b18-47b2-b974-410a0f9614b9\") " pod="openshift-marketplace/redhat-marketplace-fsz47" Nov 28 13:22:59 crc kubenswrapper[4631]: I1128 13:22:59.002993 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4nwg8\" (UniqueName: \"kubernetes.io/projected/abdc286b-6b18-47b2-b974-410a0f9614b9-kube-api-access-4nwg8\") pod \"redhat-marketplace-fsz47\" (UID: \"abdc286b-6b18-47b2-b974-410a0f9614b9\") " pod="openshift-marketplace/redhat-marketplace-fsz47" Nov 28 13:22:59 crc kubenswrapper[4631]: I1128 13:22:59.094954 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fsz47" Nov 28 13:22:59 crc kubenswrapper[4631]: I1128 13:22:59.279824 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-fsz47"] Nov 28 13:22:59 crc kubenswrapper[4631]: I1128 13:22:59.375820 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-jhwlx" Nov 28 13:22:59 crc kubenswrapper[4631]: I1128 13:22:59.441744 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-5v5xp"] Nov 28 13:22:59 crc kubenswrapper[4631]: I1128 13:22:59.460026 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fsz47" event={"ID":"abdc286b-6b18-47b2-b974-410a0f9614b9","Type":"ContainerStarted","Data":"5208e399357f67a362a2b1701f29be70f8dc5568255d6ed3ac4331226385ac7c"} Nov 28 13:22:59 crc kubenswrapper[4631]: I1128 13:22:59.468432 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mzsr9" event={"ID":"d2115a7c-dc8e-4f02-bded-25e080c7de51","Type":"ContainerStarted","Data":"9475c8df03a1175e6b229f7a743f0cc6c8d9ff9a723a506fcd955ca5ea26a55c"} Nov 28 13:23:00 crc kubenswrapper[4631]: I1128 13:23:00.475947 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7hvgh" event={"ID":"bbd0356c-2316-430e-81d3-c3c505c86e9e","Type":"ContainerStarted","Data":"d462082d5e53c7db23bd88bd0d9940e1ee748592a211debaeb630ef5196e6b11"} Nov 28 13:23:00 crc kubenswrapper[4631]: I1128 13:23:00.477689 4631 generic.go:334] "Generic (PLEG): container finished" podID="abdc286b-6b18-47b2-b974-410a0f9614b9" containerID="823a788e7a5bdcd7650d1b6ffa34d187c6b9ccb32439705f274ff90c5c12a05f" exitCode=0 Nov 28 13:23:00 crc kubenswrapper[4631]: I1128 13:23:00.477749 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fsz47" event={"ID":"abdc286b-6b18-47b2-b974-410a0f9614b9","Type":"ContainerDied","Data":"823a788e7a5bdcd7650d1b6ffa34d187c6b9ccb32439705f274ff90c5c12a05f"} Nov 28 13:23:00 crc kubenswrapper[4631]: I1128 13:23:00.480912 4631 generic.go:334] "Generic (PLEG): container finished" podID="d2115a7c-dc8e-4f02-bded-25e080c7de51" containerID="d49b574c2f33ec5d516d03edd1a7d671a733107c9233ae71ca441468d4394bfd" exitCode=0 Nov 28 13:23:00 crc kubenswrapper[4631]: I1128 13:23:00.480944 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mzsr9" event={"ID":"d2115a7c-dc8e-4f02-bded-25e080c7de51","Type":"ContainerDied","Data":"d49b574c2f33ec5d516d03edd1a7d671a733107c9233ae71ca441468d4394bfd"} Nov 28 13:23:01 crc kubenswrapper[4631]: I1128 13:23:01.488316 4631 generic.go:334] "Generic (PLEG): container finished" podID="bbd0356c-2316-430e-81d3-c3c505c86e9e" containerID="d462082d5e53c7db23bd88bd0d9940e1ee748592a211debaeb630ef5196e6b11" exitCode=0 Nov 28 13:23:01 crc kubenswrapper[4631]: I1128 13:23:01.488482 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7hvgh" event={"ID":"bbd0356c-2316-430e-81d3-c3c505c86e9e","Type":"ContainerDied","Data":"d462082d5e53c7db23bd88bd0d9940e1ee748592a211debaeb630ef5196e6b11"} Nov 28 13:23:01 crc kubenswrapper[4631]: I1128 13:23:01.492592 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-j6d86" event={"ID":"e55d89ce-6920-4cb5-951f-5b36cb73b6df","Type":"ContainerStarted","Data":"bf3bde6fa4d0b3a9fa83477c1e20b76eb3779b1e7c943d18b7e4edb87cd5d0b5"} Nov 28 13:23:01 crc kubenswrapper[4631]: I1128 13:23:01.539346 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-j6d86" podStartSLOduration=2.414181735 podStartE2EDuration="6.539307217s" podCreationTimestamp="2025-11-28 13:22:55 +0000 UTC" firstStartedPulling="2025-11-28 13:22:56.417238693 +0000 UTC m=+133.224542037" lastFinishedPulling="2025-11-28 13:23:00.542364175 +0000 UTC m=+137.349667519" observedRunningTime="2025-11-28 13:23:01.53421499 +0000 UTC m=+138.341518364" watchObservedRunningTime="2025-11-28 13:23:01.539307217 +0000 UTC m=+138.346610571" Nov 28 13:23:02 crc kubenswrapper[4631]: I1128 13:23:02.501590 4631 generic.go:334] "Generic (PLEG): container finished" podID="abdc286b-6b18-47b2-b974-410a0f9614b9" containerID="2ae794398462d77b1afc6da5b339c6245e0e726238306b0912f974e85661cb4d" exitCode=0 Nov 28 13:23:02 crc kubenswrapper[4631]: I1128 13:23:02.501693 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fsz47" event={"ID":"abdc286b-6b18-47b2-b974-410a0f9614b9","Type":"ContainerDied","Data":"2ae794398462d77b1afc6da5b339c6245e0e726238306b0912f974e85661cb4d"} Nov 28 13:23:02 crc kubenswrapper[4631]: I1128 13:23:02.507011 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mzsr9" event={"ID":"d2115a7c-dc8e-4f02-bded-25e080c7de51","Type":"ContainerStarted","Data":"4be6b0bf13b88fa7541f49b6187e0b33f06f661f352aa70fe3a6c4e689fb7629"} Nov 28 13:23:03 crc kubenswrapper[4631]: I1128 13:23:03.513870 4631 generic.go:334] "Generic (PLEG): container finished" podID="d2115a7c-dc8e-4f02-bded-25e080c7de51" containerID="4be6b0bf13b88fa7541f49b6187e0b33f06f661f352aa70fe3a6c4e689fb7629" exitCode=0 Nov 28 13:23:03 crc kubenswrapper[4631]: I1128 13:23:03.526741 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mzsr9" event={"ID":"d2115a7c-dc8e-4f02-bded-25e080c7de51","Type":"ContainerDied","Data":"4be6b0bf13b88fa7541f49b6187e0b33f06f661f352aa70fe3a6c4e689fb7629"} Nov 28 13:23:04 crc kubenswrapper[4631]: I1128 13:23:04.521917 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fsz47" event={"ID":"abdc286b-6b18-47b2-b974-410a0f9614b9","Type":"ContainerStarted","Data":"d864c28ffdcde326ea9024a46431df986a02b3144275a9d304efc6c9f5bd40e3"} Nov 28 13:23:04 crc kubenswrapper[4631]: I1128 13:23:04.524406 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mzsr9" event={"ID":"d2115a7c-dc8e-4f02-bded-25e080c7de51","Type":"ContainerStarted","Data":"ed2c633244aba8c85446b271fb93f3717258f59b58df37cbfa56ed1683881f56"} Nov 28 13:23:04 crc kubenswrapper[4631]: I1128 13:23:04.527370 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7hvgh" event={"ID":"bbd0356c-2316-430e-81d3-c3c505c86e9e","Type":"ContainerStarted","Data":"6d9caa760e07eeae5b6f7a1342b4934ee6c5510f157be429c56a88a041858267"} Nov 28 13:23:04 crc kubenswrapper[4631]: I1128 13:23:04.543549 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-fsz47" podStartSLOduration=2.8579745130000003 podStartE2EDuration="6.543527855s" podCreationTimestamp="2025-11-28 13:22:58 +0000 UTC" firstStartedPulling="2025-11-28 13:23:00.479394678 +0000 UTC m=+137.286698022" lastFinishedPulling="2025-11-28 13:23:04.16494802 +0000 UTC m=+140.972251364" observedRunningTime="2025-11-28 13:23:04.540436511 +0000 UTC m=+141.347739845" watchObservedRunningTime="2025-11-28 13:23:04.543527855 +0000 UTC m=+141.350831199" Nov 28 13:23:04 crc kubenswrapper[4631]: I1128 13:23:04.586601 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-mzsr9" podStartSLOduration=3.829427581 podStartE2EDuration="7.586577005s" podCreationTimestamp="2025-11-28 13:22:57 +0000 UTC" firstStartedPulling="2025-11-28 13:23:00.481861673 +0000 UTC m=+137.289165017" lastFinishedPulling="2025-11-28 13:23:04.239011107 +0000 UTC m=+141.046314441" observedRunningTime="2025-11-28 13:23:04.583466312 +0000 UTC m=+141.390769656" watchObservedRunningTime="2025-11-28 13:23:04.586577005 +0000 UTC m=+141.393880349" Nov 28 13:23:04 crc kubenswrapper[4631]: I1128 13:23:04.588580 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-7hvgh" podStartSLOduration=2.284595477 podStartE2EDuration="8.588572539s" podCreationTimestamp="2025-11-28 13:22:56 +0000 UTC" firstStartedPulling="2025-11-28 13:22:57.433163897 +0000 UTC m=+134.240467241" lastFinishedPulling="2025-11-28 13:23:03.737140969 +0000 UTC m=+140.544444303" observedRunningTime="2025-11-28 13:23:04.568009665 +0000 UTC m=+141.375313009" watchObservedRunningTime="2025-11-28 13:23:04.588572539 +0000 UTC m=+141.395875883" Nov 28 13:23:05 crc kubenswrapper[4631]: I1128 13:23:05.683694 4631 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-j6d86" Nov 28 13:23:05 crc kubenswrapper[4631]: I1128 13:23:05.684910 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-j6d86" Nov 28 13:23:05 crc kubenswrapper[4631]: I1128 13:23:05.729747 4631 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-j6d86" Nov 28 13:23:06 crc kubenswrapper[4631]: I1128 13:23:06.324199 4631 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Nov 28 13:23:06 crc kubenswrapper[4631]: I1128 13:23:06.325230 4631 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Nov 28 13:23:06 crc kubenswrapper[4631]: I1128 13:23:06.325465 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 28 13:23:06 crc kubenswrapper[4631]: I1128 13:23:06.325680 4631 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" containerID="cri-o://c4f323ff537cd79159a51b592401e92fc06d54c490ff005486f27f41d42a35f2" gracePeriod=15 Nov 28 13:23:06 crc kubenswrapper[4631]: I1128 13:23:06.325718 4631 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" containerID="cri-o://eff71f12d67a15adda3183a742f2869b104d773904a6c5954d1ccb6a2c64dd71" gracePeriod=15 Nov 28 13:23:06 crc kubenswrapper[4631]: I1128 13:23:06.325716 4631 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" containerID="cri-o://8ac30b44d2fe8ddb4ede530d4b99dab4dc641a097e21d4ae69cfcbeddc38802d" gracePeriod=15 Nov 28 13:23:06 crc kubenswrapper[4631]: I1128 13:23:06.325750 4631 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" containerID="cri-o://53ea681ea681973dd5177ec7f2925ef48fec88ac004b41fca867ce671f5a76d1" gracePeriod=15 Nov 28 13:23:06 crc kubenswrapper[4631]: I1128 13:23:06.325781 4631 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" containerID="cri-o://6138e1e9121a843fd730def8c3f019dd68a6f38d963ddb00124f2c532762e012" gracePeriod=15 Nov 28 13:23:06 crc kubenswrapper[4631]: I1128 13:23:06.326679 4631 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Nov 28 13:23:06 crc kubenswrapper[4631]: E1128 13:23:06.326870 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Nov 28 13:23:06 crc kubenswrapper[4631]: I1128 13:23:06.326886 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Nov 28 13:23:06 crc kubenswrapper[4631]: E1128 13:23:06.326900 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Nov 28 13:23:06 crc kubenswrapper[4631]: I1128 13:23:06.326912 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Nov 28 13:23:06 crc kubenswrapper[4631]: E1128 13:23:06.326926 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Nov 28 13:23:06 crc kubenswrapper[4631]: I1128 13:23:06.326935 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Nov 28 13:23:06 crc kubenswrapper[4631]: E1128 13:23:06.326951 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Nov 28 13:23:06 crc kubenswrapper[4631]: I1128 13:23:06.326960 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Nov 28 13:23:06 crc kubenswrapper[4631]: E1128 13:23:06.326971 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Nov 28 13:23:06 crc kubenswrapper[4631]: I1128 13:23:06.326980 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Nov 28 13:23:06 crc kubenswrapper[4631]: E1128 13:23:06.326991 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Nov 28 13:23:06 crc kubenswrapper[4631]: I1128 13:23:06.326999 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Nov 28 13:23:06 crc kubenswrapper[4631]: I1128 13:23:06.327137 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Nov 28 13:23:06 crc kubenswrapper[4631]: I1128 13:23:06.327152 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Nov 28 13:23:06 crc kubenswrapper[4631]: I1128 13:23:06.327165 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Nov 28 13:23:06 crc kubenswrapper[4631]: I1128 13:23:06.327178 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Nov 28 13:23:06 crc kubenswrapper[4631]: I1128 13:23:06.327192 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Nov 28 13:23:06 crc kubenswrapper[4631]: I1128 13:23:06.377587 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Nov 28 13:23:06 crc kubenswrapper[4631]: I1128 13:23:06.389863 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 28 13:23:06 crc kubenswrapper[4631]: I1128 13:23:06.389914 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 28 13:23:06 crc kubenswrapper[4631]: I1128 13:23:06.389951 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 28 13:23:06 crc kubenswrapper[4631]: I1128 13:23:06.389974 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 28 13:23:06 crc kubenswrapper[4631]: I1128 13:23:06.390041 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 28 13:23:06 crc kubenswrapper[4631]: I1128 13:23:06.390074 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 13:23:06 crc kubenswrapper[4631]: I1128 13:23:06.390103 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 13:23:06 crc kubenswrapper[4631]: I1128 13:23:06.390190 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 13:23:06 crc kubenswrapper[4631]: I1128 13:23:06.491843 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 28 13:23:06 crc kubenswrapper[4631]: I1128 13:23:06.491898 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 28 13:23:06 crc kubenswrapper[4631]: I1128 13:23:06.491929 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 28 13:23:06 crc kubenswrapper[4631]: I1128 13:23:06.491951 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 28 13:23:06 crc kubenswrapper[4631]: I1128 13:23:06.491967 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 28 13:23:06 crc kubenswrapper[4631]: I1128 13:23:06.491987 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 13:23:06 crc kubenswrapper[4631]: I1128 13:23:06.492007 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 13:23:06 crc kubenswrapper[4631]: I1128 13:23:06.492034 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 13:23:06 crc kubenswrapper[4631]: I1128 13:23:06.492100 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 13:23:06 crc kubenswrapper[4631]: I1128 13:23:06.492139 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 28 13:23:06 crc kubenswrapper[4631]: I1128 13:23:06.492160 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 28 13:23:06 crc kubenswrapper[4631]: I1128 13:23:06.492182 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 28 13:23:06 crc kubenswrapper[4631]: I1128 13:23:06.492202 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 28 13:23:06 crc kubenswrapper[4631]: I1128 13:23:06.492222 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 28 13:23:06 crc kubenswrapper[4631]: I1128 13:23:06.492242 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 13:23:06 crc kubenswrapper[4631]: I1128 13:23:06.492258 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 13:23:06 crc kubenswrapper[4631]: I1128 13:23:06.609831 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-j6d86" Nov 28 13:23:06 crc kubenswrapper[4631]: I1128 13:23:06.610793 4631 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.194:6443: connect: connection refused" Nov 28 13:23:06 crc kubenswrapper[4631]: I1128 13:23:06.611095 4631 status_manager.go:851] "Failed to get status for pod" podUID="e55d89ce-6920-4cb5-951f-5b36cb73b6df" pod="openshift-marketplace/certified-operators-j6d86" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-j6d86\": dial tcp 38.102.83.194:6443: connect: connection refused" Nov 28 13:23:06 crc kubenswrapper[4631]: I1128 13:23:06.611405 4631 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.194:6443: connect: connection refused" Nov 28 13:23:06 crc kubenswrapper[4631]: I1128 13:23:06.674948 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 28 13:23:06 crc kubenswrapper[4631]: I1128 13:23:06.685993 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-7hvgh" Nov 28 13:23:06 crc kubenswrapper[4631]: I1128 13:23:06.689282 4631 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-7hvgh" Nov 28 13:23:06 crc kubenswrapper[4631]: E1128 13:23:06.717611 4631 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.194:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.187c2e69f998fb6c openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-11-28 13:23:06.716388204 +0000 UTC m=+143.523691548,LastTimestamp:2025-11-28 13:23:06.716388204 +0000 UTC m=+143.523691548,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Nov 28 13:23:06 crc kubenswrapper[4631]: I1128 13:23:06.731339 4631 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-7hvgh" Nov 28 13:23:06 crc kubenswrapper[4631]: I1128 13:23:06.732056 4631 status_manager.go:851] "Failed to get status for pod" podUID="bbd0356c-2316-430e-81d3-c3c505c86e9e" pod="openshift-marketplace/community-operators-7hvgh" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-7hvgh\": dial tcp 38.102.83.194:6443: connect: connection refused" Nov 28 13:23:06 crc kubenswrapper[4631]: I1128 13:23:06.732227 4631 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.194:6443: connect: connection refused" Nov 28 13:23:06 crc kubenswrapper[4631]: I1128 13:23:06.732410 4631 status_manager.go:851] "Failed to get status for pod" podUID="e55d89ce-6920-4cb5-951f-5b36cb73b6df" pod="openshift-marketplace/certified-operators-j6d86" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-j6d86\": dial tcp 38.102.83.194:6443: connect: connection refused" Nov 28 13:23:06 crc kubenswrapper[4631]: I1128 13:23:06.732559 4631 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.194:6443: connect: connection refused" Nov 28 13:23:07 crc kubenswrapper[4631]: I1128 13:23:07.548573 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Nov 28 13:23:07 crc kubenswrapper[4631]: I1128 13:23:07.549771 4631 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="6138e1e9121a843fd730def8c3f019dd68a6f38d963ddb00124f2c532762e012" exitCode=0 Nov 28 13:23:07 crc kubenswrapper[4631]: I1128 13:23:07.549806 4631 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="8ac30b44d2fe8ddb4ede530d4b99dab4dc641a097e21d4ae69cfcbeddc38802d" exitCode=0 Nov 28 13:23:07 crc kubenswrapper[4631]: I1128 13:23:07.549821 4631 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="eff71f12d67a15adda3183a742f2869b104d773904a6c5954d1ccb6a2c64dd71" exitCode=0 Nov 28 13:23:07 crc kubenswrapper[4631]: I1128 13:23:07.549830 4631 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="53ea681ea681973dd5177ec7f2925ef48fec88ac004b41fca867ce671f5a76d1" exitCode=2 Nov 28 13:23:07 crc kubenswrapper[4631]: I1128 13:23:07.552460 4631 generic.go:334] "Generic (PLEG): container finished" podID="3a23494b-5ae8-4ace-8763-b94ffe64c5a2" containerID="93f973696558d2b35823f54d37651dffd0dca0b0a270471899448eeae23ca4d2" exitCode=0 Nov 28 13:23:07 crc kubenswrapper[4631]: I1128 13:23:07.552569 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"3a23494b-5ae8-4ace-8763-b94ffe64c5a2","Type":"ContainerDied","Data":"93f973696558d2b35823f54d37651dffd0dca0b0a270471899448eeae23ca4d2"} Nov 28 13:23:07 crc kubenswrapper[4631]: I1128 13:23:07.553961 4631 status_manager.go:851] "Failed to get status for pod" podUID="3a23494b-5ae8-4ace-8763-b94ffe64c5a2" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.194:6443: connect: connection refused" Nov 28 13:23:07 crc kubenswrapper[4631]: I1128 13:23:07.556054 4631 status_manager.go:851] "Failed to get status for pod" podUID="bbd0356c-2316-430e-81d3-c3c505c86e9e" pod="openshift-marketplace/community-operators-7hvgh" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-7hvgh\": dial tcp 38.102.83.194:6443: connect: connection refused" Nov 28 13:23:07 crc kubenswrapper[4631]: I1128 13:23:07.556879 4631 status_manager.go:851] "Failed to get status for pod" podUID="e55d89ce-6920-4cb5-951f-5b36cb73b6df" pod="openshift-marketplace/certified-operators-j6d86" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-j6d86\": dial tcp 38.102.83.194:6443: connect: connection refused" Nov 28 13:23:07 crc kubenswrapper[4631]: I1128 13:23:07.557457 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"1c53926191bf67e68a0e0230740de428c280f79550fd964cf4bef72a592a0f57"} Nov 28 13:23:07 crc kubenswrapper[4631]: I1128 13:23:07.557513 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"df8a6455cd365b09cdc2fd7bcf56227708d1ec6845bb813aec03725198478e44"} Nov 28 13:23:07 crc kubenswrapper[4631]: I1128 13:23:07.557603 4631 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.194:6443: connect: connection refused" Nov 28 13:23:07 crc kubenswrapper[4631]: I1128 13:23:07.559463 4631 status_manager.go:851] "Failed to get status for pod" podUID="bbd0356c-2316-430e-81d3-c3c505c86e9e" pod="openshift-marketplace/community-operators-7hvgh" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-7hvgh\": dial tcp 38.102.83.194:6443: connect: connection refused" Nov 28 13:23:07 crc kubenswrapper[4631]: I1128 13:23:07.559989 4631 status_manager.go:851] "Failed to get status for pod" podUID="e55d89ce-6920-4cb5-951f-5b36cb73b6df" pod="openshift-marketplace/certified-operators-j6d86" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-j6d86\": dial tcp 38.102.83.194:6443: connect: connection refused" Nov 28 13:23:07 crc kubenswrapper[4631]: I1128 13:23:07.560378 4631 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.194:6443: connect: connection refused" Nov 28 13:23:07 crc kubenswrapper[4631]: I1128 13:23:07.560705 4631 status_manager.go:851] "Failed to get status for pod" podUID="3a23494b-5ae8-4ace-8763-b94ffe64c5a2" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.194:6443: connect: connection refused" Nov 28 13:23:08 crc kubenswrapper[4631]: I1128 13:23:08.081869 4631 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-mzsr9" Nov 28 13:23:08 crc kubenswrapper[4631]: I1128 13:23:08.081916 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-mzsr9" Nov 28 13:23:08 crc kubenswrapper[4631]: I1128 13:23:08.794125 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-7hvgh" Nov 28 13:23:08 crc kubenswrapper[4631]: I1128 13:23:08.795090 4631 status_manager.go:851] "Failed to get status for pod" podUID="e55d89ce-6920-4cb5-951f-5b36cb73b6df" pod="openshift-marketplace/certified-operators-j6d86" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-j6d86\": dial tcp 38.102.83.194:6443: connect: connection refused" Nov 28 13:23:08 crc kubenswrapper[4631]: I1128 13:23:08.795745 4631 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.194:6443: connect: connection refused" Nov 28 13:23:08 crc kubenswrapper[4631]: I1128 13:23:08.796227 4631 status_manager.go:851] "Failed to get status for pod" podUID="3a23494b-5ae8-4ace-8763-b94ffe64c5a2" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.194:6443: connect: connection refused" Nov 28 13:23:08 crc kubenswrapper[4631]: I1128 13:23:08.796756 4631 status_manager.go:851] "Failed to get status for pod" podUID="bbd0356c-2316-430e-81d3-c3c505c86e9e" pod="openshift-marketplace/community-operators-7hvgh" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-7hvgh\": dial tcp 38.102.83.194:6443: connect: connection refused" Nov 28 13:23:09 crc kubenswrapper[4631]: I1128 13:23:09.090466 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Nov 28 13:23:09 crc kubenswrapper[4631]: I1128 13:23:09.091701 4631 status_manager.go:851] "Failed to get status for pod" podUID="bbd0356c-2316-430e-81d3-c3c505c86e9e" pod="openshift-marketplace/community-operators-7hvgh" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-7hvgh\": dial tcp 38.102.83.194:6443: connect: connection refused" Nov 28 13:23:09 crc kubenswrapper[4631]: I1128 13:23:09.092512 4631 status_manager.go:851] "Failed to get status for pod" podUID="e55d89ce-6920-4cb5-951f-5b36cb73b6df" pod="openshift-marketplace/certified-operators-j6d86" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-j6d86\": dial tcp 38.102.83.194:6443: connect: connection refused" Nov 28 13:23:09 crc kubenswrapper[4631]: I1128 13:23:09.092726 4631 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.194:6443: connect: connection refused" Nov 28 13:23:09 crc kubenswrapper[4631]: I1128 13:23:09.092919 4631 status_manager.go:851] "Failed to get status for pod" podUID="3a23494b-5ae8-4ace-8763-b94ffe64c5a2" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.194:6443: connect: connection refused" Nov 28 13:23:09 crc kubenswrapper[4631]: I1128 13:23:09.096029 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-fsz47" Nov 28 13:23:09 crc kubenswrapper[4631]: I1128 13:23:09.096956 4631 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-fsz47" Nov 28 13:23:09 crc kubenswrapper[4631]: I1128 13:23:09.129414 4631 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-mzsr9" podUID="d2115a7c-dc8e-4f02-bded-25e080c7de51" containerName="registry-server" probeResult="failure" output=< Nov 28 13:23:09 crc kubenswrapper[4631]: timeout: failed to connect service ":50051" within 1s Nov 28 13:23:09 crc kubenswrapper[4631]: > Nov 28 13:23:09 crc kubenswrapper[4631]: I1128 13:23:09.138194 4631 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-fsz47" Nov 28 13:23:09 crc kubenswrapper[4631]: I1128 13:23:09.138813 4631 status_manager.go:851] "Failed to get status for pod" podUID="3a23494b-5ae8-4ace-8763-b94ffe64c5a2" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.194:6443: connect: connection refused" Nov 28 13:23:09 crc kubenswrapper[4631]: I1128 13:23:09.139009 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/3a23494b-5ae8-4ace-8763-b94ffe64c5a2-kubelet-dir\") pod \"3a23494b-5ae8-4ace-8763-b94ffe64c5a2\" (UID: \"3a23494b-5ae8-4ace-8763-b94ffe64c5a2\") " Nov 28 13:23:09 crc kubenswrapper[4631]: I1128 13:23:09.139047 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/3a23494b-5ae8-4ace-8763-b94ffe64c5a2-var-lock\") pod \"3a23494b-5ae8-4ace-8763-b94ffe64c5a2\" (UID: \"3a23494b-5ae8-4ace-8763-b94ffe64c5a2\") " Nov 28 13:23:09 crc kubenswrapper[4631]: I1128 13:23:09.139097 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3a23494b-5ae8-4ace-8763-b94ffe64c5a2-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "3a23494b-5ae8-4ace-8763-b94ffe64c5a2" (UID: "3a23494b-5ae8-4ace-8763-b94ffe64c5a2"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 13:23:09 crc kubenswrapper[4631]: I1128 13:23:09.139135 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3a23494b-5ae8-4ace-8763-b94ffe64c5a2-kube-api-access\") pod \"3a23494b-5ae8-4ace-8763-b94ffe64c5a2\" (UID: \"3a23494b-5ae8-4ace-8763-b94ffe64c5a2\") " Nov 28 13:23:09 crc kubenswrapper[4631]: I1128 13:23:09.139168 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3a23494b-5ae8-4ace-8763-b94ffe64c5a2-var-lock" (OuterVolumeSpecName: "var-lock") pod "3a23494b-5ae8-4ace-8763-b94ffe64c5a2" (UID: "3a23494b-5ae8-4ace-8763-b94ffe64c5a2"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 13:23:09 crc kubenswrapper[4631]: I1128 13:23:09.139315 4631 status_manager.go:851] "Failed to get status for pod" podUID="bbd0356c-2316-430e-81d3-c3c505c86e9e" pod="openshift-marketplace/community-operators-7hvgh" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-7hvgh\": dial tcp 38.102.83.194:6443: connect: connection refused" Nov 28 13:23:09 crc kubenswrapper[4631]: I1128 13:23:09.139437 4631 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/3a23494b-5ae8-4ace-8763-b94ffe64c5a2-kubelet-dir\") on node \"crc\" DevicePath \"\"" Nov 28 13:23:09 crc kubenswrapper[4631]: I1128 13:23:09.139454 4631 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/3a23494b-5ae8-4ace-8763-b94ffe64c5a2-var-lock\") on node \"crc\" DevicePath \"\"" Nov 28 13:23:09 crc kubenswrapper[4631]: I1128 13:23:09.140047 4631 status_manager.go:851] "Failed to get status for pod" podUID="e55d89ce-6920-4cb5-951f-5b36cb73b6df" pod="openshift-marketplace/certified-operators-j6d86" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-j6d86\": dial tcp 38.102.83.194:6443: connect: connection refused" Nov 28 13:23:09 crc kubenswrapper[4631]: I1128 13:23:09.140242 4631 status_manager.go:851] "Failed to get status for pod" podUID="abdc286b-6b18-47b2-b974-410a0f9614b9" pod="openshift-marketplace/redhat-marketplace-fsz47" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-fsz47\": dial tcp 38.102.83.194:6443: connect: connection refused" Nov 28 13:23:09 crc kubenswrapper[4631]: I1128 13:23:09.140463 4631 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.194:6443: connect: connection refused" Nov 28 13:23:09 crc kubenswrapper[4631]: I1128 13:23:09.149449 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3a23494b-5ae8-4ace-8763-b94ffe64c5a2-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "3a23494b-5ae8-4ace-8763-b94ffe64c5a2" (UID: "3a23494b-5ae8-4ace-8763-b94ffe64c5a2"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:23:09 crc kubenswrapper[4631]: I1128 13:23:09.240618 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3a23494b-5ae8-4ace-8763-b94ffe64c5a2-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 28 13:23:09 crc kubenswrapper[4631]: I1128 13:23:09.572424 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Nov 28 13:23:09 crc kubenswrapper[4631]: I1128 13:23:09.573032 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"3a23494b-5ae8-4ace-8763-b94ffe64c5a2","Type":"ContainerDied","Data":"6c5ff36ffbac32af2835a89b780efe3f57ac371f22108b849656f4e985416966"} Nov 28 13:23:09 crc kubenswrapper[4631]: I1128 13:23:09.573066 4631 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6c5ff36ffbac32af2835a89b780efe3f57ac371f22108b849656f4e985416966" Nov 28 13:23:09 crc kubenswrapper[4631]: I1128 13:23:09.575896 4631 status_manager.go:851] "Failed to get status for pod" podUID="e55d89ce-6920-4cb5-951f-5b36cb73b6df" pod="openshift-marketplace/certified-operators-j6d86" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-j6d86\": dial tcp 38.102.83.194:6443: connect: connection refused" Nov 28 13:23:09 crc kubenswrapper[4631]: I1128 13:23:09.576420 4631 status_manager.go:851] "Failed to get status for pod" podUID="abdc286b-6b18-47b2-b974-410a0f9614b9" pod="openshift-marketplace/redhat-marketplace-fsz47" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-fsz47\": dial tcp 38.102.83.194:6443: connect: connection refused" Nov 28 13:23:09 crc kubenswrapper[4631]: I1128 13:23:09.576841 4631 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.194:6443: connect: connection refused" Nov 28 13:23:09 crc kubenswrapper[4631]: I1128 13:23:09.577246 4631 status_manager.go:851] "Failed to get status for pod" podUID="3a23494b-5ae8-4ace-8763-b94ffe64c5a2" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.194:6443: connect: connection refused" Nov 28 13:23:09 crc kubenswrapper[4631]: I1128 13:23:09.577711 4631 status_manager.go:851] "Failed to get status for pod" podUID="bbd0356c-2316-430e-81d3-c3c505c86e9e" pod="openshift-marketplace/community-operators-7hvgh" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-7hvgh\": dial tcp 38.102.83.194:6443: connect: connection refused" Nov 28 13:23:09 crc kubenswrapper[4631]: I1128 13:23:09.628019 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-fsz47" Nov 28 13:23:09 crc kubenswrapper[4631]: I1128 13:23:09.628377 4631 status_manager.go:851] "Failed to get status for pod" podUID="e55d89ce-6920-4cb5-951f-5b36cb73b6df" pod="openshift-marketplace/certified-operators-j6d86" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-j6d86\": dial tcp 38.102.83.194:6443: connect: connection refused" Nov 28 13:23:09 crc kubenswrapper[4631]: I1128 13:23:09.628554 4631 status_manager.go:851] "Failed to get status for pod" podUID="abdc286b-6b18-47b2-b974-410a0f9614b9" pod="openshift-marketplace/redhat-marketplace-fsz47" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-fsz47\": dial tcp 38.102.83.194:6443: connect: connection refused" Nov 28 13:23:09 crc kubenswrapper[4631]: I1128 13:23:09.628884 4631 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.194:6443: connect: connection refused" Nov 28 13:23:09 crc kubenswrapper[4631]: I1128 13:23:09.629508 4631 status_manager.go:851] "Failed to get status for pod" podUID="3a23494b-5ae8-4ace-8763-b94ffe64c5a2" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.194:6443: connect: connection refused" Nov 28 13:23:09 crc kubenswrapper[4631]: I1128 13:23:09.629794 4631 status_manager.go:851] "Failed to get status for pod" podUID="bbd0356c-2316-430e-81d3-c3c505c86e9e" pod="openshift-marketplace/community-operators-7hvgh" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-7hvgh\": dial tcp 38.102.83.194:6443: connect: connection refused" Nov 28 13:23:11 crc kubenswrapper[4631]: E1128 13:23:11.274909 4631 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.194:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.187c2e69f998fb6c openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-11-28 13:23:06.716388204 +0000 UTC m=+143.523691548,LastTimestamp:2025-11-28 13:23:06.716388204 +0000 UTC m=+143.523691548,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Nov 28 13:23:11 crc kubenswrapper[4631]: I1128 13:23:11.590088 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Nov 28 13:23:11 crc kubenswrapper[4631]: I1128 13:23:11.592026 4631 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="c4f323ff537cd79159a51b592401e92fc06d54c490ff005486f27f41d42a35f2" exitCode=0 Nov 28 13:23:12 crc kubenswrapper[4631]: I1128 13:23:12.474521 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Nov 28 13:23:12 crc kubenswrapper[4631]: I1128 13:23:12.475918 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 13:23:12 crc kubenswrapper[4631]: I1128 13:23:12.476737 4631 status_manager.go:851] "Failed to get status for pod" podUID="bbd0356c-2316-430e-81d3-c3c505c86e9e" pod="openshift-marketplace/community-operators-7hvgh" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-7hvgh\": dial tcp 38.102.83.194:6443: connect: connection refused" Nov 28 13:23:12 crc kubenswrapper[4631]: I1128 13:23:12.477600 4631 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.194:6443: connect: connection refused" Nov 28 13:23:12 crc kubenswrapper[4631]: I1128 13:23:12.477939 4631 status_manager.go:851] "Failed to get status for pod" podUID="e55d89ce-6920-4cb5-951f-5b36cb73b6df" pod="openshift-marketplace/certified-operators-j6d86" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-j6d86\": dial tcp 38.102.83.194:6443: connect: connection refused" Nov 28 13:23:12 crc kubenswrapper[4631]: I1128 13:23:12.478209 4631 status_manager.go:851] "Failed to get status for pod" podUID="abdc286b-6b18-47b2-b974-410a0f9614b9" pod="openshift-marketplace/redhat-marketplace-fsz47" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-fsz47\": dial tcp 38.102.83.194:6443: connect: connection refused" Nov 28 13:23:12 crc kubenswrapper[4631]: I1128 13:23:12.478486 4631 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.194:6443: connect: connection refused" Nov 28 13:23:12 crc kubenswrapper[4631]: I1128 13:23:12.478804 4631 status_manager.go:851] "Failed to get status for pod" podUID="3a23494b-5ae8-4ace-8763-b94ffe64c5a2" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.194:6443: connect: connection refused" Nov 28 13:23:12 crc kubenswrapper[4631]: I1128 13:23:12.583783 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Nov 28 13:23:12 crc kubenswrapper[4631]: I1128 13:23:12.583905 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Nov 28 13:23:12 crc kubenswrapper[4631]: I1128 13:23:12.583934 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 13:23:12 crc kubenswrapper[4631]: I1128 13:23:12.584030 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Nov 28 13:23:12 crc kubenswrapper[4631]: I1128 13:23:12.584054 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir" (OuterVolumeSpecName: "cert-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "cert-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 13:23:12 crc kubenswrapper[4631]: I1128 13:23:12.584156 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 13:23:12 crc kubenswrapper[4631]: I1128 13:23:12.584468 4631 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") on node \"crc\" DevicePath \"\"" Nov 28 13:23:12 crc kubenswrapper[4631]: I1128 13:23:12.584498 4631 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") on node \"crc\" DevicePath \"\"" Nov 28 13:23:12 crc kubenswrapper[4631]: I1128 13:23:12.584511 4631 reconciler_common.go:293] "Volume detached for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") on node \"crc\" DevicePath \"\"" Nov 28 13:23:12 crc kubenswrapper[4631]: I1128 13:23:12.601935 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Nov 28 13:23:12 crc kubenswrapper[4631]: I1128 13:23:12.603064 4631 scope.go:117] "RemoveContainer" containerID="6138e1e9121a843fd730def8c3f019dd68a6f38d963ddb00124f2c532762e012" Nov 28 13:23:12 crc kubenswrapper[4631]: I1128 13:23:12.603307 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 13:23:12 crc kubenswrapper[4631]: I1128 13:23:12.621081 4631 status_manager.go:851] "Failed to get status for pod" podUID="bbd0356c-2316-430e-81d3-c3c505c86e9e" pod="openshift-marketplace/community-operators-7hvgh" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-7hvgh\": dial tcp 38.102.83.194:6443: connect: connection refused" Nov 28 13:23:12 crc kubenswrapper[4631]: I1128 13:23:12.621551 4631 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.194:6443: connect: connection refused" Nov 28 13:23:12 crc kubenswrapper[4631]: I1128 13:23:12.621703 4631 status_manager.go:851] "Failed to get status for pod" podUID="e55d89ce-6920-4cb5-951f-5b36cb73b6df" pod="openshift-marketplace/certified-operators-j6d86" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-j6d86\": dial tcp 38.102.83.194:6443: connect: connection refused" Nov 28 13:23:12 crc kubenswrapper[4631]: I1128 13:23:12.621851 4631 status_manager.go:851] "Failed to get status for pod" podUID="abdc286b-6b18-47b2-b974-410a0f9614b9" pod="openshift-marketplace/redhat-marketplace-fsz47" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-fsz47\": dial tcp 38.102.83.194:6443: connect: connection refused" Nov 28 13:23:12 crc kubenswrapper[4631]: I1128 13:23:12.622501 4631 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.194:6443: connect: connection refused" Nov 28 13:23:12 crc kubenswrapper[4631]: I1128 13:23:12.622670 4631 status_manager.go:851] "Failed to get status for pod" podUID="3a23494b-5ae8-4ace-8763-b94ffe64c5a2" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.194:6443: connect: connection refused" Nov 28 13:23:12 crc kubenswrapper[4631]: I1128 13:23:12.627906 4631 scope.go:117] "RemoveContainer" containerID="8ac30b44d2fe8ddb4ede530d4b99dab4dc641a097e21d4ae69cfcbeddc38802d" Nov 28 13:23:12 crc kubenswrapper[4631]: I1128 13:23:12.646843 4631 scope.go:117] "RemoveContainer" containerID="eff71f12d67a15adda3183a742f2869b104d773904a6c5954d1ccb6a2c64dd71" Nov 28 13:23:12 crc kubenswrapper[4631]: I1128 13:23:12.672816 4631 scope.go:117] "RemoveContainer" containerID="53ea681ea681973dd5177ec7f2925ef48fec88ac004b41fca867ce671f5a76d1" Nov 28 13:23:12 crc kubenswrapper[4631]: I1128 13:23:12.686169 4631 scope.go:117] "RemoveContainer" containerID="c4f323ff537cd79159a51b592401e92fc06d54c490ff005486f27f41d42a35f2" Nov 28 13:23:12 crc kubenswrapper[4631]: I1128 13:23:12.702122 4631 scope.go:117] "RemoveContainer" containerID="f0b61d56ed070810021d253875a7bb5b9cb94a3446c56c4a92fb5418022dd643" Nov 28 13:23:13 crc kubenswrapper[4631]: I1128 13:23:13.516633 4631 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.194:6443: connect: connection refused" Nov 28 13:23:13 crc kubenswrapper[4631]: I1128 13:23:13.517544 4631 status_manager.go:851] "Failed to get status for pod" podUID="e55d89ce-6920-4cb5-951f-5b36cb73b6df" pod="openshift-marketplace/certified-operators-j6d86" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-j6d86\": dial tcp 38.102.83.194:6443: connect: connection refused" Nov 28 13:23:13 crc kubenswrapper[4631]: I1128 13:23:13.518491 4631 status_manager.go:851] "Failed to get status for pod" podUID="abdc286b-6b18-47b2-b974-410a0f9614b9" pod="openshift-marketplace/redhat-marketplace-fsz47" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-fsz47\": dial tcp 38.102.83.194:6443: connect: connection refused" Nov 28 13:23:13 crc kubenswrapper[4631]: I1128 13:23:13.519397 4631 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.194:6443: connect: connection refused" Nov 28 13:23:13 crc kubenswrapper[4631]: I1128 13:23:13.519963 4631 status_manager.go:851] "Failed to get status for pod" podUID="3a23494b-5ae8-4ace-8763-b94ffe64c5a2" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.194:6443: connect: connection refused" Nov 28 13:23:13 crc kubenswrapper[4631]: I1128 13:23:13.520464 4631 status_manager.go:851] "Failed to get status for pod" podUID="bbd0356c-2316-430e-81d3-c3c505c86e9e" pod="openshift-marketplace/community-operators-7hvgh" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-7hvgh\": dial tcp 38.102.83.194:6443: connect: connection refused" Nov 28 13:23:13 crc kubenswrapper[4631]: I1128 13:23:13.525324 4631 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4b27818a5e8e43d0dc095d08835c792" path="/var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/volumes" Nov 28 13:23:13 crc kubenswrapper[4631]: I1128 13:23:13.786280 4631 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-hgsfk" podUID="145c4da2-bb50-4900-84d7-974b0cfaa4d8" containerName="oauth-openshift" containerID="cri-o://ebf95e54de61f41d4a27cf2da056c3eb14c5d892726916d1ad9ca6b1ed823a0f" gracePeriod=15 Nov 28 13:23:15 crc kubenswrapper[4631]: I1128 13:23:15.266490 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-hgsfk" Nov 28 13:23:15 crc kubenswrapper[4631]: I1128 13:23:15.267347 4631 status_manager.go:851] "Failed to get status for pod" podUID="e55d89ce-6920-4cb5-951f-5b36cb73b6df" pod="openshift-marketplace/certified-operators-j6d86" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-j6d86\": dial tcp 38.102.83.194:6443: connect: connection refused" Nov 28 13:23:15 crc kubenswrapper[4631]: I1128 13:23:15.267511 4631 status_manager.go:851] "Failed to get status for pod" podUID="abdc286b-6b18-47b2-b974-410a0f9614b9" pod="openshift-marketplace/redhat-marketplace-fsz47" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-fsz47\": dial tcp 38.102.83.194:6443: connect: connection refused" Nov 28 13:23:15 crc kubenswrapper[4631]: I1128 13:23:15.267757 4631 status_manager.go:851] "Failed to get status for pod" podUID="145c4da2-bb50-4900-84d7-974b0cfaa4d8" pod="openshift-authentication/oauth-openshift-558db77b4-hgsfk" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-hgsfk\": dial tcp 38.102.83.194:6443: connect: connection refused" Nov 28 13:23:15 crc kubenswrapper[4631]: I1128 13:23:15.268057 4631 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.194:6443: connect: connection refused" Nov 28 13:23:15 crc kubenswrapper[4631]: I1128 13:23:15.268202 4631 status_manager.go:851] "Failed to get status for pod" podUID="3a23494b-5ae8-4ace-8763-b94ffe64c5a2" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.194:6443: connect: connection refused" Nov 28 13:23:15 crc kubenswrapper[4631]: I1128 13:23:15.268376 4631 status_manager.go:851] "Failed to get status for pod" podUID="bbd0356c-2316-430e-81d3-c3c505c86e9e" pod="openshift-marketplace/community-operators-7hvgh" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-7hvgh\": dial tcp 38.102.83.194:6443: connect: connection refused" Nov 28 13:23:15 crc kubenswrapper[4631]: I1128 13:23:15.323741 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/145c4da2-bb50-4900-84d7-974b0cfaa4d8-v4-0-config-user-idp-0-file-data\") pod \"145c4da2-bb50-4900-84d7-974b0cfaa4d8\" (UID: \"145c4da2-bb50-4900-84d7-974b0cfaa4d8\") " Nov 28 13:23:15 crc kubenswrapper[4631]: I1128 13:23:15.323790 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/145c4da2-bb50-4900-84d7-974b0cfaa4d8-v4-0-config-system-serving-cert\") pod \"145c4da2-bb50-4900-84d7-974b0cfaa4d8\" (UID: \"145c4da2-bb50-4900-84d7-974b0cfaa4d8\") " Nov 28 13:23:15 crc kubenswrapper[4631]: I1128 13:23:15.323810 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/145c4da2-bb50-4900-84d7-974b0cfaa4d8-audit-policies\") pod \"145c4da2-bb50-4900-84d7-974b0cfaa4d8\" (UID: \"145c4da2-bb50-4900-84d7-974b0cfaa4d8\") " Nov 28 13:23:15 crc kubenswrapper[4631]: I1128 13:23:15.323853 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/145c4da2-bb50-4900-84d7-974b0cfaa4d8-v4-0-config-user-template-provider-selection\") pod \"145c4da2-bb50-4900-84d7-974b0cfaa4d8\" (UID: \"145c4da2-bb50-4900-84d7-974b0cfaa4d8\") " Nov 28 13:23:15 crc kubenswrapper[4631]: I1128 13:23:15.323873 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/145c4da2-bb50-4900-84d7-974b0cfaa4d8-v4-0-config-system-ocp-branding-template\") pod \"145c4da2-bb50-4900-84d7-974b0cfaa4d8\" (UID: \"145c4da2-bb50-4900-84d7-974b0cfaa4d8\") " Nov 28 13:23:15 crc kubenswrapper[4631]: I1128 13:23:15.323900 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/145c4da2-bb50-4900-84d7-974b0cfaa4d8-v4-0-config-system-session\") pod \"145c4da2-bb50-4900-84d7-974b0cfaa4d8\" (UID: \"145c4da2-bb50-4900-84d7-974b0cfaa4d8\") " Nov 28 13:23:15 crc kubenswrapper[4631]: I1128 13:23:15.323919 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2l8sz\" (UniqueName: \"kubernetes.io/projected/145c4da2-bb50-4900-84d7-974b0cfaa4d8-kube-api-access-2l8sz\") pod \"145c4da2-bb50-4900-84d7-974b0cfaa4d8\" (UID: \"145c4da2-bb50-4900-84d7-974b0cfaa4d8\") " Nov 28 13:23:15 crc kubenswrapper[4631]: I1128 13:23:15.323936 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/145c4da2-bb50-4900-84d7-974b0cfaa4d8-audit-dir\") pod \"145c4da2-bb50-4900-84d7-974b0cfaa4d8\" (UID: \"145c4da2-bb50-4900-84d7-974b0cfaa4d8\") " Nov 28 13:23:15 crc kubenswrapper[4631]: I1128 13:23:15.323971 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/145c4da2-bb50-4900-84d7-974b0cfaa4d8-v4-0-config-user-template-login\") pod \"145c4da2-bb50-4900-84d7-974b0cfaa4d8\" (UID: \"145c4da2-bb50-4900-84d7-974b0cfaa4d8\") " Nov 28 13:23:15 crc kubenswrapper[4631]: I1128 13:23:15.323991 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/145c4da2-bb50-4900-84d7-974b0cfaa4d8-v4-0-config-system-cliconfig\") pod \"145c4da2-bb50-4900-84d7-974b0cfaa4d8\" (UID: \"145c4da2-bb50-4900-84d7-974b0cfaa4d8\") " Nov 28 13:23:15 crc kubenswrapper[4631]: I1128 13:23:15.324029 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/145c4da2-bb50-4900-84d7-974b0cfaa4d8-v4-0-config-system-router-certs\") pod \"145c4da2-bb50-4900-84d7-974b0cfaa4d8\" (UID: \"145c4da2-bb50-4900-84d7-974b0cfaa4d8\") " Nov 28 13:23:15 crc kubenswrapper[4631]: I1128 13:23:15.324045 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/145c4da2-bb50-4900-84d7-974b0cfaa4d8-v4-0-config-system-service-ca\") pod \"145c4da2-bb50-4900-84d7-974b0cfaa4d8\" (UID: \"145c4da2-bb50-4900-84d7-974b0cfaa4d8\") " Nov 28 13:23:15 crc kubenswrapper[4631]: I1128 13:23:15.324062 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/145c4da2-bb50-4900-84d7-974b0cfaa4d8-v4-0-config-system-trusted-ca-bundle\") pod \"145c4da2-bb50-4900-84d7-974b0cfaa4d8\" (UID: \"145c4da2-bb50-4900-84d7-974b0cfaa4d8\") " Nov 28 13:23:15 crc kubenswrapper[4631]: I1128 13:23:15.324080 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/145c4da2-bb50-4900-84d7-974b0cfaa4d8-v4-0-config-user-template-error\") pod \"145c4da2-bb50-4900-84d7-974b0cfaa4d8\" (UID: \"145c4da2-bb50-4900-84d7-974b0cfaa4d8\") " Nov 28 13:23:15 crc kubenswrapper[4631]: I1128 13:23:15.324958 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/145c4da2-bb50-4900-84d7-974b0cfaa4d8-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "145c4da2-bb50-4900-84d7-974b0cfaa4d8" (UID: "145c4da2-bb50-4900-84d7-974b0cfaa4d8"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 13:23:15 crc kubenswrapper[4631]: I1128 13:23:15.326025 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/145c4da2-bb50-4900-84d7-974b0cfaa4d8-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "145c4da2-bb50-4900-84d7-974b0cfaa4d8" (UID: "145c4da2-bb50-4900-84d7-974b0cfaa4d8"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 13:23:15 crc kubenswrapper[4631]: I1128 13:23:15.326254 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/145c4da2-bb50-4900-84d7-974b0cfaa4d8-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "145c4da2-bb50-4900-84d7-974b0cfaa4d8" (UID: "145c4da2-bb50-4900-84d7-974b0cfaa4d8"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 13:23:15 crc kubenswrapper[4631]: I1128 13:23:15.326311 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/145c4da2-bb50-4900-84d7-974b0cfaa4d8-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "145c4da2-bb50-4900-84d7-974b0cfaa4d8" (UID: "145c4da2-bb50-4900-84d7-974b0cfaa4d8"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 13:23:15 crc kubenswrapper[4631]: I1128 13:23:15.330951 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/145c4da2-bb50-4900-84d7-974b0cfaa4d8-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "145c4da2-bb50-4900-84d7-974b0cfaa4d8" (UID: "145c4da2-bb50-4900-84d7-974b0cfaa4d8"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:23:15 crc kubenswrapper[4631]: I1128 13:23:15.331511 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/145c4da2-bb50-4900-84d7-974b0cfaa4d8-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "145c4da2-bb50-4900-84d7-974b0cfaa4d8" (UID: "145c4da2-bb50-4900-84d7-974b0cfaa4d8"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:23:15 crc kubenswrapper[4631]: I1128 13:23:15.331560 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/145c4da2-bb50-4900-84d7-974b0cfaa4d8-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "145c4da2-bb50-4900-84d7-974b0cfaa4d8" (UID: "145c4da2-bb50-4900-84d7-974b0cfaa4d8"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 13:23:15 crc kubenswrapper[4631]: I1128 13:23:15.331849 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/145c4da2-bb50-4900-84d7-974b0cfaa4d8-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "145c4da2-bb50-4900-84d7-974b0cfaa4d8" (UID: "145c4da2-bb50-4900-84d7-974b0cfaa4d8"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:23:15 crc kubenswrapper[4631]: I1128 13:23:15.332592 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/145c4da2-bb50-4900-84d7-974b0cfaa4d8-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "145c4da2-bb50-4900-84d7-974b0cfaa4d8" (UID: "145c4da2-bb50-4900-84d7-974b0cfaa4d8"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:23:15 crc kubenswrapper[4631]: I1128 13:23:15.332794 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/145c4da2-bb50-4900-84d7-974b0cfaa4d8-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "145c4da2-bb50-4900-84d7-974b0cfaa4d8" (UID: "145c4da2-bb50-4900-84d7-974b0cfaa4d8"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:23:15 crc kubenswrapper[4631]: I1128 13:23:15.335979 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/145c4da2-bb50-4900-84d7-974b0cfaa4d8-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "145c4da2-bb50-4900-84d7-974b0cfaa4d8" (UID: "145c4da2-bb50-4900-84d7-974b0cfaa4d8"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:23:15 crc kubenswrapper[4631]: I1128 13:23:15.336271 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/145c4da2-bb50-4900-84d7-974b0cfaa4d8-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "145c4da2-bb50-4900-84d7-974b0cfaa4d8" (UID: "145c4da2-bb50-4900-84d7-974b0cfaa4d8"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:23:15 crc kubenswrapper[4631]: I1128 13:23:15.337655 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/145c4da2-bb50-4900-84d7-974b0cfaa4d8-kube-api-access-2l8sz" (OuterVolumeSpecName: "kube-api-access-2l8sz") pod "145c4da2-bb50-4900-84d7-974b0cfaa4d8" (UID: "145c4da2-bb50-4900-84d7-974b0cfaa4d8"). InnerVolumeSpecName "kube-api-access-2l8sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:23:15 crc kubenswrapper[4631]: I1128 13:23:15.341405 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/145c4da2-bb50-4900-84d7-974b0cfaa4d8-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "145c4da2-bb50-4900-84d7-974b0cfaa4d8" (UID: "145c4da2-bb50-4900-84d7-974b0cfaa4d8"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:23:15 crc kubenswrapper[4631]: I1128 13:23:15.425727 4631 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/145c4da2-bb50-4900-84d7-974b0cfaa4d8-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Nov 28 13:23:15 crc kubenswrapper[4631]: I1128 13:23:15.425759 4631 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/145c4da2-bb50-4900-84d7-974b0cfaa4d8-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Nov 28 13:23:15 crc kubenswrapper[4631]: I1128 13:23:15.425774 4631 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/145c4da2-bb50-4900-84d7-974b0cfaa4d8-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Nov 28 13:23:15 crc kubenswrapper[4631]: I1128 13:23:15.425785 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2l8sz\" (UniqueName: \"kubernetes.io/projected/145c4da2-bb50-4900-84d7-974b0cfaa4d8-kube-api-access-2l8sz\") on node \"crc\" DevicePath \"\"" Nov 28 13:23:15 crc kubenswrapper[4631]: I1128 13:23:15.425796 4631 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/145c4da2-bb50-4900-84d7-974b0cfaa4d8-audit-dir\") on node \"crc\" DevicePath \"\"" Nov 28 13:23:15 crc kubenswrapper[4631]: I1128 13:23:15.425808 4631 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/145c4da2-bb50-4900-84d7-974b0cfaa4d8-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Nov 28 13:23:15 crc kubenswrapper[4631]: I1128 13:23:15.425817 4631 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/145c4da2-bb50-4900-84d7-974b0cfaa4d8-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Nov 28 13:23:15 crc kubenswrapper[4631]: I1128 13:23:15.425826 4631 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/145c4da2-bb50-4900-84d7-974b0cfaa4d8-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Nov 28 13:23:15 crc kubenswrapper[4631]: I1128 13:23:15.425835 4631 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/145c4da2-bb50-4900-84d7-974b0cfaa4d8-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Nov 28 13:23:15 crc kubenswrapper[4631]: I1128 13:23:15.425846 4631 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/145c4da2-bb50-4900-84d7-974b0cfaa4d8-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 13:23:15 crc kubenswrapper[4631]: I1128 13:23:15.425855 4631 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/145c4da2-bb50-4900-84d7-974b0cfaa4d8-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Nov 28 13:23:15 crc kubenswrapper[4631]: I1128 13:23:15.425866 4631 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/145c4da2-bb50-4900-84d7-974b0cfaa4d8-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Nov 28 13:23:15 crc kubenswrapper[4631]: I1128 13:23:15.425875 4631 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/145c4da2-bb50-4900-84d7-974b0cfaa4d8-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 28 13:23:15 crc kubenswrapper[4631]: I1128 13:23:15.425884 4631 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/145c4da2-bb50-4900-84d7-974b0cfaa4d8-audit-policies\") on node \"crc\" DevicePath \"\"" Nov 28 13:23:15 crc kubenswrapper[4631]: I1128 13:23:15.623487 4631 generic.go:334] "Generic (PLEG): container finished" podID="145c4da2-bb50-4900-84d7-974b0cfaa4d8" containerID="ebf95e54de61f41d4a27cf2da056c3eb14c5d892726916d1ad9ca6b1ed823a0f" exitCode=0 Nov 28 13:23:15 crc kubenswrapper[4631]: I1128 13:23:15.623540 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-hgsfk" event={"ID":"145c4da2-bb50-4900-84d7-974b0cfaa4d8","Type":"ContainerDied","Data":"ebf95e54de61f41d4a27cf2da056c3eb14c5d892726916d1ad9ca6b1ed823a0f"} Nov 28 13:23:15 crc kubenswrapper[4631]: I1128 13:23:15.623597 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-hgsfk" event={"ID":"145c4da2-bb50-4900-84d7-974b0cfaa4d8","Type":"ContainerDied","Data":"4a8d6262961fe4ede84838daf319eff68a0398144877b19b21af09e42cac844e"} Nov 28 13:23:15 crc kubenswrapper[4631]: I1128 13:23:15.623620 4631 scope.go:117] "RemoveContainer" containerID="ebf95e54de61f41d4a27cf2da056c3eb14c5d892726916d1ad9ca6b1ed823a0f" Nov 28 13:23:15 crc kubenswrapper[4631]: I1128 13:23:15.623652 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-hgsfk" Nov 28 13:23:15 crc kubenswrapper[4631]: I1128 13:23:15.624821 4631 status_manager.go:851] "Failed to get status for pod" podUID="e55d89ce-6920-4cb5-951f-5b36cb73b6df" pod="openshift-marketplace/certified-operators-j6d86" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-j6d86\": dial tcp 38.102.83.194:6443: connect: connection refused" Nov 28 13:23:15 crc kubenswrapper[4631]: I1128 13:23:15.625173 4631 status_manager.go:851] "Failed to get status for pod" podUID="abdc286b-6b18-47b2-b974-410a0f9614b9" pod="openshift-marketplace/redhat-marketplace-fsz47" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-fsz47\": dial tcp 38.102.83.194:6443: connect: connection refused" Nov 28 13:23:15 crc kubenswrapper[4631]: I1128 13:23:15.625875 4631 status_manager.go:851] "Failed to get status for pod" podUID="145c4da2-bb50-4900-84d7-974b0cfaa4d8" pod="openshift-authentication/oauth-openshift-558db77b4-hgsfk" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-hgsfk\": dial tcp 38.102.83.194:6443: connect: connection refused" Nov 28 13:23:15 crc kubenswrapper[4631]: I1128 13:23:15.626197 4631 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.194:6443: connect: connection refused" Nov 28 13:23:15 crc kubenswrapper[4631]: I1128 13:23:15.626436 4631 status_manager.go:851] "Failed to get status for pod" podUID="3a23494b-5ae8-4ace-8763-b94ffe64c5a2" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.194:6443: connect: connection refused" Nov 28 13:23:15 crc kubenswrapper[4631]: I1128 13:23:15.626689 4631 status_manager.go:851] "Failed to get status for pod" podUID="bbd0356c-2316-430e-81d3-c3c505c86e9e" pod="openshift-marketplace/community-operators-7hvgh" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-7hvgh\": dial tcp 38.102.83.194:6443: connect: connection refused" Nov 28 13:23:15 crc kubenswrapper[4631]: I1128 13:23:15.627249 4631 status_manager.go:851] "Failed to get status for pod" podUID="3a23494b-5ae8-4ace-8763-b94ffe64c5a2" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.194:6443: connect: connection refused" Nov 28 13:23:15 crc kubenswrapper[4631]: I1128 13:23:15.627581 4631 status_manager.go:851] "Failed to get status for pod" podUID="bbd0356c-2316-430e-81d3-c3c505c86e9e" pod="openshift-marketplace/community-operators-7hvgh" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-7hvgh\": dial tcp 38.102.83.194:6443: connect: connection refused" Nov 28 13:23:15 crc kubenswrapper[4631]: I1128 13:23:15.627908 4631 status_manager.go:851] "Failed to get status for pod" podUID="e55d89ce-6920-4cb5-951f-5b36cb73b6df" pod="openshift-marketplace/certified-operators-j6d86" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-j6d86\": dial tcp 38.102.83.194:6443: connect: connection refused" Nov 28 13:23:15 crc kubenswrapper[4631]: I1128 13:23:15.628396 4631 status_manager.go:851] "Failed to get status for pod" podUID="abdc286b-6b18-47b2-b974-410a0f9614b9" pod="openshift-marketplace/redhat-marketplace-fsz47" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-fsz47\": dial tcp 38.102.83.194:6443: connect: connection refused" Nov 28 13:23:15 crc kubenswrapper[4631]: I1128 13:23:15.628689 4631 status_manager.go:851] "Failed to get status for pod" podUID="145c4da2-bb50-4900-84d7-974b0cfaa4d8" pod="openshift-authentication/oauth-openshift-558db77b4-hgsfk" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-hgsfk\": dial tcp 38.102.83.194:6443: connect: connection refused" Nov 28 13:23:15 crc kubenswrapper[4631]: I1128 13:23:15.629066 4631 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.194:6443: connect: connection refused" Nov 28 13:23:15 crc kubenswrapper[4631]: I1128 13:23:15.643703 4631 scope.go:117] "RemoveContainer" containerID="ebf95e54de61f41d4a27cf2da056c3eb14c5d892726916d1ad9ca6b1ed823a0f" Nov 28 13:23:15 crc kubenswrapper[4631]: E1128 13:23:15.644607 4631 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ebf95e54de61f41d4a27cf2da056c3eb14c5d892726916d1ad9ca6b1ed823a0f\": container with ID starting with ebf95e54de61f41d4a27cf2da056c3eb14c5d892726916d1ad9ca6b1ed823a0f not found: ID does not exist" containerID="ebf95e54de61f41d4a27cf2da056c3eb14c5d892726916d1ad9ca6b1ed823a0f" Nov 28 13:23:15 crc kubenswrapper[4631]: I1128 13:23:15.644665 4631 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ebf95e54de61f41d4a27cf2da056c3eb14c5d892726916d1ad9ca6b1ed823a0f"} err="failed to get container status \"ebf95e54de61f41d4a27cf2da056c3eb14c5d892726916d1ad9ca6b1ed823a0f\": rpc error: code = NotFound desc = could not find container \"ebf95e54de61f41d4a27cf2da056c3eb14c5d892726916d1ad9ca6b1ed823a0f\": container with ID starting with ebf95e54de61f41d4a27cf2da056c3eb14c5d892726916d1ad9ca6b1ed823a0f not found: ID does not exist" Nov 28 13:23:15 crc kubenswrapper[4631]: E1128 13:23:15.963423 4631 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.194:6443: connect: connection refused" Nov 28 13:23:15 crc kubenswrapper[4631]: E1128 13:23:15.963919 4631 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.194:6443: connect: connection refused" Nov 28 13:23:15 crc kubenswrapper[4631]: E1128 13:23:15.964402 4631 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.194:6443: connect: connection refused" Nov 28 13:23:15 crc kubenswrapper[4631]: E1128 13:23:15.964838 4631 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.194:6443: connect: connection refused" Nov 28 13:23:15 crc kubenswrapper[4631]: E1128 13:23:15.965372 4631 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.194:6443: connect: connection refused" Nov 28 13:23:15 crc kubenswrapper[4631]: I1128 13:23:15.965433 4631 controller.go:115] "failed to update lease using latest lease, fallback to ensure lease" err="failed 5 attempts to update lease" Nov 28 13:23:15 crc kubenswrapper[4631]: E1128 13:23:15.965837 4631 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.194:6443: connect: connection refused" interval="200ms" Nov 28 13:23:16 crc kubenswrapper[4631]: E1128 13:23:16.167884 4631 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.194:6443: connect: connection refused" interval="400ms" Nov 28 13:23:16 crc kubenswrapper[4631]: E1128 13:23:16.569114 4631 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.194:6443: connect: connection refused" interval="800ms" Nov 28 13:23:17 crc kubenswrapper[4631]: E1128 13:23:17.371187 4631 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.194:6443: connect: connection refused" interval="1.6s" Nov 28 13:23:18 crc kubenswrapper[4631]: I1128 13:23:18.138430 4631 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-mzsr9" Nov 28 13:23:18 crc kubenswrapper[4631]: I1128 13:23:18.138980 4631 status_manager.go:851] "Failed to get status for pod" podUID="3a23494b-5ae8-4ace-8763-b94ffe64c5a2" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.194:6443: connect: connection refused" Nov 28 13:23:18 crc kubenswrapper[4631]: I1128 13:23:18.139359 4631 status_manager.go:851] "Failed to get status for pod" podUID="bbd0356c-2316-430e-81d3-c3c505c86e9e" pod="openshift-marketplace/community-operators-7hvgh" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-7hvgh\": dial tcp 38.102.83.194:6443: connect: connection refused" Nov 28 13:23:18 crc kubenswrapper[4631]: I1128 13:23:18.139651 4631 status_manager.go:851] "Failed to get status for pod" podUID="e55d89ce-6920-4cb5-951f-5b36cb73b6df" pod="openshift-marketplace/certified-operators-j6d86" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-j6d86\": dial tcp 38.102.83.194:6443: connect: connection refused" Nov 28 13:23:18 crc kubenswrapper[4631]: I1128 13:23:18.140543 4631 status_manager.go:851] "Failed to get status for pod" podUID="abdc286b-6b18-47b2-b974-410a0f9614b9" pod="openshift-marketplace/redhat-marketplace-fsz47" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-fsz47\": dial tcp 38.102.83.194:6443: connect: connection refused" Nov 28 13:23:18 crc kubenswrapper[4631]: I1128 13:23:18.141048 4631 status_manager.go:851] "Failed to get status for pod" podUID="145c4da2-bb50-4900-84d7-974b0cfaa4d8" pod="openshift-authentication/oauth-openshift-558db77b4-hgsfk" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-hgsfk\": dial tcp 38.102.83.194:6443: connect: connection refused" Nov 28 13:23:18 crc kubenswrapper[4631]: I1128 13:23:18.141531 4631 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.194:6443: connect: connection refused" Nov 28 13:23:18 crc kubenswrapper[4631]: I1128 13:23:18.142004 4631 status_manager.go:851] "Failed to get status for pod" podUID="d2115a7c-dc8e-4f02-bded-25e080c7de51" pod="openshift-marketplace/redhat-operators-mzsr9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-mzsr9\": dial tcp 38.102.83.194:6443: connect: connection refused" Nov 28 13:23:18 crc kubenswrapper[4631]: I1128 13:23:18.183468 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-mzsr9" Nov 28 13:23:18 crc kubenswrapper[4631]: I1128 13:23:18.184075 4631 status_manager.go:851] "Failed to get status for pod" podUID="e55d89ce-6920-4cb5-951f-5b36cb73b6df" pod="openshift-marketplace/certified-operators-j6d86" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-j6d86\": dial tcp 38.102.83.194:6443: connect: connection refused" Nov 28 13:23:18 crc kubenswrapper[4631]: I1128 13:23:18.184679 4631 status_manager.go:851] "Failed to get status for pod" podUID="abdc286b-6b18-47b2-b974-410a0f9614b9" pod="openshift-marketplace/redhat-marketplace-fsz47" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-fsz47\": dial tcp 38.102.83.194:6443: connect: connection refused" Nov 28 13:23:18 crc kubenswrapper[4631]: I1128 13:23:18.185242 4631 status_manager.go:851] "Failed to get status for pod" podUID="145c4da2-bb50-4900-84d7-974b0cfaa4d8" pod="openshift-authentication/oauth-openshift-558db77b4-hgsfk" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-hgsfk\": dial tcp 38.102.83.194:6443: connect: connection refused" Nov 28 13:23:18 crc kubenswrapper[4631]: I1128 13:23:18.185672 4631 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.194:6443: connect: connection refused" Nov 28 13:23:18 crc kubenswrapper[4631]: I1128 13:23:18.186083 4631 status_manager.go:851] "Failed to get status for pod" podUID="d2115a7c-dc8e-4f02-bded-25e080c7de51" pod="openshift-marketplace/redhat-operators-mzsr9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-mzsr9\": dial tcp 38.102.83.194:6443: connect: connection refused" Nov 28 13:23:18 crc kubenswrapper[4631]: I1128 13:23:18.186591 4631 status_manager.go:851] "Failed to get status for pod" podUID="3a23494b-5ae8-4ace-8763-b94ffe64c5a2" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.194:6443: connect: connection refused" Nov 28 13:23:18 crc kubenswrapper[4631]: I1128 13:23:18.186924 4631 status_manager.go:851] "Failed to get status for pod" podUID="bbd0356c-2316-430e-81d3-c3c505c86e9e" pod="openshift-marketplace/community-operators-7hvgh" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-7hvgh\": dial tcp 38.102.83.194:6443: connect: connection refused" Nov 28 13:23:19 crc kubenswrapper[4631]: E1128 13:23:19.126154 4631 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.194:6443: connect: connection refused" interval="3.2s" Nov 28 13:23:19 crc kubenswrapper[4631]: I1128 13:23:19.513283 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 13:23:19 crc kubenswrapper[4631]: I1128 13:23:19.515321 4631 status_manager.go:851] "Failed to get status for pod" podUID="d2115a7c-dc8e-4f02-bded-25e080c7de51" pod="openshift-marketplace/redhat-operators-mzsr9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-mzsr9\": dial tcp 38.102.83.194:6443: connect: connection refused" Nov 28 13:23:19 crc kubenswrapper[4631]: I1128 13:23:19.516041 4631 status_manager.go:851] "Failed to get status for pod" podUID="3a23494b-5ae8-4ace-8763-b94ffe64c5a2" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.194:6443: connect: connection refused" Nov 28 13:23:19 crc kubenswrapper[4631]: I1128 13:23:19.517654 4631 status_manager.go:851] "Failed to get status for pod" podUID="bbd0356c-2316-430e-81d3-c3c505c86e9e" pod="openshift-marketplace/community-operators-7hvgh" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-7hvgh\": dial tcp 38.102.83.194:6443: connect: connection refused" Nov 28 13:23:19 crc kubenswrapper[4631]: I1128 13:23:19.518114 4631 status_manager.go:851] "Failed to get status for pod" podUID="e55d89ce-6920-4cb5-951f-5b36cb73b6df" pod="openshift-marketplace/certified-operators-j6d86" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-j6d86\": dial tcp 38.102.83.194:6443: connect: connection refused" Nov 28 13:23:19 crc kubenswrapper[4631]: I1128 13:23:19.518374 4631 status_manager.go:851] "Failed to get status for pod" podUID="abdc286b-6b18-47b2-b974-410a0f9614b9" pod="openshift-marketplace/redhat-marketplace-fsz47" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-fsz47\": dial tcp 38.102.83.194:6443: connect: connection refused" Nov 28 13:23:19 crc kubenswrapper[4631]: I1128 13:23:19.518693 4631 status_manager.go:851] "Failed to get status for pod" podUID="145c4da2-bb50-4900-84d7-974b0cfaa4d8" pod="openshift-authentication/oauth-openshift-558db77b4-hgsfk" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-hgsfk\": dial tcp 38.102.83.194:6443: connect: connection refused" Nov 28 13:23:19 crc kubenswrapper[4631]: I1128 13:23:19.519039 4631 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.194:6443: connect: connection refused" Nov 28 13:23:19 crc kubenswrapper[4631]: I1128 13:23:19.535068 4631 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="07b4914f-d3c8-4e6a-909c-3af046e1759c" Nov 28 13:23:19 crc kubenswrapper[4631]: I1128 13:23:19.535125 4631 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="07b4914f-d3c8-4e6a-909c-3af046e1759c" Nov 28 13:23:19 crc kubenswrapper[4631]: E1128 13:23:19.535965 4631 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.194:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 13:23:19 crc kubenswrapper[4631]: I1128 13:23:19.536577 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 13:23:19 crc kubenswrapper[4631]: W1128 13:23:19.563937 4631 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod71bb4a3aecc4ba5b26c4b7318770ce13.slice/crio-7b936a3756f9d2b2c85df4fab736caccce41869abf3cd09fbfab10a638903d6b WatchSource:0}: Error finding container 7b936a3756f9d2b2c85df4fab736caccce41869abf3cd09fbfab10a638903d6b: Status 404 returned error can't find the container with id 7b936a3756f9d2b2c85df4fab736caccce41869abf3cd09fbfab10a638903d6b Nov 28 13:23:19 crc kubenswrapper[4631]: I1128 13:23:19.635038 4631 patch_prober.go:28] interesting pod/machine-config-daemon-47sxw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 28 13:23:19 crc kubenswrapper[4631]: I1128 13:23:19.635136 4631 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" podUID="8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 28 13:23:19 crc kubenswrapper[4631]: I1128 13:23:19.664839 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"7b936a3756f9d2b2c85df4fab736caccce41869abf3cd09fbfab10a638903d6b"} Nov 28 13:23:20 crc kubenswrapper[4631]: I1128 13:23:20.670119 4631 generic.go:334] "Generic (PLEG): container finished" podID="71bb4a3aecc4ba5b26c4b7318770ce13" containerID="db6cf18bc439cbe7c947ac82fe373ba0682e008caa73fdc1a4016bbbda7ff6ee" exitCode=0 Nov 28 13:23:20 crc kubenswrapper[4631]: I1128 13:23:20.670176 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerDied","Data":"db6cf18bc439cbe7c947ac82fe373ba0682e008caa73fdc1a4016bbbda7ff6ee"} Nov 28 13:23:20 crc kubenswrapper[4631]: I1128 13:23:20.670442 4631 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="07b4914f-d3c8-4e6a-909c-3af046e1759c" Nov 28 13:23:20 crc kubenswrapper[4631]: I1128 13:23:20.670459 4631 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="07b4914f-d3c8-4e6a-909c-3af046e1759c" Nov 28 13:23:20 crc kubenswrapper[4631]: E1128 13:23:20.670946 4631 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.194:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 13:23:20 crc kubenswrapper[4631]: I1128 13:23:20.671118 4631 status_manager.go:851] "Failed to get status for pod" podUID="d2115a7c-dc8e-4f02-bded-25e080c7de51" pod="openshift-marketplace/redhat-operators-mzsr9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-mzsr9\": dial tcp 38.102.83.194:6443: connect: connection refused" Nov 28 13:23:20 crc kubenswrapper[4631]: I1128 13:23:20.671514 4631 status_manager.go:851] "Failed to get status for pod" podUID="3a23494b-5ae8-4ace-8763-b94ffe64c5a2" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.194:6443: connect: connection refused" Nov 28 13:23:20 crc kubenswrapper[4631]: I1128 13:23:20.672460 4631 status_manager.go:851] "Failed to get status for pod" podUID="bbd0356c-2316-430e-81d3-c3c505c86e9e" pod="openshift-marketplace/community-operators-7hvgh" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-7hvgh\": dial tcp 38.102.83.194:6443: connect: connection refused" Nov 28 13:23:20 crc kubenswrapper[4631]: I1128 13:23:20.673116 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Nov 28 13:23:20 crc kubenswrapper[4631]: I1128 13:23:20.673149 4631 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="35d912d041da9a7ed307ebea6e999cf087dacf84a3cefbfd0f3a9c891c979408" exitCode=1 Nov 28 13:23:20 crc kubenswrapper[4631]: I1128 13:23:20.673172 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"35d912d041da9a7ed307ebea6e999cf087dacf84a3cefbfd0f3a9c891c979408"} Nov 28 13:23:20 crc kubenswrapper[4631]: I1128 13:23:20.673640 4631 scope.go:117] "RemoveContainer" containerID="35d912d041da9a7ed307ebea6e999cf087dacf84a3cefbfd0f3a9c891c979408" Nov 28 13:23:20 crc kubenswrapper[4631]: I1128 13:23:20.674163 4631 status_manager.go:851] "Failed to get status for pod" podUID="e55d89ce-6920-4cb5-951f-5b36cb73b6df" pod="openshift-marketplace/certified-operators-j6d86" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-j6d86\": dial tcp 38.102.83.194:6443: connect: connection refused" Nov 28 13:23:20 crc kubenswrapper[4631]: I1128 13:23:20.674418 4631 status_manager.go:851] "Failed to get status for pod" podUID="abdc286b-6b18-47b2-b974-410a0f9614b9" pod="openshift-marketplace/redhat-marketplace-fsz47" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-fsz47\": dial tcp 38.102.83.194:6443: connect: connection refused" Nov 28 13:23:20 crc kubenswrapper[4631]: I1128 13:23:20.674662 4631 status_manager.go:851] "Failed to get status for pod" podUID="145c4da2-bb50-4900-84d7-974b0cfaa4d8" pod="openshift-authentication/oauth-openshift-558db77b4-hgsfk" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-hgsfk\": dial tcp 38.102.83.194:6443: connect: connection refused" Nov 28 13:23:20 crc kubenswrapper[4631]: I1128 13:23:20.674898 4631 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.194:6443: connect: connection refused" Nov 28 13:23:20 crc kubenswrapper[4631]: I1128 13:23:20.675206 4631 status_manager.go:851] "Failed to get status for pod" podUID="3a23494b-5ae8-4ace-8763-b94ffe64c5a2" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.194:6443: connect: connection refused" Nov 28 13:23:20 crc kubenswrapper[4631]: I1128 13:23:20.676123 4631 status_manager.go:851] "Failed to get status for pod" podUID="bbd0356c-2316-430e-81d3-c3c505c86e9e" pod="openshift-marketplace/community-operators-7hvgh" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-7hvgh\": dial tcp 38.102.83.194:6443: connect: connection refused" Nov 28 13:23:20 crc kubenswrapper[4631]: I1128 13:23:20.677771 4631 status_manager.go:851] "Failed to get status for pod" podUID="e55d89ce-6920-4cb5-951f-5b36cb73b6df" pod="openshift-marketplace/certified-operators-j6d86" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-j6d86\": dial tcp 38.102.83.194:6443: connect: connection refused" Nov 28 13:23:20 crc kubenswrapper[4631]: I1128 13:23:20.678406 4631 status_manager.go:851] "Failed to get status for pod" podUID="abdc286b-6b18-47b2-b974-410a0f9614b9" pod="openshift-marketplace/redhat-marketplace-fsz47" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-fsz47\": dial tcp 38.102.83.194:6443: connect: connection refused" Nov 28 13:23:20 crc kubenswrapper[4631]: I1128 13:23:20.678816 4631 status_manager.go:851] "Failed to get status for pod" podUID="145c4da2-bb50-4900-84d7-974b0cfaa4d8" pod="openshift-authentication/oauth-openshift-558db77b4-hgsfk" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-hgsfk\": dial tcp 38.102.83.194:6443: connect: connection refused" Nov 28 13:23:20 crc kubenswrapper[4631]: I1128 13:23:20.679340 4631 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.194:6443: connect: connection refused" Nov 28 13:23:20 crc kubenswrapper[4631]: I1128 13:23:20.679961 4631 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.194:6443: connect: connection refused" Nov 28 13:23:20 crc kubenswrapper[4631]: I1128 13:23:20.680278 4631 status_manager.go:851] "Failed to get status for pod" podUID="d2115a7c-dc8e-4f02-bded-25e080c7de51" pod="openshift-marketplace/redhat-operators-mzsr9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-mzsr9\": dial tcp 38.102.83.194:6443: connect: connection refused" Nov 28 13:23:21 crc kubenswrapper[4631]: I1128 13:23:21.693519 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"0628be8bc2ce6ff4df1d8a5de711ca69a887db0338ec087091755a2b1f8caa30"} Nov 28 13:23:21 crc kubenswrapper[4631]: I1128 13:23:21.693847 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"5d7b014e968e1cb208c1ef5164f14cb60daf81b7dc9edc125b1b7cecc5f1aaee"} Nov 28 13:23:21 crc kubenswrapper[4631]: I1128 13:23:21.701561 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Nov 28 13:23:21 crc kubenswrapper[4631]: I1128 13:23:21.701616 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"836a798ea54bfd7096c4ada3e228a81d9c4741a57a0396dabf0712e6e139aaf3"} Nov 28 13:23:22 crc kubenswrapper[4631]: I1128 13:23:22.708672 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"d32b45bbc1271ff4e83aaa44da410f09aa3b3ebf5ced61fbcd524c24df553ade"} Nov 28 13:23:22 crc kubenswrapper[4631]: I1128 13:23:22.708926 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"72d46b65c66b8a6b34a064a3c826a59678b6274a038d865eb6eb85a1c2f7c0e0"} Nov 28 13:23:22 crc kubenswrapper[4631]: I1128 13:23:22.708940 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"851f3b04b1e827d5421ec553c387b989ce4469bf8b694b24c68356701f4fcf9c"} Nov 28 13:23:22 crc kubenswrapper[4631]: I1128 13:23:22.709243 4631 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="07b4914f-d3c8-4e6a-909c-3af046e1759c" Nov 28 13:23:22 crc kubenswrapper[4631]: I1128 13:23:22.709256 4631 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="07b4914f-d3c8-4e6a-909c-3af046e1759c" Nov 28 13:23:22 crc kubenswrapper[4631]: I1128 13:23:22.709428 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 13:23:24 crc kubenswrapper[4631]: I1128 13:23:24.494113 4631 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-5v5xp" podUID="14717577-3fab-482b-bdbe-0e604d23d6d8" containerName="registry" containerID="cri-o://a9523ebf1eca553f15218be6b889124d3e5f5d105019f43c00e66ba5e3270a43" gracePeriod=30 Nov 28 13:23:24 crc kubenswrapper[4631]: I1128 13:23:24.537432 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 13:23:24 crc kubenswrapper[4631]: I1128 13:23:24.537634 4631 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 13:23:24 crc kubenswrapper[4631]: I1128 13:23:24.547395 4631 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 13:23:24 crc kubenswrapper[4631]: I1128 13:23:24.720741 4631 generic.go:334] "Generic (PLEG): container finished" podID="14717577-3fab-482b-bdbe-0e604d23d6d8" containerID="a9523ebf1eca553f15218be6b889124d3e5f5d105019f43c00e66ba5e3270a43" exitCode=0 Nov 28 13:23:24 crc kubenswrapper[4631]: I1128 13:23:24.720882 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-5v5xp" event={"ID":"14717577-3fab-482b-bdbe-0e604d23d6d8","Type":"ContainerDied","Data":"a9523ebf1eca553f15218be6b889124d3e5f5d105019f43c00e66ba5e3270a43"} Nov 28 13:23:25 crc kubenswrapper[4631]: I1128 13:23:25.125178 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-5v5xp" Nov 28 13:23:25 crc kubenswrapper[4631]: I1128 13:23:25.307782 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/14717577-3fab-482b-bdbe-0e604d23d6d8-installation-pull-secrets\") pod \"14717577-3fab-482b-bdbe-0e604d23d6d8\" (UID: \"14717577-3fab-482b-bdbe-0e604d23d6d8\") " Nov 28 13:23:25 crc kubenswrapper[4631]: I1128 13:23:25.308068 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"14717577-3fab-482b-bdbe-0e604d23d6d8\" (UID: \"14717577-3fab-482b-bdbe-0e604d23d6d8\") " Nov 28 13:23:25 crc kubenswrapper[4631]: I1128 13:23:25.308123 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/14717577-3fab-482b-bdbe-0e604d23d6d8-bound-sa-token\") pod \"14717577-3fab-482b-bdbe-0e604d23d6d8\" (UID: \"14717577-3fab-482b-bdbe-0e604d23d6d8\") " Nov 28 13:23:25 crc kubenswrapper[4631]: I1128 13:23:25.308185 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/14717577-3fab-482b-bdbe-0e604d23d6d8-registry-certificates\") pod \"14717577-3fab-482b-bdbe-0e604d23d6d8\" (UID: \"14717577-3fab-482b-bdbe-0e604d23d6d8\") " Nov 28 13:23:25 crc kubenswrapper[4631]: I1128 13:23:25.308223 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/14717577-3fab-482b-bdbe-0e604d23d6d8-trusted-ca\") pod \"14717577-3fab-482b-bdbe-0e604d23d6d8\" (UID: \"14717577-3fab-482b-bdbe-0e604d23d6d8\") " Nov 28 13:23:25 crc kubenswrapper[4631]: I1128 13:23:25.308255 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/14717577-3fab-482b-bdbe-0e604d23d6d8-registry-tls\") pod \"14717577-3fab-482b-bdbe-0e604d23d6d8\" (UID: \"14717577-3fab-482b-bdbe-0e604d23d6d8\") " Nov 28 13:23:25 crc kubenswrapper[4631]: I1128 13:23:25.308337 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/14717577-3fab-482b-bdbe-0e604d23d6d8-ca-trust-extracted\") pod \"14717577-3fab-482b-bdbe-0e604d23d6d8\" (UID: \"14717577-3fab-482b-bdbe-0e604d23d6d8\") " Nov 28 13:23:25 crc kubenswrapper[4631]: I1128 13:23:25.308401 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lqvsn\" (UniqueName: \"kubernetes.io/projected/14717577-3fab-482b-bdbe-0e604d23d6d8-kube-api-access-lqvsn\") pod \"14717577-3fab-482b-bdbe-0e604d23d6d8\" (UID: \"14717577-3fab-482b-bdbe-0e604d23d6d8\") " Nov 28 13:23:25 crc kubenswrapper[4631]: I1128 13:23:25.309538 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/14717577-3fab-482b-bdbe-0e604d23d6d8-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "14717577-3fab-482b-bdbe-0e604d23d6d8" (UID: "14717577-3fab-482b-bdbe-0e604d23d6d8"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 13:23:25 crc kubenswrapper[4631]: I1128 13:23:25.310256 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/14717577-3fab-482b-bdbe-0e604d23d6d8-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "14717577-3fab-482b-bdbe-0e604d23d6d8" (UID: "14717577-3fab-482b-bdbe-0e604d23d6d8"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 13:23:25 crc kubenswrapper[4631]: I1128 13:23:25.327125 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/14717577-3fab-482b-bdbe-0e604d23d6d8-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "14717577-3fab-482b-bdbe-0e604d23d6d8" (UID: "14717577-3fab-482b-bdbe-0e604d23d6d8"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:23:25 crc kubenswrapper[4631]: I1128 13:23:25.327154 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/14717577-3fab-482b-bdbe-0e604d23d6d8-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "14717577-3fab-482b-bdbe-0e604d23d6d8" (UID: "14717577-3fab-482b-bdbe-0e604d23d6d8"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:23:25 crc kubenswrapper[4631]: I1128 13:23:25.327630 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/14717577-3fab-482b-bdbe-0e604d23d6d8-kube-api-access-lqvsn" (OuterVolumeSpecName: "kube-api-access-lqvsn") pod "14717577-3fab-482b-bdbe-0e604d23d6d8" (UID: "14717577-3fab-482b-bdbe-0e604d23d6d8"). InnerVolumeSpecName "kube-api-access-lqvsn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:23:25 crc kubenswrapper[4631]: I1128 13:23:25.328089 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/14717577-3fab-482b-bdbe-0e604d23d6d8-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "14717577-3fab-482b-bdbe-0e604d23d6d8" (UID: "14717577-3fab-482b-bdbe-0e604d23d6d8"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:23:25 crc kubenswrapper[4631]: I1128 13:23:25.332066 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "14717577-3fab-482b-bdbe-0e604d23d6d8" (UID: "14717577-3fab-482b-bdbe-0e604d23d6d8"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Nov 28 13:23:25 crc kubenswrapper[4631]: I1128 13:23:25.347186 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/14717577-3fab-482b-bdbe-0e604d23d6d8-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "14717577-3fab-482b-bdbe-0e604d23d6d8" (UID: "14717577-3fab-482b-bdbe-0e604d23d6d8"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 13:23:25 crc kubenswrapper[4631]: I1128 13:23:25.409622 4631 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/14717577-3fab-482b-bdbe-0e604d23d6d8-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 28 13:23:25 crc kubenswrapper[4631]: I1128 13:23:25.409663 4631 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/14717577-3fab-482b-bdbe-0e604d23d6d8-registry-certificates\") on node \"crc\" DevicePath \"\"" Nov 28 13:23:25 crc kubenswrapper[4631]: I1128 13:23:25.409675 4631 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/14717577-3fab-482b-bdbe-0e604d23d6d8-registry-tls\") on node \"crc\" DevicePath \"\"" Nov 28 13:23:25 crc kubenswrapper[4631]: I1128 13:23:25.409683 4631 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/14717577-3fab-482b-bdbe-0e604d23d6d8-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Nov 28 13:23:25 crc kubenswrapper[4631]: I1128 13:23:25.409692 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lqvsn\" (UniqueName: \"kubernetes.io/projected/14717577-3fab-482b-bdbe-0e604d23d6d8-kube-api-access-lqvsn\") on node \"crc\" DevicePath \"\"" Nov 28 13:23:25 crc kubenswrapper[4631]: I1128 13:23:25.409701 4631 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/14717577-3fab-482b-bdbe-0e604d23d6d8-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Nov 28 13:23:25 crc kubenswrapper[4631]: I1128 13:23:25.409708 4631 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/14717577-3fab-482b-bdbe-0e604d23d6d8-bound-sa-token\") on node \"crc\" DevicePath \"\"" Nov 28 13:23:25 crc kubenswrapper[4631]: I1128 13:23:25.730924 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-5v5xp" event={"ID":"14717577-3fab-482b-bdbe-0e604d23d6d8","Type":"ContainerDied","Data":"6f8e107523760a208fc044fa471cb14ffbde0fe03ac81ffabdf1604ae7e2582a"} Nov 28 13:23:25 crc kubenswrapper[4631]: I1128 13:23:25.730992 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-5v5xp" Nov 28 13:23:25 crc kubenswrapper[4631]: I1128 13:23:25.731046 4631 scope.go:117] "RemoveContainer" containerID="a9523ebf1eca553f15218be6b889124d3e5f5d105019f43c00e66ba5e3270a43" Nov 28 13:23:27 crc kubenswrapper[4631]: I1128 13:23:27.291278 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 28 13:23:27 crc kubenswrapper[4631]: I1128 13:23:27.729999 4631 kubelet.go:1914] "Deleted mirror pod because it is outdated" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 13:23:27 crc kubenswrapper[4631]: I1128 13:23:27.869221 4631 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="7317ccae-ce82-4a5b-9bdd-d3fb3e8be3f5" Nov 28 13:23:28 crc kubenswrapper[4631]: I1128 13:23:28.764990 4631 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="07b4914f-d3c8-4e6a-909c-3af046e1759c" Nov 28 13:23:28 crc kubenswrapper[4631]: I1128 13:23:28.765034 4631 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="07b4914f-d3c8-4e6a-909c-3af046e1759c" Nov 28 13:23:28 crc kubenswrapper[4631]: I1128 13:23:28.769826 4631 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="7317ccae-ce82-4a5b-9bdd-d3fb3e8be3f5" Nov 28 13:23:28 crc kubenswrapper[4631]: I1128 13:23:28.772512 4631 status_manager.go:308] "Container readiness changed before pod has synced" pod="openshift-kube-apiserver/kube-apiserver-crc" containerID="cri-o://5d7b014e968e1cb208c1ef5164f14cb60daf81b7dc9edc125b1b7cecc5f1aaee" Nov 28 13:23:28 crc kubenswrapper[4631]: I1128 13:23:28.772574 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 13:23:29 crc kubenswrapper[4631]: I1128 13:23:29.771542 4631 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="07b4914f-d3c8-4e6a-909c-3af046e1759c" Nov 28 13:23:29 crc kubenswrapper[4631]: I1128 13:23:29.771583 4631 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="07b4914f-d3c8-4e6a-909c-3af046e1759c" Nov 28 13:23:29 crc kubenswrapper[4631]: I1128 13:23:29.775784 4631 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="7317ccae-ce82-4a5b-9bdd-d3fb3e8be3f5" Nov 28 13:23:30 crc kubenswrapper[4631]: I1128 13:23:30.232897 4631 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 28 13:23:30 crc kubenswrapper[4631]: I1128 13:23:30.240814 4631 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 28 13:23:37 crc kubenswrapper[4631]: I1128 13:23:37.298838 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 28 13:23:38 crc kubenswrapper[4631]: I1128 13:23:38.013086 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Nov 28 13:23:38 crc kubenswrapper[4631]: I1128 13:23:38.015839 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Nov 28 13:23:38 crc kubenswrapper[4631]: I1128 13:23:38.037104 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Nov 28 13:23:38 crc kubenswrapper[4631]: I1128 13:23:38.750557 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Nov 28 13:23:38 crc kubenswrapper[4631]: I1128 13:23:38.905846 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Nov 28 13:23:38 crc kubenswrapper[4631]: I1128 13:23:38.995551 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Nov 28 13:23:39 crc kubenswrapper[4631]: I1128 13:23:39.159875 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Nov 28 13:23:39 crc kubenswrapper[4631]: I1128 13:23:39.380726 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Nov 28 13:23:39 crc kubenswrapper[4631]: I1128 13:23:39.497068 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Nov 28 13:23:39 crc kubenswrapper[4631]: I1128 13:23:39.699116 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Nov 28 13:23:39 crc kubenswrapper[4631]: I1128 13:23:39.721857 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Nov 28 13:23:39 crc kubenswrapper[4631]: I1128 13:23:39.895554 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Nov 28 13:23:39 crc kubenswrapper[4631]: I1128 13:23:39.911684 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Nov 28 13:23:39 crc kubenswrapper[4631]: I1128 13:23:39.949862 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Nov 28 13:23:39 crc kubenswrapper[4631]: I1128 13:23:39.970185 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Nov 28 13:23:40 crc kubenswrapper[4631]: I1128 13:23:40.050485 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Nov 28 13:23:40 crc kubenswrapper[4631]: I1128 13:23:40.059169 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Nov 28 13:23:40 crc kubenswrapper[4631]: I1128 13:23:40.073513 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Nov 28 13:23:40 crc kubenswrapper[4631]: I1128 13:23:40.101904 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Nov 28 13:23:40 crc kubenswrapper[4631]: I1128 13:23:40.264865 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Nov 28 13:23:40 crc kubenswrapper[4631]: I1128 13:23:40.355759 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Nov 28 13:23:40 crc kubenswrapper[4631]: I1128 13:23:40.394798 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Nov 28 13:23:40 crc kubenswrapper[4631]: I1128 13:23:40.543545 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Nov 28 13:23:40 crc kubenswrapper[4631]: I1128 13:23:40.572608 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Nov 28 13:23:40 crc kubenswrapper[4631]: I1128 13:23:40.579159 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Nov 28 13:23:40 crc kubenswrapper[4631]: I1128 13:23:40.592427 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Nov 28 13:23:40 crc kubenswrapper[4631]: I1128 13:23:40.692104 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Nov 28 13:23:40 crc kubenswrapper[4631]: I1128 13:23:40.706651 4631 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Nov 28 13:23:40 crc kubenswrapper[4631]: I1128 13:23:40.720247 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Nov 28 13:23:40 crc kubenswrapper[4631]: I1128 13:23:40.720392 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Nov 28 13:23:40 crc kubenswrapper[4631]: I1128 13:23:40.769453 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Nov 28 13:23:40 crc kubenswrapper[4631]: I1128 13:23:40.829983 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Nov 28 13:23:40 crc kubenswrapper[4631]: I1128 13:23:40.976412 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Nov 28 13:23:40 crc kubenswrapper[4631]: I1128 13:23:40.976543 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Nov 28 13:23:41 crc kubenswrapper[4631]: I1128 13:23:41.096594 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Nov 28 13:23:41 crc kubenswrapper[4631]: I1128 13:23:41.103282 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Nov 28 13:23:41 crc kubenswrapper[4631]: I1128 13:23:41.206664 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Nov 28 13:23:41 crc kubenswrapper[4631]: I1128 13:23:41.465876 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Nov 28 13:23:41 crc kubenswrapper[4631]: I1128 13:23:41.475922 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Nov 28 13:23:41 crc kubenswrapper[4631]: I1128 13:23:41.605848 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Nov 28 13:23:41 crc kubenswrapper[4631]: I1128 13:23:41.661913 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Nov 28 13:23:41 crc kubenswrapper[4631]: I1128 13:23:41.758979 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Nov 28 13:23:41 crc kubenswrapper[4631]: I1128 13:23:41.770509 4631 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Nov 28 13:23:41 crc kubenswrapper[4631]: I1128 13:23:41.804527 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Nov 28 13:23:41 crc kubenswrapper[4631]: I1128 13:23:41.874825 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Nov 28 13:23:41 crc kubenswrapper[4631]: I1128 13:23:41.958546 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Nov 28 13:23:42 crc kubenswrapper[4631]: I1128 13:23:42.037188 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Nov 28 13:23:42 crc kubenswrapper[4631]: I1128 13:23:42.111025 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Nov 28 13:23:42 crc kubenswrapper[4631]: I1128 13:23:42.111041 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Nov 28 13:23:42 crc kubenswrapper[4631]: I1128 13:23:42.331268 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Nov 28 13:23:42 crc kubenswrapper[4631]: I1128 13:23:42.369847 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Nov 28 13:23:42 crc kubenswrapper[4631]: I1128 13:23:42.423853 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Nov 28 13:23:42 crc kubenswrapper[4631]: I1128 13:23:42.640892 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Nov 28 13:23:42 crc kubenswrapper[4631]: I1128 13:23:42.644588 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Nov 28 13:23:42 crc kubenswrapper[4631]: I1128 13:23:42.645361 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Nov 28 13:23:42 crc kubenswrapper[4631]: I1128 13:23:42.697070 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Nov 28 13:23:42 crc kubenswrapper[4631]: I1128 13:23:42.745530 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Nov 28 13:23:42 crc kubenswrapper[4631]: I1128 13:23:42.781897 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Nov 28 13:23:42 crc kubenswrapper[4631]: I1128 13:23:42.868137 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Nov 28 13:23:42 crc kubenswrapper[4631]: I1128 13:23:42.908902 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Nov 28 13:23:42 crc kubenswrapper[4631]: I1128 13:23:42.971012 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Nov 28 13:23:42 crc kubenswrapper[4631]: I1128 13:23:42.989834 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Nov 28 13:23:43 crc kubenswrapper[4631]: I1128 13:23:43.152007 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Nov 28 13:23:43 crc kubenswrapper[4631]: I1128 13:23:43.189918 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Nov 28 13:23:43 crc kubenswrapper[4631]: I1128 13:23:43.192214 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Nov 28 13:23:43 crc kubenswrapper[4631]: I1128 13:23:43.213671 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Nov 28 13:23:43 crc kubenswrapper[4631]: I1128 13:23:43.241906 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Nov 28 13:23:43 crc kubenswrapper[4631]: I1128 13:23:43.265955 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Nov 28 13:23:43 crc kubenswrapper[4631]: I1128 13:23:43.314033 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Nov 28 13:23:43 crc kubenswrapper[4631]: I1128 13:23:43.369034 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Nov 28 13:23:43 crc kubenswrapper[4631]: I1128 13:23:43.404532 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Nov 28 13:23:43 crc kubenswrapper[4631]: I1128 13:23:43.545058 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Nov 28 13:23:43 crc kubenswrapper[4631]: I1128 13:23:43.603489 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Nov 28 13:23:43 crc kubenswrapper[4631]: I1128 13:23:43.638931 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Nov 28 13:23:43 crc kubenswrapper[4631]: I1128 13:23:43.715731 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Nov 28 13:23:43 crc kubenswrapper[4631]: I1128 13:23:43.802515 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Nov 28 13:23:43 crc kubenswrapper[4631]: I1128 13:23:43.923651 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Nov 28 13:23:43 crc kubenswrapper[4631]: I1128 13:23:43.953900 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Nov 28 13:23:44 crc kubenswrapper[4631]: I1128 13:23:44.106575 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Nov 28 13:23:44 crc kubenswrapper[4631]: I1128 13:23:44.125040 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Nov 28 13:23:44 crc kubenswrapper[4631]: I1128 13:23:44.130106 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Nov 28 13:23:44 crc kubenswrapper[4631]: I1128 13:23:44.163602 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Nov 28 13:23:44 crc kubenswrapper[4631]: I1128 13:23:44.308423 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Nov 28 13:23:44 crc kubenswrapper[4631]: I1128 13:23:44.383354 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Nov 28 13:23:44 crc kubenswrapper[4631]: I1128 13:23:44.410674 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Nov 28 13:23:44 crc kubenswrapper[4631]: I1128 13:23:44.467277 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Nov 28 13:23:44 crc kubenswrapper[4631]: I1128 13:23:44.511386 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Nov 28 13:23:44 crc kubenswrapper[4631]: I1128 13:23:44.513206 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Nov 28 13:23:44 crc kubenswrapper[4631]: I1128 13:23:44.678967 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Nov 28 13:23:44 crc kubenswrapper[4631]: I1128 13:23:44.728423 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Nov 28 13:23:44 crc kubenswrapper[4631]: I1128 13:23:44.745779 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Nov 28 13:23:44 crc kubenswrapper[4631]: I1128 13:23:44.873058 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Nov 28 13:23:44 crc kubenswrapper[4631]: I1128 13:23:44.970855 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Nov 28 13:23:45 crc kubenswrapper[4631]: I1128 13:23:45.018159 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Nov 28 13:23:45 crc kubenswrapper[4631]: I1128 13:23:45.062184 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Nov 28 13:23:45 crc kubenswrapper[4631]: I1128 13:23:45.086025 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Nov 28 13:23:45 crc kubenswrapper[4631]: I1128 13:23:45.227244 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Nov 28 13:23:45 crc kubenswrapper[4631]: I1128 13:23:45.357247 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Nov 28 13:23:45 crc kubenswrapper[4631]: I1128 13:23:45.383166 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Nov 28 13:23:45 crc kubenswrapper[4631]: I1128 13:23:45.594940 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Nov 28 13:23:45 crc kubenswrapper[4631]: I1128 13:23:45.720163 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Nov 28 13:23:45 crc kubenswrapper[4631]: I1128 13:23:45.749446 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Nov 28 13:23:45 crc kubenswrapper[4631]: I1128 13:23:45.758465 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Nov 28 13:23:45 crc kubenswrapper[4631]: I1128 13:23:45.767718 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Nov 28 13:23:45 crc kubenswrapper[4631]: I1128 13:23:45.862109 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Nov 28 13:23:45 crc kubenswrapper[4631]: I1128 13:23:45.910522 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Nov 28 13:23:45 crc kubenswrapper[4631]: I1128 13:23:45.937763 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Nov 28 13:23:45 crc kubenswrapper[4631]: I1128 13:23:45.941768 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Nov 28 13:23:45 crc kubenswrapper[4631]: I1128 13:23:45.976159 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Nov 28 13:23:46 crc kubenswrapper[4631]: I1128 13:23:46.015513 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Nov 28 13:23:46 crc kubenswrapper[4631]: I1128 13:23:46.092960 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Nov 28 13:23:46 crc kubenswrapper[4631]: I1128 13:23:46.126975 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Nov 28 13:23:46 crc kubenswrapper[4631]: I1128 13:23:46.139065 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Nov 28 13:23:46 crc kubenswrapper[4631]: I1128 13:23:46.170177 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Nov 28 13:23:46 crc kubenswrapper[4631]: I1128 13:23:46.231066 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Nov 28 13:23:46 crc kubenswrapper[4631]: I1128 13:23:46.281017 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Nov 28 13:23:46 crc kubenswrapper[4631]: I1128 13:23:46.459376 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Nov 28 13:23:46 crc kubenswrapper[4631]: I1128 13:23:46.464504 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Nov 28 13:23:46 crc kubenswrapper[4631]: I1128 13:23:46.536159 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Nov 28 13:23:46 crc kubenswrapper[4631]: I1128 13:23:46.538615 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Nov 28 13:23:46 crc kubenswrapper[4631]: I1128 13:23:46.581996 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Nov 28 13:23:46 crc kubenswrapper[4631]: I1128 13:23:46.633383 4631 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Nov 28 13:23:46 crc kubenswrapper[4631]: I1128 13:23:46.637310 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podStartSLOduration=40.637270262 podStartE2EDuration="40.637270262s" podCreationTimestamp="2025-11-28 13:23:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 13:23:27.674707461 +0000 UTC m=+164.482010805" watchObservedRunningTime="2025-11-28 13:23:46.637270262 +0000 UTC m=+183.444573606" Nov 28 13:23:46 crc kubenswrapper[4631]: I1128 13:23:46.638664 4631 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc","openshift-image-registry/image-registry-697d97f7c8-5v5xp","openshift-authentication/oauth-openshift-558db77b4-hgsfk"] Nov 28 13:23:46 crc kubenswrapper[4631]: I1128 13:23:46.638738 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Nov 28 13:23:46 crc kubenswrapper[4631]: I1128 13:23:46.646329 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 13:23:46 crc kubenswrapper[4631]: I1128 13:23:46.665831 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=19.665809382 podStartE2EDuration="19.665809382s" podCreationTimestamp="2025-11-28 13:23:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 13:23:46.661151065 +0000 UTC m=+183.468454489" watchObservedRunningTime="2025-11-28 13:23:46.665809382 +0000 UTC m=+183.473112736" Nov 28 13:23:46 crc kubenswrapper[4631]: I1128 13:23:46.822594 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Nov 28 13:23:46 crc kubenswrapper[4631]: I1128 13:23:46.825205 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Nov 28 13:23:46 crc kubenswrapper[4631]: I1128 13:23:46.842459 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Nov 28 13:23:46 crc kubenswrapper[4631]: I1128 13:23:46.845498 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Nov 28 13:23:46 crc kubenswrapper[4631]: I1128 13:23:46.888532 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Nov 28 13:23:46 crc kubenswrapper[4631]: I1128 13:23:46.900136 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Nov 28 13:23:46 crc kubenswrapper[4631]: I1128 13:23:46.906059 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Nov 28 13:23:46 crc kubenswrapper[4631]: I1128 13:23:46.916438 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Nov 28 13:23:46 crc kubenswrapper[4631]: I1128 13:23:46.921344 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Nov 28 13:23:46 crc kubenswrapper[4631]: I1128 13:23:46.950725 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Nov 28 13:23:46 crc kubenswrapper[4631]: I1128 13:23:46.962233 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Nov 28 13:23:47 crc kubenswrapper[4631]: I1128 13:23:47.232054 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Nov 28 13:23:47 crc kubenswrapper[4631]: I1128 13:23:47.291451 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Nov 28 13:23:47 crc kubenswrapper[4631]: I1128 13:23:47.327072 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Nov 28 13:23:47 crc kubenswrapper[4631]: I1128 13:23:47.424470 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Nov 28 13:23:47 crc kubenswrapper[4631]: I1128 13:23:47.480279 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Nov 28 13:23:47 crc kubenswrapper[4631]: I1128 13:23:47.523845 4631 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="145c4da2-bb50-4900-84d7-974b0cfaa4d8" path="/var/lib/kubelet/pods/145c4da2-bb50-4900-84d7-974b0cfaa4d8/volumes" Nov 28 13:23:47 crc kubenswrapper[4631]: I1128 13:23:47.526403 4631 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="14717577-3fab-482b-bdbe-0e604d23d6d8" path="/var/lib/kubelet/pods/14717577-3fab-482b-bdbe-0e604d23d6d8/volumes" Nov 28 13:23:47 crc kubenswrapper[4631]: I1128 13:23:47.567044 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Nov 28 13:23:47 crc kubenswrapper[4631]: I1128 13:23:47.598520 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Nov 28 13:23:47 crc kubenswrapper[4631]: I1128 13:23:47.637058 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Nov 28 13:23:47 crc kubenswrapper[4631]: I1128 13:23:47.746906 4631 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Nov 28 13:23:47 crc kubenswrapper[4631]: I1128 13:23:47.747540 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Nov 28 13:23:47 crc kubenswrapper[4631]: I1128 13:23:47.822486 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Nov 28 13:23:47 crc kubenswrapper[4631]: I1128 13:23:47.896790 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Nov 28 13:23:47 crc kubenswrapper[4631]: I1128 13:23:47.916513 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Nov 28 13:23:47 crc kubenswrapper[4631]: I1128 13:23:47.949931 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Nov 28 13:23:47 crc kubenswrapper[4631]: I1128 13:23:47.958880 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Nov 28 13:23:48 crc kubenswrapper[4631]: I1128 13:23:48.013644 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Nov 28 13:23:48 crc kubenswrapper[4631]: I1128 13:23:48.047077 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Nov 28 13:23:48 crc kubenswrapper[4631]: I1128 13:23:48.067431 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Nov 28 13:23:48 crc kubenswrapper[4631]: I1128 13:23:48.117911 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Nov 28 13:23:48 crc kubenswrapper[4631]: I1128 13:23:48.159933 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Nov 28 13:23:48 crc kubenswrapper[4631]: I1128 13:23:48.166505 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Nov 28 13:23:48 crc kubenswrapper[4631]: I1128 13:23:48.237229 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Nov 28 13:23:48 crc kubenswrapper[4631]: I1128 13:23:48.298492 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Nov 28 13:23:48 crc kubenswrapper[4631]: I1128 13:23:48.302580 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Nov 28 13:23:48 crc kubenswrapper[4631]: I1128 13:23:48.322240 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Nov 28 13:23:48 crc kubenswrapper[4631]: I1128 13:23:48.330718 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Nov 28 13:23:48 crc kubenswrapper[4631]: I1128 13:23:48.439782 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Nov 28 13:23:48 crc kubenswrapper[4631]: I1128 13:23:48.590039 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Nov 28 13:23:48 crc kubenswrapper[4631]: I1128 13:23:48.608322 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Nov 28 13:23:48 crc kubenswrapper[4631]: I1128 13:23:48.687906 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Nov 28 13:23:49 crc kubenswrapper[4631]: I1128 13:23:49.021462 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Nov 28 13:23:49 crc kubenswrapper[4631]: I1128 13:23:49.043152 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Nov 28 13:23:49 crc kubenswrapper[4631]: I1128 13:23:49.066848 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Nov 28 13:23:49 crc kubenswrapper[4631]: I1128 13:23:49.151031 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Nov 28 13:23:49 crc kubenswrapper[4631]: I1128 13:23:49.173164 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Nov 28 13:23:49 crc kubenswrapper[4631]: I1128 13:23:49.231722 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Nov 28 13:23:49 crc kubenswrapper[4631]: I1128 13:23:49.299409 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Nov 28 13:23:49 crc kubenswrapper[4631]: I1128 13:23:49.308021 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Nov 28 13:23:49 crc kubenswrapper[4631]: I1128 13:23:49.326999 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Nov 28 13:23:49 crc kubenswrapper[4631]: I1128 13:23:49.349550 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Nov 28 13:23:49 crc kubenswrapper[4631]: I1128 13:23:49.635409 4631 patch_prober.go:28] interesting pod/machine-config-daemon-47sxw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 28 13:23:49 crc kubenswrapper[4631]: I1128 13:23:49.635493 4631 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" podUID="8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 28 13:23:49 crc kubenswrapper[4631]: I1128 13:23:49.691141 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Nov 28 13:23:49 crc kubenswrapper[4631]: I1128 13:23:49.716322 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Nov 28 13:23:49 crc kubenswrapper[4631]: I1128 13:23:49.737222 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Nov 28 13:23:49 crc kubenswrapper[4631]: I1128 13:23:49.797217 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Nov 28 13:23:49 crc kubenswrapper[4631]: I1128 13:23:49.828579 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Nov 28 13:23:49 crc kubenswrapper[4631]: I1128 13:23:49.897874 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Nov 28 13:23:49 crc kubenswrapper[4631]: I1128 13:23:49.993094 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Nov 28 13:23:50 crc kubenswrapper[4631]: I1128 13:23:50.073157 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Nov 28 13:23:50 crc kubenswrapper[4631]: I1128 13:23:50.092365 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Nov 28 13:23:50 crc kubenswrapper[4631]: I1128 13:23:50.124075 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Nov 28 13:23:50 crc kubenswrapper[4631]: I1128 13:23:50.200630 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Nov 28 13:23:50 crc kubenswrapper[4631]: I1128 13:23:50.266592 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Nov 28 13:23:50 crc kubenswrapper[4631]: I1128 13:23:50.281774 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Nov 28 13:23:50 crc kubenswrapper[4631]: I1128 13:23:50.286603 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Nov 28 13:23:50 crc kubenswrapper[4631]: I1128 13:23:50.362743 4631 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Nov 28 13:23:50 crc kubenswrapper[4631]: I1128 13:23:50.363159 4631 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" containerID="cri-o://1c53926191bf67e68a0e0230740de428c280f79550fd964cf4bef72a592a0f57" gracePeriod=5 Nov 28 13:23:50 crc kubenswrapper[4631]: I1128 13:23:50.368729 4631 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Nov 28 13:23:50 crc kubenswrapper[4631]: I1128 13:23:50.403924 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Nov 28 13:23:50 crc kubenswrapper[4631]: I1128 13:23:50.447696 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Nov 28 13:23:50 crc kubenswrapper[4631]: I1128 13:23:50.595210 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Nov 28 13:23:50 crc kubenswrapper[4631]: I1128 13:23:50.809098 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Nov 28 13:23:50 crc kubenswrapper[4631]: I1128 13:23:50.876716 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Nov 28 13:23:50 crc kubenswrapper[4631]: I1128 13:23:50.959800 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Nov 28 13:23:51 crc kubenswrapper[4631]: I1128 13:23:51.075901 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Nov 28 13:23:51 crc kubenswrapper[4631]: I1128 13:23:51.109659 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Nov 28 13:23:51 crc kubenswrapper[4631]: I1128 13:23:51.287641 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Nov 28 13:23:51 crc kubenswrapper[4631]: I1128 13:23:51.289892 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Nov 28 13:23:51 crc kubenswrapper[4631]: I1128 13:23:51.300499 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Nov 28 13:23:51 crc kubenswrapper[4631]: I1128 13:23:51.477362 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Nov 28 13:23:51 crc kubenswrapper[4631]: I1128 13:23:51.482078 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Nov 28 13:23:51 crc kubenswrapper[4631]: I1128 13:23:51.562776 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Nov 28 13:23:51 crc kubenswrapper[4631]: I1128 13:23:51.653906 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Nov 28 13:23:51 crc kubenswrapper[4631]: I1128 13:23:51.672521 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Nov 28 13:23:51 crc kubenswrapper[4631]: I1128 13:23:51.765714 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Nov 28 13:23:51 crc kubenswrapper[4631]: I1128 13:23:51.781380 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Nov 28 13:23:51 crc kubenswrapper[4631]: I1128 13:23:51.792806 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Nov 28 13:23:51 crc kubenswrapper[4631]: I1128 13:23:51.832843 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Nov 28 13:23:51 crc kubenswrapper[4631]: I1128 13:23:51.906409 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Nov 28 13:23:51 crc kubenswrapper[4631]: I1128 13:23:51.927269 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Nov 28 13:23:51 crc kubenswrapper[4631]: I1128 13:23:51.950364 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-68b4d6ff9f-trblx"] Nov 28 13:23:51 crc kubenswrapper[4631]: E1128 13:23:51.950712 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="14717577-3fab-482b-bdbe-0e604d23d6d8" containerName="registry" Nov 28 13:23:51 crc kubenswrapper[4631]: I1128 13:23:51.950733 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="14717577-3fab-482b-bdbe-0e604d23d6d8" containerName="registry" Nov 28 13:23:51 crc kubenswrapper[4631]: E1128 13:23:51.950754 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="145c4da2-bb50-4900-84d7-974b0cfaa4d8" containerName="oauth-openshift" Nov 28 13:23:51 crc kubenswrapper[4631]: I1128 13:23:51.950764 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="145c4da2-bb50-4900-84d7-974b0cfaa4d8" containerName="oauth-openshift" Nov 28 13:23:51 crc kubenswrapper[4631]: E1128 13:23:51.950782 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3a23494b-5ae8-4ace-8763-b94ffe64c5a2" containerName="installer" Nov 28 13:23:51 crc kubenswrapper[4631]: I1128 13:23:51.950794 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a23494b-5ae8-4ace-8763-b94ffe64c5a2" containerName="installer" Nov 28 13:23:51 crc kubenswrapper[4631]: E1128 13:23:51.950805 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Nov 28 13:23:51 crc kubenswrapper[4631]: I1128 13:23:51.950815 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Nov 28 13:23:51 crc kubenswrapper[4631]: I1128 13:23:51.950980 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Nov 28 13:23:51 crc kubenswrapper[4631]: I1128 13:23:51.951000 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="3a23494b-5ae8-4ace-8763-b94ffe64c5a2" containerName="installer" Nov 28 13:23:51 crc kubenswrapper[4631]: I1128 13:23:51.951019 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="145c4da2-bb50-4900-84d7-974b0cfaa4d8" containerName="oauth-openshift" Nov 28 13:23:51 crc kubenswrapper[4631]: I1128 13:23:51.951038 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="14717577-3fab-482b-bdbe-0e604d23d6d8" containerName="registry" Nov 28 13:23:51 crc kubenswrapper[4631]: I1128 13:23:51.951735 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-68b4d6ff9f-trblx" Nov 28 13:23:51 crc kubenswrapper[4631]: I1128 13:23:51.956781 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Nov 28 13:23:51 crc kubenswrapper[4631]: I1128 13:23:51.957098 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Nov 28 13:23:51 crc kubenswrapper[4631]: I1128 13:23:51.959708 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Nov 28 13:23:51 crc kubenswrapper[4631]: I1128 13:23:51.959721 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Nov 28 13:23:51 crc kubenswrapper[4631]: I1128 13:23:51.959827 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Nov 28 13:23:51 crc kubenswrapper[4631]: I1128 13:23:51.960045 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Nov 28 13:23:51 crc kubenswrapper[4631]: I1128 13:23:51.960280 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Nov 28 13:23:51 crc kubenswrapper[4631]: I1128 13:23:51.961433 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Nov 28 13:23:51 crc kubenswrapper[4631]: I1128 13:23:51.961599 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Nov 28 13:23:51 crc kubenswrapper[4631]: I1128 13:23:51.963991 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Nov 28 13:23:51 crc kubenswrapper[4631]: I1128 13:23:51.963999 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Nov 28 13:23:51 crc kubenswrapper[4631]: I1128 13:23:51.975452 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Nov 28 13:23:51 crc kubenswrapper[4631]: I1128 13:23:51.977426 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Nov 28 13:23:51 crc kubenswrapper[4631]: I1128 13:23:51.982008 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Nov 28 13:23:51 crc kubenswrapper[4631]: I1128 13:23:51.985573 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Nov 28 13:23:51 crc kubenswrapper[4631]: I1128 13:23:51.992653 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-68b4d6ff9f-trblx"] Nov 28 13:23:52 crc kubenswrapper[4631]: I1128 13:23:52.004496 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Nov 28 13:23:52 crc kubenswrapper[4631]: I1128 13:23:52.040823 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/cae7f5a3-c3e4-442c-9b3b-e18bebc76b98-v4-0-config-system-service-ca\") pod \"oauth-openshift-68b4d6ff9f-trblx\" (UID: \"cae7f5a3-c3e4-442c-9b3b-e18bebc76b98\") " pod="openshift-authentication/oauth-openshift-68b4d6ff9f-trblx" Nov 28 13:23:52 crc kubenswrapper[4631]: I1128 13:23:52.040888 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/cae7f5a3-c3e4-442c-9b3b-e18bebc76b98-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-68b4d6ff9f-trblx\" (UID: \"cae7f5a3-c3e4-442c-9b3b-e18bebc76b98\") " pod="openshift-authentication/oauth-openshift-68b4d6ff9f-trblx" Nov 28 13:23:52 crc kubenswrapper[4631]: I1128 13:23:52.040922 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/cae7f5a3-c3e4-442c-9b3b-e18bebc76b98-v4-0-config-user-template-login\") pod \"oauth-openshift-68b4d6ff9f-trblx\" (UID: \"cae7f5a3-c3e4-442c-9b3b-e18bebc76b98\") " pod="openshift-authentication/oauth-openshift-68b4d6ff9f-trblx" Nov 28 13:23:52 crc kubenswrapper[4631]: I1128 13:23:52.040942 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/cae7f5a3-c3e4-442c-9b3b-e18bebc76b98-v4-0-config-system-router-certs\") pod \"oauth-openshift-68b4d6ff9f-trblx\" (UID: \"cae7f5a3-c3e4-442c-9b3b-e18bebc76b98\") " pod="openshift-authentication/oauth-openshift-68b4d6ff9f-trblx" Nov 28 13:23:52 crc kubenswrapper[4631]: I1128 13:23:52.041092 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/cae7f5a3-c3e4-442c-9b3b-e18bebc76b98-v4-0-config-system-session\") pod \"oauth-openshift-68b4d6ff9f-trblx\" (UID: \"cae7f5a3-c3e4-442c-9b3b-e18bebc76b98\") " pod="openshift-authentication/oauth-openshift-68b4d6ff9f-trblx" Nov 28 13:23:52 crc kubenswrapper[4631]: I1128 13:23:52.041154 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/cae7f5a3-c3e4-442c-9b3b-e18bebc76b98-v4-0-config-system-serving-cert\") pod \"oauth-openshift-68b4d6ff9f-trblx\" (UID: \"cae7f5a3-c3e4-442c-9b3b-e18bebc76b98\") " pod="openshift-authentication/oauth-openshift-68b4d6ff9f-trblx" Nov 28 13:23:52 crc kubenswrapper[4631]: I1128 13:23:52.041266 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/cae7f5a3-c3e4-442c-9b3b-e18bebc76b98-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-68b4d6ff9f-trblx\" (UID: \"cae7f5a3-c3e4-442c-9b3b-e18bebc76b98\") " pod="openshift-authentication/oauth-openshift-68b4d6ff9f-trblx" Nov 28 13:23:52 crc kubenswrapper[4631]: I1128 13:23:52.041325 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/cae7f5a3-c3e4-442c-9b3b-e18bebc76b98-v4-0-config-system-cliconfig\") pod \"oauth-openshift-68b4d6ff9f-trblx\" (UID: \"cae7f5a3-c3e4-442c-9b3b-e18bebc76b98\") " pod="openshift-authentication/oauth-openshift-68b4d6ff9f-trblx" Nov 28 13:23:52 crc kubenswrapper[4631]: I1128 13:23:52.041357 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/cae7f5a3-c3e4-442c-9b3b-e18bebc76b98-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-68b4d6ff9f-trblx\" (UID: \"cae7f5a3-c3e4-442c-9b3b-e18bebc76b98\") " pod="openshift-authentication/oauth-openshift-68b4d6ff9f-trblx" Nov 28 13:23:52 crc kubenswrapper[4631]: I1128 13:23:52.041416 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/cae7f5a3-c3e4-442c-9b3b-e18bebc76b98-audit-dir\") pod \"oauth-openshift-68b4d6ff9f-trblx\" (UID: \"cae7f5a3-c3e4-442c-9b3b-e18bebc76b98\") " pod="openshift-authentication/oauth-openshift-68b4d6ff9f-trblx" Nov 28 13:23:52 crc kubenswrapper[4631]: I1128 13:23:52.041487 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nhhjb\" (UniqueName: \"kubernetes.io/projected/cae7f5a3-c3e4-442c-9b3b-e18bebc76b98-kube-api-access-nhhjb\") pod \"oauth-openshift-68b4d6ff9f-trblx\" (UID: \"cae7f5a3-c3e4-442c-9b3b-e18bebc76b98\") " pod="openshift-authentication/oauth-openshift-68b4d6ff9f-trblx" Nov 28 13:23:52 crc kubenswrapper[4631]: I1128 13:23:52.041600 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/cae7f5a3-c3e4-442c-9b3b-e18bebc76b98-audit-policies\") pod \"oauth-openshift-68b4d6ff9f-trblx\" (UID: \"cae7f5a3-c3e4-442c-9b3b-e18bebc76b98\") " pod="openshift-authentication/oauth-openshift-68b4d6ff9f-trblx" Nov 28 13:23:52 crc kubenswrapper[4631]: I1128 13:23:52.041652 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/cae7f5a3-c3e4-442c-9b3b-e18bebc76b98-v4-0-config-user-template-error\") pod \"oauth-openshift-68b4d6ff9f-trblx\" (UID: \"cae7f5a3-c3e4-442c-9b3b-e18bebc76b98\") " pod="openshift-authentication/oauth-openshift-68b4d6ff9f-trblx" Nov 28 13:23:52 crc kubenswrapper[4631]: I1128 13:23:52.041714 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/cae7f5a3-c3e4-442c-9b3b-e18bebc76b98-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-68b4d6ff9f-trblx\" (UID: \"cae7f5a3-c3e4-442c-9b3b-e18bebc76b98\") " pod="openshift-authentication/oauth-openshift-68b4d6ff9f-trblx" Nov 28 13:23:52 crc kubenswrapper[4631]: I1128 13:23:52.143696 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nhhjb\" (UniqueName: \"kubernetes.io/projected/cae7f5a3-c3e4-442c-9b3b-e18bebc76b98-kube-api-access-nhhjb\") pod \"oauth-openshift-68b4d6ff9f-trblx\" (UID: \"cae7f5a3-c3e4-442c-9b3b-e18bebc76b98\") " pod="openshift-authentication/oauth-openshift-68b4d6ff9f-trblx" Nov 28 13:23:52 crc kubenswrapper[4631]: I1128 13:23:52.143785 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/cae7f5a3-c3e4-442c-9b3b-e18bebc76b98-audit-policies\") pod \"oauth-openshift-68b4d6ff9f-trblx\" (UID: \"cae7f5a3-c3e4-442c-9b3b-e18bebc76b98\") " pod="openshift-authentication/oauth-openshift-68b4d6ff9f-trblx" Nov 28 13:23:52 crc kubenswrapper[4631]: I1128 13:23:52.143817 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/cae7f5a3-c3e4-442c-9b3b-e18bebc76b98-v4-0-config-user-template-error\") pod \"oauth-openshift-68b4d6ff9f-trblx\" (UID: \"cae7f5a3-c3e4-442c-9b3b-e18bebc76b98\") " pod="openshift-authentication/oauth-openshift-68b4d6ff9f-trblx" Nov 28 13:23:52 crc kubenswrapper[4631]: I1128 13:23:52.143846 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/cae7f5a3-c3e4-442c-9b3b-e18bebc76b98-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-68b4d6ff9f-trblx\" (UID: \"cae7f5a3-c3e4-442c-9b3b-e18bebc76b98\") " pod="openshift-authentication/oauth-openshift-68b4d6ff9f-trblx" Nov 28 13:23:52 crc kubenswrapper[4631]: I1128 13:23:52.143872 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/cae7f5a3-c3e4-442c-9b3b-e18bebc76b98-v4-0-config-system-service-ca\") pod \"oauth-openshift-68b4d6ff9f-trblx\" (UID: \"cae7f5a3-c3e4-442c-9b3b-e18bebc76b98\") " pod="openshift-authentication/oauth-openshift-68b4d6ff9f-trblx" Nov 28 13:23:52 crc kubenswrapper[4631]: I1128 13:23:52.143894 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/cae7f5a3-c3e4-442c-9b3b-e18bebc76b98-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-68b4d6ff9f-trblx\" (UID: \"cae7f5a3-c3e4-442c-9b3b-e18bebc76b98\") " pod="openshift-authentication/oauth-openshift-68b4d6ff9f-trblx" Nov 28 13:23:52 crc kubenswrapper[4631]: I1128 13:23:52.143916 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/cae7f5a3-c3e4-442c-9b3b-e18bebc76b98-v4-0-config-user-template-login\") pod \"oauth-openshift-68b4d6ff9f-trblx\" (UID: \"cae7f5a3-c3e4-442c-9b3b-e18bebc76b98\") " pod="openshift-authentication/oauth-openshift-68b4d6ff9f-trblx" Nov 28 13:23:52 crc kubenswrapper[4631]: I1128 13:23:52.143936 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/cae7f5a3-c3e4-442c-9b3b-e18bebc76b98-v4-0-config-system-router-certs\") pod \"oauth-openshift-68b4d6ff9f-trblx\" (UID: \"cae7f5a3-c3e4-442c-9b3b-e18bebc76b98\") " pod="openshift-authentication/oauth-openshift-68b4d6ff9f-trblx" Nov 28 13:23:52 crc kubenswrapper[4631]: I1128 13:23:52.143973 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/cae7f5a3-c3e4-442c-9b3b-e18bebc76b98-v4-0-config-system-session\") pod \"oauth-openshift-68b4d6ff9f-trblx\" (UID: \"cae7f5a3-c3e4-442c-9b3b-e18bebc76b98\") " pod="openshift-authentication/oauth-openshift-68b4d6ff9f-trblx" Nov 28 13:23:52 crc kubenswrapper[4631]: I1128 13:23:52.143993 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/cae7f5a3-c3e4-442c-9b3b-e18bebc76b98-v4-0-config-system-serving-cert\") pod \"oauth-openshift-68b4d6ff9f-trblx\" (UID: \"cae7f5a3-c3e4-442c-9b3b-e18bebc76b98\") " pod="openshift-authentication/oauth-openshift-68b4d6ff9f-trblx" Nov 28 13:23:52 crc kubenswrapper[4631]: I1128 13:23:52.144020 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/cae7f5a3-c3e4-442c-9b3b-e18bebc76b98-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-68b4d6ff9f-trblx\" (UID: \"cae7f5a3-c3e4-442c-9b3b-e18bebc76b98\") " pod="openshift-authentication/oauth-openshift-68b4d6ff9f-trblx" Nov 28 13:23:52 crc kubenswrapper[4631]: I1128 13:23:52.144042 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/cae7f5a3-c3e4-442c-9b3b-e18bebc76b98-v4-0-config-system-cliconfig\") pod \"oauth-openshift-68b4d6ff9f-trblx\" (UID: \"cae7f5a3-c3e4-442c-9b3b-e18bebc76b98\") " pod="openshift-authentication/oauth-openshift-68b4d6ff9f-trblx" Nov 28 13:23:52 crc kubenswrapper[4631]: I1128 13:23:52.144062 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/cae7f5a3-c3e4-442c-9b3b-e18bebc76b98-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-68b4d6ff9f-trblx\" (UID: \"cae7f5a3-c3e4-442c-9b3b-e18bebc76b98\") " pod="openshift-authentication/oauth-openshift-68b4d6ff9f-trblx" Nov 28 13:23:52 crc kubenswrapper[4631]: I1128 13:23:52.144083 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/cae7f5a3-c3e4-442c-9b3b-e18bebc76b98-audit-dir\") pod \"oauth-openshift-68b4d6ff9f-trblx\" (UID: \"cae7f5a3-c3e4-442c-9b3b-e18bebc76b98\") " pod="openshift-authentication/oauth-openshift-68b4d6ff9f-trblx" Nov 28 13:23:52 crc kubenswrapper[4631]: I1128 13:23:52.144150 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/cae7f5a3-c3e4-442c-9b3b-e18bebc76b98-audit-dir\") pod \"oauth-openshift-68b4d6ff9f-trblx\" (UID: \"cae7f5a3-c3e4-442c-9b3b-e18bebc76b98\") " pod="openshift-authentication/oauth-openshift-68b4d6ff9f-trblx" Nov 28 13:23:52 crc kubenswrapper[4631]: I1128 13:23:52.144793 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/cae7f5a3-c3e4-442c-9b3b-e18bebc76b98-audit-policies\") pod \"oauth-openshift-68b4d6ff9f-trblx\" (UID: \"cae7f5a3-c3e4-442c-9b3b-e18bebc76b98\") " pod="openshift-authentication/oauth-openshift-68b4d6ff9f-trblx" Nov 28 13:23:52 crc kubenswrapper[4631]: I1128 13:23:52.144915 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/cae7f5a3-c3e4-442c-9b3b-e18bebc76b98-v4-0-config-system-service-ca\") pod \"oauth-openshift-68b4d6ff9f-trblx\" (UID: \"cae7f5a3-c3e4-442c-9b3b-e18bebc76b98\") " pod="openshift-authentication/oauth-openshift-68b4d6ff9f-trblx" Nov 28 13:23:52 crc kubenswrapper[4631]: I1128 13:23:52.145532 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/cae7f5a3-c3e4-442c-9b3b-e18bebc76b98-v4-0-config-system-cliconfig\") pod \"oauth-openshift-68b4d6ff9f-trblx\" (UID: \"cae7f5a3-c3e4-442c-9b3b-e18bebc76b98\") " pod="openshift-authentication/oauth-openshift-68b4d6ff9f-trblx" Nov 28 13:23:52 crc kubenswrapper[4631]: I1128 13:23:52.145640 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/cae7f5a3-c3e4-442c-9b3b-e18bebc76b98-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-68b4d6ff9f-trblx\" (UID: \"cae7f5a3-c3e4-442c-9b3b-e18bebc76b98\") " pod="openshift-authentication/oauth-openshift-68b4d6ff9f-trblx" Nov 28 13:23:52 crc kubenswrapper[4631]: I1128 13:23:52.151161 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/cae7f5a3-c3e4-442c-9b3b-e18bebc76b98-v4-0-config-system-router-certs\") pod \"oauth-openshift-68b4d6ff9f-trblx\" (UID: \"cae7f5a3-c3e4-442c-9b3b-e18bebc76b98\") " pod="openshift-authentication/oauth-openshift-68b4d6ff9f-trblx" Nov 28 13:23:52 crc kubenswrapper[4631]: I1128 13:23:52.153140 4631 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Nov 28 13:23:52 crc kubenswrapper[4631]: I1128 13:23:52.153363 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/cae7f5a3-c3e4-442c-9b3b-e18bebc76b98-v4-0-config-system-serving-cert\") pod \"oauth-openshift-68b4d6ff9f-trblx\" (UID: \"cae7f5a3-c3e4-442c-9b3b-e18bebc76b98\") " pod="openshift-authentication/oauth-openshift-68b4d6ff9f-trblx" Nov 28 13:23:52 crc kubenswrapper[4631]: I1128 13:23:52.154200 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/cae7f5a3-c3e4-442c-9b3b-e18bebc76b98-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-68b4d6ff9f-trblx\" (UID: \"cae7f5a3-c3e4-442c-9b3b-e18bebc76b98\") " pod="openshift-authentication/oauth-openshift-68b4d6ff9f-trblx" Nov 28 13:23:52 crc kubenswrapper[4631]: I1128 13:23:52.154283 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/cae7f5a3-c3e4-442c-9b3b-e18bebc76b98-v4-0-config-user-template-error\") pod \"oauth-openshift-68b4d6ff9f-trblx\" (UID: \"cae7f5a3-c3e4-442c-9b3b-e18bebc76b98\") " pod="openshift-authentication/oauth-openshift-68b4d6ff9f-trblx" Nov 28 13:23:52 crc kubenswrapper[4631]: I1128 13:23:52.155620 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/cae7f5a3-c3e4-442c-9b3b-e18bebc76b98-v4-0-config-system-session\") pod \"oauth-openshift-68b4d6ff9f-trblx\" (UID: \"cae7f5a3-c3e4-442c-9b3b-e18bebc76b98\") " pod="openshift-authentication/oauth-openshift-68b4d6ff9f-trblx" Nov 28 13:23:52 crc kubenswrapper[4631]: I1128 13:23:52.158049 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/cae7f5a3-c3e4-442c-9b3b-e18bebc76b98-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-68b4d6ff9f-trblx\" (UID: \"cae7f5a3-c3e4-442c-9b3b-e18bebc76b98\") " pod="openshift-authentication/oauth-openshift-68b4d6ff9f-trblx" Nov 28 13:23:52 crc kubenswrapper[4631]: I1128 13:23:52.159242 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/cae7f5a3-c3e4-442c-9b3b-e18bebc76b98-v4-0-config-user-template-login\") pod \"oauth-openshift-68b4d6ff9f-trblx\" (UID: \"cae7f5a3-c3e4-442c-9b3b-e18bebc76b98\") " pod="openshift-authentication/oauth-openshift-68b4d6ff9f-trblx" Nov 28 13:23:52 crc kubenswrapper[4631]: I1128 13:23:52.162554 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/cae7f5a3-c3e4-442c-9b3b-e18bebc76b98-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-68b4d6ff9f-trblx\" (UID: \"cae7f5a3-c3e4-442c-9b3b-e18bebc76b98\") " pod="openshift-authentication/oauth-openshift-68b4d6ff9f-trblx" Nov 28 13:23:52 crc kubenswrapper[4631]: I1128 13:23:52.165460 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nhhjb\" (UniqueName: \"kubernetes.io/projected/cae7f5a3-c3e4-442c-9b3b-e18bebc76b98-kube-api-access-nhhjb\") pod \"oauth-openshift-68b4d6ff9f-trblx\" (UID: \"cae7f5a3-c3e4-442c-9b3b-e18bebc76b98\") " pod="openshift-authentication/oauth-openshift-68b4d6ff9f-trblx" Nov 28 13:23:52 crc kubenswrapper[4631]: I1128 13:23:52.170458 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Nov 28 13:23:52 crc kubenswrapper[4631]: I1128 13:23:52.273732 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Nov 28 13:23:52 crc kubenswrapper[4631]: I1128 13:23:52.280691 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-68b4d6ff9f-trblx" Nov 28 13:23:52 crc kubenswrapper[4631]: I1128 13:23:52.326382 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Nov 28 13:23:52 crc kubenswrapper[4631]: I1128 13:23:52.363374 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Nov 28 13:23:52 crc kubenswrapper[4631]: I1128 13:23:52.420170 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Nov 28 13:23:52 crc kubenswrapper[4631]: I1128 13:23:52.520508 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Nov 28 13:23:52 crc kubenswrapper[4631]: I1128 13:23:52.533856 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Nov 28 13:23:52 crc kubenswrapper[4631]: I1128 13:23:52.564272 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Nov 28 13:23:52 crc kubenswrapper[4631]: I1128 13:23:52.573912 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Nov 28 13:23:52 crc kubenswrapper[4631]: I1128 13:23:52.588334 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-68b4d6ff9f-trblx"] Nov 28 13:23:52 crc kubenswrapper[4631]: I1128 13:23:52.588586 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Nov 28 13:23:52 crc kubenswrapper[4631]: I1128 13:23:52.616643 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Nov 28 13:23:52 crc kubenswrapper[4631]: I1128 13:23:52.628050 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Nov 28 13:23:52 crc kubenswrapper[4631]: I1128 13:23:52.719761 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Nov 28 13:23:52 crc kubenswrapper[4631]: I1128 13:23:52.778505 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Nov 28 13:23:52 crc kubenswrapper[4631]: I1128 13:23:52.929044 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Nov 28 13:23:52 crc kubenswrapper[4631]: I1128 13:23:52.963380 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-68b4d6ff9f-trblx" event={"ID":"cae7f5a3-c3e4-442c-9b3b-e18bebc76b98","Type":"ContainerStarted","Data":"35a975f8f913854fc5daa0f16349ea35804fca8645e91cace459906af150d777"} Nov 28 13:23:52 crc kubenswrapper[4631]: I1128 13:23:52.963445 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-68b4d6ff9f-trblx" event={"ID":"cae7f5a3-c3e4-442c-9b3b-e18bebc76b98","Type":"ContainerStarted","Data":"a7b1761a6250e095c42a5a98f76ee22fb71cf77ee023b16a991ecb339d26e8e6"} Nov 28 13:23:52 crc kubenswrapper[4631]: I1128 13:23:52.965178 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-68b4d6ff9f-trblx" Nov 28 13:23:52 crc kubenswrapper[4631]: I1128 13:23:52.986776 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-68b4d6ff9f-trblx" podStartSLOduration=64.98675253 podStartE2EDuration="1m4.98675253s" podCreationTimestamp="2025-11-28 13:22:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 13:23:52.986316039 +0000 UTC m=+189.793619383" watchObservedRunningTime="2025-11-28 13:23:52.98675253 +0000 UTC m=+189.794055874" Nov 28 13:23:53 crc kubenswrapper[4631]: I1128 13:23:53.010907 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Nov 28 13:23:53 crc kubenswrapper[4631]: I1128 13:23:53.053491 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Nov 28 13:23:53 crc kubenswrapper[4631]: I1128 13:23:53.284326 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Nov 28 13:23:53 crc kubenswrapper[4631]: I1128 13:23:53.489121 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-68b4d6ff9f-trblx" Nov 28 13:23:53 crc kubenswrapper[4631]: I1128 13:23:53.498843 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Nov 28 13:23:53 crc kubenswrapper[4631]: I1128 13:23:53.597728 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Nov 28 13:23:53 crc kubenswrapper[4631]: I1128 13:23:53.605491 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Nov 28 13:23:53 crc kubenswrapper[4631]: I1128 13:23:53.738523 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Nov 28 13:23:54 crc kubenswrapper[4631]: I1128 13:23:54.038190 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Nov 28 13:23:54 crc kubenswrapper[4631]: I1128 13:23:54.136540 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Nov 28 13:23:54 crc kubenswrapper[4631]: I1128 13:23:54.485813 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Nov 28 13:23:55 crc kubenswrapper[4631]: I1128 13:23:55.988690 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Nov 28 13:23:55 crc kubenswrapper[4631]: I1128 13:23:55.989203 4631 generic.go:334] "Generic (PLEG): container finished" podID="f85e55b1a89d02b0cb034b1ea31ed45a" containerID="1c53926191bf67e68a0e0230740de428c280f79550fd964cf4bef72a592a0f57" exitCode=137 Nov 28 13:23:55 crc kubenswrapper[4631]: I1128 13:23:55.989253 4631 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="df8a6455cd365b09cdc2fd7bcf56227708d1ec6845bb813aec03725198478e44" Nov 28 13:23:56 crc kubenswrapper[4631]: I1128 13:23:56.022558 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Nov 28 13:23:56 crc kubenswrapper[4631]: I1128 13:23:56.022684 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 28 13:23:56 crc kubenswrapper[4631]: I1128 13:23:56.114183 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Nov 28 13:23:56 crc kubenswrapper[4631]: I1128 13:23:56.114370 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests" (OuterVolumeSpecName: "manifests") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "manifests". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 13:23:56 crc kubenswrapper[4631]: I1128 13:23:56.114746 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Nov 28 13:23:56 crc kubenswrapper[4631]: I1128 13:23:56.114862 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Nov 28 13:23:56 crc kubenswrapper[4631]: I1128 13:23:56.114940 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Nov 28 13:23:56 crc kubenswrapper[4631]: I1128 13:23:56.115039 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Nov 28 13:23:56 crc kubenswrapper[4631]: I1128 13:23:56.115059 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock" (OuterVolumeSpecName: "var-lock") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 13:23:56 crc kubenswrapper[4631]: I1128 13:23:56.115107 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log" (OuterVolumeSpecName: "var-log") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 13:23:56 crc kubenswrapper[4631]: I1128 13:23:56.115335 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 13:23:56 crc kubenswrapper[4631]: I1128 13:23:56.115651 4631 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") on node \"crc\" DevicePath \"\"" Nov 28 13:23:56 crc kubenswrapper[4631]: I1128 13:23:56.115773 4631 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") on node \"crc\" DevicePath \"\"" Nov 28 13:23:56 crc kubenswrapper[4631]: I1128 13:23:56.115876 4631 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") on node \"crc\" DevicePath \"\"" Nov 28 13:23:56 crc kubenswrapper[4631]: I1128 13:23:56.115996 4631 reconciler_common.go:293] "Volume detached for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") on node \"crc\" DevicePath \"\"" Nov 28 13:23:56 crc kubenswrapper[4631]: I1128 13:23:56.128546 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir" (OuterVolumeSpecName: "pod-resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "pod-resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 13:23:56 crc kubenswrapper[4631]: I1128 13:23:56.217579 4631 reconciler_common.go:293] "Volume detached for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") on node \"crc\" DevicePath \"\"" Nov 28 13:23:56 crc kubenswrapper[4631]: I1128 13:23:56.994517 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 28 13:23:57 crc kubenswrapper[4631]: I1128 13:23:57.521402 4631 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" path="/var/lib/kubelet/pods/f85e55b1a89d02b0cb034b1ea31ed45a/volumes" Nov 28 13:23:57 crc kubenswrapper[4631]: I1128 13:23:57.522104 4631 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="" Nov 28 13:23:57 crc kubenswrapper[4631]: I1128 13:23:57.533237 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Nov 28 13:23:57 crc kubenswrapper[4631]: I1128 13:23:57.533298 4631 kubelet.go:2649] "Unable to find pod for mirror pod, skipping" mirrorPod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" mirrorPodUID="fd05b9db-dd4b-492e-a4ae-e8a4cc636858" Nov 28 13:23:57 crc kubenswrapper[4631]: I1128 13:23:57.537851 4631 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Nov 28 13:23:57 crc kubenswrapper[4631]: I1128 13:23:57.537905 4631 kubelet.go:2673] "Unable to find pod for mirror pod, skipping" mirrorPod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" mirrorPodUID="fd05b9db-dd4b-492e-a4ae-e8a4cc636858" Nov 28 13:24:19 crc kubenswrapper[4631]: I1128 13:24:19.635401 4631 patch_prober.go:28] interesting pod/machine-config-daemon-47sxw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 28 13:24:19 crc kubenswrapper[4631]: I1128 13:24:19.636269 4631 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" podUID="8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 28 13:24:19 crc kubenswrapper[4631]: I1128 13:24:19.636992 4631 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" Nov 28 13:24:19 crc kubenswrapper[4631]: I1128 13:24:19.638011 4631 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"7f613562449a8f5ebf17cd881967c47d44fb0296171712af40f42263f69f8136"} pod="openshift-machine-config-operator/machine-config-daemon-47sxw" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 28 13:24:19 crc kubenswrapper[4631]: I1128 13:24:19.638278 4631 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" podUID="8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1" containerName="machine-config-daemon" containerID="cri-o://7f613562449a8f5ebf17cd881967c47d44fb0296171712af40f42263f69f8136" gracePeriod=600 Nov 28 13:24:20 crc kubenswrapper[4631]: I1128 13:24:20.507150 4631 generic.go:334] "Generic (PLEG): container finished" podID="8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1" containerID="7f613562449a8f5ebf17cd881967c47d44fb0296171712af40f42263f69f8136" exitCode=0 Nov 28 13:24:20 crc kubenswrapper[4631]: I1128 13:24:20.507583 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" event={"ID":"8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1","Type":"ContainerDied","Data":"7f613562449a8f5ebf17cd881967c47d44fb0296171712af40f42263f69f8136"} Nov 28 13:24:21 crc kubenswrapper[4631]: I1128 13:24:21.522752 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" event={"ID":"8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1","Type":"ContainerStarted","Data":"b04ff78ff7998dc8957c0acf1246dd6adabf1e8af15924c3953943c9bdb3150a"} Nov 28 13:24:30 crc kubenswrapper[4631]: I1128 13:24:30.125579 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-7gmxj"] Nov 28 13:24:30 crc kubenswrapper[4631]: I1128 13:24:30.126955 4631 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-7gmxj" podUID="e7c7bd85-4add-4bda-a4fa-71a7b75787e4" containerName="controller-manager" containerID="cri-o://bd92fabfc49e57b88c4ce7a4b6957c3af7dfd28d76def445487a97dbd2876afc" gracePeriod=30 Nov 28 13:24:30 crc kubenswrapper[4631]: I1128 13:24:30.222173 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-ddx6v"] Nov 28 13:24:30 crc kubenswrapper[4631]: I1128 13:24:30.222407 4631 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ddx6v" podUID="594fba15-6b93-4f34-af64-403a1b2d694c" containerName="route-controller-manager" containerID="cri-o://e27b111a3dc6ae2d3a462faffe6d2e2109a2746588eb8eb6745f2f3c59669f45" gracePeriod=30 Nov 28 13:24:30 crc kubenswrapper[4631]: I1128 13:24:30.571083 4631 generic.go:334] "Generic (PLEG): container finished" podID="594fba15-6b93-4f34-af64-403a1b2d694c" containerID="e27b111a3dc6ae2d3a462faffe6d2e2109a2746588eb8eb6745f2f3c59669f45" exitCode=0 Nov 28 13:24:30 crc kubenswrapper[4631]: I1128 13:24:30.571175 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ddx6v" event={"ID":"594fba15-6b93-4f34-af64-403a1b2d694c","Type":"ContainerDied","Data":"e27b111a3dc6ae2d3a462faffe6d2e2109a2746588eb8eb6745f2f3c59669f45"} Nov 28 13:24:30 crc kubenswrapper[4631]: I1128 13:24:30.572996 4631 generic.go:334] "Generic (PLEG): container finished" podID="e7c7bd85-4add-4bda-a4fa-71a7b75787e4" containerID="bd92fabfc49e57b88c4ce7a4b6957c3af7dfd28d76def445487a97dbd2876afc" exitCode=0 Nov 28 13:24:30 crc kubenswrapper[4631]: I1128 13:24:30.573029 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-7gmxj" event={"ID":"e7c7bd85-4add-4bda-a4fa-71a7b75787e4","Type":"ContainerDied","Data":"bd92fabfc49e57b88c4ce7a4b6957c3af7dfd28d76def445487a97dbd2876afc"} Nov 28 13:24:31 crc kubenswrapper[4631]: I1128 13:24:31.854735 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-7gmxj" Nov 28 13:24:31 crc kubenswrapper[4631]: I1128 13:24:31.896407 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-66d87f89f6-5g9jj"] Nov 28 13:24:31 crc kubenswrapper[4631]: E1128 13:24:31.896715 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e7c7bd85-4add-4bda-a4fa-71a7b75787e4" containerName="controller-manager" Nov 28 13:24:31 crc kubenswrapper[4631]: I1128 13:24:31.896733 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="e7c7bd85-4add-4bda-a4fa-71a7b75787e4" containerName="controller-manager" Nov 28 13:24:31 crc kubenswrapper[4631]: I1128 13:24:31.896855 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="e7c7bd85-4add-4bda-a4fa-71a7b75787e4" containerName="controller-manager" Nov 28 13:24:31 crc kubenswrapper[4631]: I1128 13:24:31.897440 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-66d87f89f6-5g9jj" Nov 28 13:24:31 crc kubenswrapper[4631]: I1128 13:24:31.911812 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7c7bd85-4add-4bda-a4fa-71a7b75787e4-serving-cert\") pod \"e7c7bd85-4add-4bda-a4fa-71a7b75787e4\" (UID: \"e7c7bd85-4add-4bda-a4fa-71a7b75787e4\") " Nov 28 13:24:31 crc kubenswrapper[4631]: I1128 13:24:31.911953 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7c7bd85-4add-4bda-a4fa-71a7b75787e4-config\") pod \"e7c7bd85-4add-4bda-a4fa-71a7b75787e4\" (UID: \"e7c7bd85-4add-4bda-a4fa-71a7b75787e4\") " Nov 28 13:24:31 crc kubenswrapper[4631]: I1128 13:24:31.912011 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e7c7bd85-4add-4bda-a4fa-71a7b75787e4-proxy-ca-bundles\") pod \"e7c7bd85-4add-4bda-a4fa-71a7b75787e4\" (UID: \"e7c7bd85-4add-4bda-a4fa-71a7b75787e4\") " Nov 28 13:24:31 crc kubenswrapper[4631]: I1128 13:24:31.912072 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d8jfw\" (UniqueName: \"kubernetes.io/projected/e7c7bd85-4add-4bda-a4fa-71a7b75787e4-kube-api-access-d8jfw\") pod \"e7c7bd85-4add-4bda-a4fa-71a7b75787e4\" (UID: \"e7c7bd85-4add-4bda-a4fa-71a7b75787e4\") " Nov 28 13:24:31 crc kubenswrapper[4631]: I1128 13:24:31.912119 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e7c7bd85-4add-4bda-a4fa-71a7b75787e4-client-ca\") pod \"e7c7bd85-4add-4bda-a4fa-71a7b75787e4\" (UID: \"e7c7bd85-4add-4bda-a4fa-71a7b75787e4\") " Nov 28 13:24:31 crc kubenswrapper[4631]: I1128 13:24:31.912386 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a2c8105b-a4d6-487c-af19-668a70b2964e-serving-cert\") pod \"controller-manager-66d87f89f6-5g9jj\" (UID: \"a2c8105b-a4d6-487c-af19-668a70b2964e\") " pod="openshift-controller-manager/controller-manager-66d87f89f6-5g9jj" Nov 28 13:24:31 crc kubenswrapper[4631]: I1128 13:24:31.912471 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a2c8105b-a4d6-487c-af19-668a70b2964e-config\") pod \"controller-manager-66d87f89f6-5g9jj\" (UID: \"a2c8105b-a4d6-487c-af19-668a70b2964e\") " pod="openshift-controller-manager/controller-manager-66d87f89f6-5g9jj" Nov 28 13:24:31 crc kubenswrapper[4631]: I1128 13:24:31.912504 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a2c8105b-a4d6-487c-af19-668a70b2964e-client-ca\") pod \"controller-manager-66d87f89f6-5g9jj\" (UID: \"a2c8105b-a4d6-487c-af19-668a70b2964e\") " pod="openshift-controller-manager/controller-manager-66d87f89f6-5g9jj" Nov 28 13:24:31 crc kubenswrapper[4631]: I1128 13:24:31.912535 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a2c8105b-a4d6-487c-af19-668a70b2964e-proxy-ca-bundles\") pod \"controller-manager-66d87f89f6-5g9jj\" (UID: \"a2c8105b-a4d6-487c-af19-668a70b2964e\") " pod="openshift-controller-manager/controller-manager-66d87f89f6-5g9jj" Nov 28 13:24:31 crc kubenswrapper[4631]: I1128 13:24:31.912581 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pzh9l\" (UniqueName: \"kubernetes.io/projected/a2c8105b-a4d6-487c-af19-668a70b2964e-kube-api-access-pzh9l\") pod \"controller-manager-66d87f89f6-5g9jj\" (UID: \"a2c8105b-a4d6-487c-af19-668a70b2964e\") " pod="openshift-controller-manager/controller-manager-66d87f89f6-5g9jj" Nov 28 13:24:31 crc kubenswrapper[4631]: I1128 13:24:31.914859 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7c7bd85-4add-4bda-a4fa-71a7b75787e4-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "e7c7bd85-4add-4bda-a4fa-71a7b75787e4" (UID: "e7c7bd85-4add-4bda-a4fa-71a7b75787e4"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 13:24:31 crc kubenswrapper[4631]: I1128 13:24:31.914885 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7c7bd85-4add-4bda-a4fa-71a7b75787e4-client-ca" (OuterVolumeSpecName: "client-ca") pod "e7c7bd85-4add-4bda-a4fa-71a7b75787e4" (UID: "e7c7bd85-4add-4bda-a4fa-71a7b75787e4"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 13:24:31 crc kubenswrapper[4631]: I1128 13:24:31.915345 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7c7bd85-4add-4bda-a4fa-71a7b75787e4-config" (OuterVolumeSpecName: "config") pod "e7c7bd85-4add-4bda-a4fa-71a7b75787e4" (UID: "e7c7bd85-4add-4bda-a4fa-71a7b75787e4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 13:24:31 crc kubenswrapper[4631]: I1128 13:24:31.928887 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7c7bd85-4add-4bda-a4fa-71a7b75787e4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7c7bd85-4add-4bda-a4fa-71a7b75787e4" (UID: "e7c7bd85-4add-4bda-a4fa-71a7b75787e4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:24:31 crc kubenswrapper[4631]: I1128 13:24:31.929449 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7c7bd85-4add-4bda-a4fa-71a7b75787e4-kube-api-access-d8jfw" (OuterVolumeSpecName: "kube-api-access-d8jfw") pod "e7c7bd85-4add-4bda-a4fa-71a7b75787e4" (UID: "e7c7bd85-4add-4bda-a4fa-71a7b75787e4"). InnerVolumeSpecName "kube-api-access-d8jfw". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:24:31 crc kubenswrapper[4631]: I1128 13:24:31.938145 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-66d87f89f6-5g9jj"] Nov 28 13:24:31 crc kubenswrapper[4631]: I1128 13:24:31.983366 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ddx6v" Nov 28 13:24:32 crc kubenswrapper[4631]: I1128 13:24:32.013269 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mfgnq\" (UniqueName: \"kubernetes.io/projected/594fba15-6b93-4f34-af64-403a1b2d694c-kube-api-access-mfgnq\") pod \"594fba15-6b93-4f34-af64-403a1b2d694c\" (UID: \"594fba15-6b93-4f34-af64-403a1b2d694c\") " Nov 28 13:24:32 crc kubenswrapper[4631]: I1128 13:24:32.013422 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/594fba15-6b93-4f34-af64-403a1b2d694c-config\") pod \"594fba15-6b93-4f34-af64-403a1b2d694c\" (UID: \"594fba15-6b93-4f34-af64-403a1b2d694c\") " Nov 28 13:24:32 crc kubenswrapper[4631]: I1128 13:24:32.013542 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/594fba15-6b93-4f34-af64-403a1b2d694c-client-ca\") pod \"594fba15-6b93-4f34-af64-403a1b2d694c\" (UID: \"594fba15-6b93-4f34-af64-403a1b2d694c\") " Nov 28 13:24:32 crc kubenswrapper[4631]: I1128 13:24:32.013638 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/594fba15-6b93-4f34-af64-403a1b2d694c-serving-cert\") pod \"594fba15-6b93-4f34-af64-403a1b2d694c\" (UID: \"594fba15-6b93-4f34-af64-403a1b2d694c\") " Nov 28 13:24:32 crc kubenswrapper[4631]: I1128 13:24:32.013947 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a2c8105b-a4d6-487c-af19-668a70b2964e-proxy-ca-bundles\") pod \"controller-manager-66d87f89f6-5g9jj\" (UID: \"a2c8105b-a4d6-487c-af19-668a70b2964e\") " pod="openshift-controller-manager/controller-manager-66d87f89f6-5g9jj" Nov 28 13:24:32 crc kubenswrapper[4631]: I1128 13:24:32.014015 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pzh9l\" (UniqueName: \"kubernetes.io/projected/a2c8105b-a4d6-487c-af19-668a70b2964e-kube-api-access-pzh9l\") pod \"controller-manager-66d87f89f6-5g9jj\" (UID: \"a2c8105b-a4d6-487c-af19-668a70b2964e\") " pod="openshift-controller-manager/controller-manager-66d87f89f6-5g9jj" Nov 28 13:24:32 crc kubenswrapper[4631]: I1128 13:24:32.014064 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a2c8105b-a4d6-487c-af19-668a70b2964e-serving-cert\") pod \"controller-manager-66d87f89f6-5g9jj\" (UID: \"a2c8105b-a4d6-487c-af19-668a70b2964e\") " pod="openshift-controller-manager/controller-manager-66d87f89f6-5g9jj" Nov 28 13:24:32 crc kubenswrapper[4631]: I1128 13:24:32.014131 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a2c8105b-a4d6-487c-af19-668a70b2964e-config\") pod \"controller-manager-66d87f89f6-5g9jj\" (UID: \"a2c8105b-a4d6-487c-af19-668a70b2964e\") " pod="openshift-controller-manager/controller-manager-66d87f89f6-5g9jj" Nov 28 13:24:32 crc kubenswrapper[4631]: I1128 13:24:32.014168 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a2c8105b-a4d6-487c-af19-668a70b2964e-client-ca\") pod \"controller-manager-66d87f89f6-5g9jj\" (UID: \"a2c8105b-a4d6-487c-af19-668a70b2964e\") " pod="openshift-controller-manager/controller-manager-66d87f89f6-5g9jj" Nov 28 13:24:32 crc kubenswrapper[4631]: I1128 13:24:32.014253 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d8jfw\" (UniqueName: \"kubernetes.io/projected/e7c7bd85-4add-4bda-a4fa-71a7b75787e4-kube-api-access-d8jfw\") on node \"crc\" DevicePath \"\"" Nov 28 13:24:32 crc kubenswrapper[4631]: I1128 13:24:32.014277 4631 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e7c7bd85-4add-4bda-a4fa-71a7b75787e4-client-ca\") on node \"crc\" DevicePath \"\"" Nov 28 13:24:32 crc kubenswrapper[4631]: I1128 13:24:32.014310 4631 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7c7bd85-4add-4bda-a4fa-71a7b75787e4-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 28 13:24:32 crc kubenswrapper[4631]: I1128 13:24:32.014327 4631 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7c7bd85-4add-4bda-a4fa-71a7b75787e4-config\") on node \"crc\" DevicePath \"\"" Nov 28 13:24:32 crc kubenswrapper[4631]: I1128 13:24:32.014345 4631 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e7c7bd85-4add-4bda-a4fa-71a7b75787e4-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Nov 28 13:24:32 crc kubenswrapper[4631]: I1128 13:24:32.015776 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a2c8105b-a4d6-487c-af19-668a70b2964e-client-ca\") pod \"controller-manager-66d87f89f6-5g9jj\" (UID: \"a2c8105b-a4d6-487c-af19-668a70b2964e\") " pod="openshift-controller-manager/controller-manager-66d87f89f6-5g9jj" Nov 28 13:24:32 crc kubenswrapper[4631]: I1128 13:24:32.014101 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/594fba15-6b93-4f34-af64-403a1b2d694c-config" (OuterVolumeSpecName: "config") pod "594fba15-6b93-4f34-af64-403a1b2d694c" (UID: "594fba15-6b93-4f34-af64-403a1b2d694c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 13:24:32 crc kubenswrapper[4631]: I1128 13:24:32.014908 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/594fba15-6b93-4f34-af64-403a1b2d694c-client-ca" (OuterVolumeSpecName: "client-ca") pod "594fba15-6b93-4f34-af64-403a1b2d694c" (UID: "594fba15-6b93-4f34-af64-403a1b2d694c"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 13:24:32 crc kubenswrapper[4631]: I1128 13:24:32.016113 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a2c8105b-a4d6-487c-af19-668a70b2964e-config\") pod \"controller-manager-66d87f89f6-5g9jj\" (UID: \"a2c8105b-a4d6-487c-af19-668a70b2964e\") " pod="openshift-controller-manager/controller-manager-66d87f89f6-5g9jj" Nov 28 13:24:32 crc kubenswrapper[4631]: I1128 13:24:32.018544 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a2c8105b-a4d6-487c-af19-668a70b2964e-proxy-ca-bundles\") pod \"controller-manager-66d87f89f6-5g9jj\" (UID: \"a2c8105b-a4d6-487c-af19-668a70b2964e\") " pod="openshift-controller-manager/controller-manager-66d87f89f6-5g9jj" Nov 28 13:24:32 crc kubenswrapper[4631]: I1128 13:24:32.023104 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/594fba15-6b93-4f34-af64-403a1b2d694c-kube-api-access-mfgnq" (OuterVolumeSpecName: "kube-api-access-mfgnq") pod "594fba15-6b93-4f34-af64-403a1b2d694c" (UID: "594fba15-6b93-4f34-af64-403a1b2d694c"). InnerVolumeSpecName "kube-api-access-mfgnq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:24:32 crc kubenswrapper[4631]: I1128 13:24:32.024061 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a2c8105b-a4d6-487c-af19-668a70b2964e-serving-cert\") pod \"controller-manager-66d87f89f6-5g9jj\" (UID: \"a2c8105b-a4d6-487c-af19-668a70b2964e\") " pod="openshift-controller-manager/controller-manager-66d87f89f6-5g9jj" Nov 28 13:24:32 crc kubenswrapper[4631]: I1128 13:24:32.024660 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/594fba15-6b93-4f34-af64-403a1b2d694c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "594fba15-6b93-4f34-af64-403a1b2d694c" (UID: "594fba15-6b93-4f34-af64-403a1b2d694c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:24:32 crc kubenswrapper[4631]: I1128 13:24:32.036789 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pzh9l\" (UniqueName: \"kubernetes.io/projected/a2c8105b-a4d6-487c-af19-668a70b2964e-kube-api-access-pzh9l\") pod \"controller-manager-66d87f89f6-5g9jj\" (UID: \"a2c8105b-a4d6-487c-af19-668a70b2964e\") " pod="openshift-controller-manager/controller-manager-66d87f89f6-5g9jj" Nov 28 13:24:32 crc kubenswrapper[4631]: I1128 13:24:32.115820 4631 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/594fba15-6b93-4f34-af64-403a1b2d694c-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 28 13:24:32 crc kubenswrapper[4631]: I1128 13:24:32.115861 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mfgnq\" (UniqueName: \"kubernetes.io/projected/594fba15-6b93-4f34-af64-403a1b2d694c-kube-api-access-mfgnq\") on node \"crc\" DevicePath \"\"" Nov 28 13:24:32 crc kubenswrapper[4631]: I1128 13:24:32.115878 4631 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/594fba15-6b93-4f34-af64-403a1b2d694c-config\") on node \"crc\" DevicePath \"\"" Nov 28 13:24:32 crc kubenswrapper[4631]: I1128 13:24:32.115889 4631 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/594fba15-6b93-4f34-af64-403a1b2d694c-client-ca\") on node \"crc\" DevicePath \"\"" Nov 28 13:24:32 crc kubenswrapper[4631]: I1128 13:24:32.279851 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-66d87f89f6-5g9jj" Nov 28 13:24:32 crc kubenswrapper[4631]: I1128 13:24:32.531037 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-66d87f89f6-5g9jj"] Nov 28 13:24:32 crc kubenswrapper[4631]: I1128 13:24:32.610263 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ddx6v" event={"ID":"594fba15-6b93-4f34-af64-403a1b2d694c","Type":"ContainerDied","Data":"eca0b8da471872c541822e9b1c0a121260a0a99163c68ca217f7989ad15d52e0"} Nov 28 13:24:32 crc kubenswrapper[4631]: I1128 13:24:32.610359 4631 scope.go:117] "RemoveContainer" containerID="e27b111a3dc6ae2d3a462faffe6d2e2109a2746588eb8eb6745f2f3c59669f45" Nov 28 13:24:32 crc kubenswrapper[4631]: I1128 13:24:32.610422 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ddx6v" Nov 28 13:24:32 crc kubenswrapper[4631]: I1128 13:24:32.613245 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-7gmxj" event={"ID":"e7c7bd85-4add-4bda-a4fa-71a7b75787e4","Type":"ContainerDied","Data":"807704f91ff09049c560dee0013b23bfdbc5fb270c9483b032886f494feb9fd7"} Nov 28 13:24:32 crc kubenswrapper[4631]: I1128 13:24:32.613378 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-7gmxj" Nov 28 13:24:32 crc kubenswrapper[4631]: I1128 13:24:32.614310 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-66d87f89f6-5g9jj" event={"ID":"a2c8105b-a4d6-487c-af19-668a70b2964e","Type":"ContainerStarted","Data":"7347b8f46aeeaf3fce3305864d4e1d3c36edfe7b6e6aee003c854230e436fdcd"} Nov 28 13:24:32 crc kubenswrapper[4631]: I1128 13:24:32.658082 4631 scope.go:117] "RemoveContainer" containerID="bd92fabfc49e57b88c4ce7a4b6957c3af7dfd28d76def445487a97dbd2876afc" Nov 28 13:24:32 crc kubenswrapper[4631]: I1128 13:24:32.688345 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-ddx6v"] Nov 28 13:24:32 crc kubenswrapper[4631]: I1128 13:24:32.691323 4631 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-ddx6v"] Nov 28 13:24:32 crc kubenswrapper[4631]: I1128 13:24:32.695562 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-7gmxj"] Nov 28 13:24:32 crc kubenswrapper[4631]: I1128 13:24:32.698677 4631 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-7gmxj"] Nov 28 13:24:33 crc kubenswrapper[4631]: I1128 13:24:33.522113 4631 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="594fba15-6b93-4f34-af64-403a1b2d694c" path="/var/lib/kubelet/pods/594fba15-6b93-4f34-af64-403a1b2d694c/volumes" Nov 28 13:24:33 crc kubenswrapper[4631]: I1128 13:24:33.524652 4631 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7c7bd85-4add-4bda-a4fa-71a7b75787e4" path="/var/lib/kubelet/pods/e7c7bd85-4add-4bda-a4fa-71a7b75787e4/volumes" Nov 28 13:24:33 crc kubenswrapper[4631]: I1128 13:24:33.623713 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-66d87f89f6-5g9jj" event={"ID":"a2c8105b-a4d6-487c-af19-668a70b2964e","Type":"ContainerStarted","Data":"1d2dad314463a6bd66baf8e0493fd1f25a1f7b6d10ed1e900fda55f0fb71961b"} Nov 28 13:24:33 crc kubenswrapper[4631]: I1128 13:24:33.624601 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-66d87f89f6-5g9jj" Nov 28 13:24:33 crc kubenswrapper[4631]: I1128 13:24:33.630426 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-66d87f89f6-5g9jj" Nov 28 13:24:33 crc kubenswrapper[4631]: I1128 13:24:33.648579 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-66d87f89f6-5g9jj" podStartSLOduration=3.648553948 podStartE2EDuration="3.648553948s" podCreationTimestamp="2025-11-28 13:24:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 13:24:33.647761608 +0000 UTC m=+230.455064962" watchObservedRunningTime="2025-11-28 13:24:33.648553948 +0000 UTC m=+230.455857292" Nov 28 13:24:33 crc kubenswrapper[4631]: I1128 13:24:33.981927 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-bbff9c674-sjkrj"] Nov 28 13:24:33 crc kubenswrapper[4631]: E1128 13:24:33.982266 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="594fba15-6b93-4f34-af64-403a1b2d694c" containerName="route-controller-manager" Nov 28 13:24:33 crc kubenswrapper[4631]: I1128 13:24:33.982304 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="594fba15-6b93-4f34-af64-403a1b2d694c" containerName="route-controller-manager" Nov 28 13:24:33 crc kubenswrapper[4631]: I1128 13:24:33.982449 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="594fba15-6b93-4f34-af64-403a1b2d694c" containerName="route-controller-manager" Nov 28 13:24:33 crc kubenswrapper[4631]: I1128 13:24:33.983036 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-bbff9c674-sjkrj" Nov 28 13:24:33 crc kubenswrapper[4631]: I1128 13:24:33.986963 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Nov 28 13:24:33 crc kubenswrapper[4631]: I1128 13:24:33.987129 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Nov 28 13:24:33 crc kubenswrapper[4631]: I1128 13:24:33.987786 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Nov 28 13:24:33 crc kubenswrapper[4631]: I1128 13:24:33.990170 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Nov 28 13:24:33 crc kubenswrapper[4631]: I1128 13:24:33.992364 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Nov 28 13:24:33 crc kubenswrapper[4631]: I1128 13:24:33.994706 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Nov 28 13:24:34 crc kubenswrapper[4631]: I1128 13:24:34.001672 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-bbff9c674-sjkrj"] Nov 28 13:24:34 crc kubenswrapper[4631]: I1128 13:24:34.053543 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8a0e31f0-4c01-4e0a-b0ac-49f731ef26e3-serving-cert\") pod \"route-controller-manager-bbff9c674-sjkrj\" (UID: \"8a0e31f0-4c01-4e0a-b0ac-49f731ef26e3\") " pod="openshift-route-controller-manager/route-controller-manager-bbff9c674-sjkrj" Nov 28 13:24:34 crc kubenswrapper[4631]: I1128 13:24:34.053605 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sgv2h\" (UniqueName: \"kubernetes.io/projected/8a0e31f0-4c01-4e0a-b0ac-49f731ef26e3-kube-api-access-sgv2h\") pod \"route-controller-manager-bbff9c674-sjkrj\" (UID: \"8a0e31f0-4c01-4e0a-b0ac-49f731ef26e3\") " pod="openshift-route-controller-manager/route-controller-manager-bbff9c674-sjkrj" Nov 28 13:24:34 crc kubenswrapper[4631]: I1128 13:24:34.053647 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8a0e31f0-4c01-4e0a-b0ac-49f731ef26e3-config\") pod \"route-controller-manager-bbff9c674-sjkrj\" (UID: \"8a0e31f0-4c01-4e0a-b0ac-49f731ef26e3\") " pod="openshift-route-controller-manager/route-controller-manager-bbff9c674-sjkrj" Nov 28 13:24:34 crc kubenswrapper[4631]: I1128 13:24:34.053716 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8a0e31f0-4c01-4e0a-b0ac-49f731ef26e3-client-ca\") pod \"route-controller-manager-bbff9c674-sjkrj\" (UID: \"8a0e31f0-4c01-4e0a-b0ac-49f731ef26e3\") " pod="openshift-route-controller-manager/route-controller-manager-bbff9c674-sjkrj" Nov 28 13:24:34 crc kubenswrapper[4631]: I1128 13:24:34.155614 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8a0e31f0-4c01-4e0a-b0ac-49f731ef26e3-config\") pod \"route-controller-manager-bbff9c674-sjkrj\" (UID: \"8a0e31f0-4c01-4e0a-b0ac-49f731ef26e3\") " pod="openshift-route-controller-manager/route-controller-manager-bbff9c674-sjkrj" Nov 28 13:24:34 crc kubenswrapper[4631]: I1128 13:24:34.156114 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8a0e31f0-4c01-4e0a-b0ac-49f731ef26e3-client-ca\") pod \"route-controller-manager-bbff9c674-sjkrj\" (UID: \"8a0e31f0-4c01-4e0a-b0ac-49f731ef26e3\") " pod="openshift-route-controller-manager/route-controller-manager-bbff9c674-sjkrj" Nov 28 13:24:34 crc kubenswrapper[4631]: I1128 13:24:34.156244 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8a0e31f0-4c01-4e0a-b0ac-49f731ef26e3-serving-cert\") pod \"route-controller-manager-bbff9c674-sjkrj\" (UID: \"8a0e31f0-4c01-4e0a-b0ac-49f731ef26e3\") " pod="openshift-route-controller-manager/route-controller-manager-bbff9c674-sjkrj" Nov 28 13:24:34 crc kubenswrapper[4631]: I1128 13:24:34.156371 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sgv2h\" (UniqueName: \"kubernetes.io/projected/8a0e31f0-4c01-4e0a-b0ac-49f731ef26e3-kube-api-access-sgv2h\") pod \"route-controller-manager-bbff9c674-sjkrj\" (UID: \"8a0e31f0-4c01-4e0a-b0ac-49f731ef26e3\") " pod="openshift-route-controller-manager/route-controller-manager-bbff9c674-sjkrj" Nov 28 13:24:34 crc kubenswrapper[4631]: I1128 13:24:34.157136 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8a0e31f0-4c01-4e0a-b0ac-49f731ef26e3-config\") pod \"route-controller-manager-bbff9c674-sjkrj\" (UID: \"8a0e31f0-4c01-4e0a-b0ac-49f731ef26e3\") " pod="openshift-route-controller-manager/route-controller-manager-bbff9c674-sjkrj" Nov 28 13:24:34 crc kubenswrapper[4631]: I1128 13:24:34.157505 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8a0e31f0-4c01-4e0a-b0ac-49f731ef26e3-client-ca\") pod \"route-controller-manager-bbff9c674-sjkrj\" (UID: \"8a0e31f0-4c01-4e0a-b0ac-49f731ef26e3\") " pod="openshift-route-controller-manager/route-controller-manager-bbff9c674-sjkrj" Nov 28 13:24:34 crc kubenswrapper[4631]: I1128 13:24:34.178883 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sgv2h\" (UniqueName: \"kubernetes.io/projected/8a0e31f0-4c01-4e0a-b0ac-49f731ef26e3-kube-api-access-sgv2h\") pod \"route-controller-manager-bbff9c674-sjkrj\" (UID: \"8a0e31f0-4c01-4e0a-b0ac-49f731ef26e3\") " pod="openshift-route-controller-manager/route-controller-manager-bbff9c674-sjkrj" Nov 28 13:24:34 crc kubenswrapper[4631]: I1128 13:24:34.178966 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8a0e31f0-4c01-4e0a-b0ac-49f731ef26e3-serving-cert\") pod \"route-controller-manager-bbff9c674-sjkrj\" (UID: \"8a0e31f0-4c01-4e0a-b0ac-49f731ef26e3\") " pod="openshift-route-controller-manager/route-controller-manager-bbff9c674-sjkrj" Nov 28 13:24:34 crc kubenswrapper[4631]: I1128 13:24:34.301722 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-bbff9c674-sjkrj" Nov 28 13:24:34 crc kubenswrapper[4631]: I1128 13:24:34.563926 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-bbff9c674-sjkrj"] Nov 28 13:24:34 crc kubenswrapper[4631]: I1128 13:24:34.632915 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-bbff9c674-sjkrj" event={"ID":"8a0e31f0-4c01-4e0a-b0ac-49f731ef26e3","Type":"ContainerStarted","Data":"9c6ac2fdd76ad3893a951e30f6300d0f0ff63f2288e9c20a9ed438d708e327fd"} Nov 28 13:24:35 crc kubenswrapper[4631]: I1128 13:24:35.648352 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-bbff9c674-sjkrj" event={"ID":"8a0e31f0-4c01-4e0a-b0ac-49f731ef26e3","Type":"ContainerStarted","Data":"0f53650917609699a2516a6ef534d6cfb13b473fa2e01232fc5a5ed3a3775c11"} Nov 28 13:24:35 crc kubenswrapper[4631]: I1128 13:24:35.668821 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-bbff9c674-sjkrj" podStartSLOduration=5.668797529 podStartE2EDuration="5.668797529s" podCreationTimestamp="2025-11-28 13:24:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 13:24:35.664789133 +0000 UTC m=+232.472092477" watchObservedRunningTime="2025-11-28 13:24:35.668797529 +0000 UTC m=+232.476100873" Nov 28 13:24:36 crc kubenswrapper[4631]: I1128 13:24:36.652738 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-bbff9c674-sjkrj" Nov 28 13:24:36 crc kubenswrapper[4631]: I1128 13:24:36.661100 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-bbff9c674-sjkrj" Nov 28 13:25:10 crc kubenswrapper[4631]: I1128 13:25:10.131222 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-bbff9c674-sjkrj"] Nov 28 13:25:10 crc kubenswrapper[4631]: I1128 13:25:10.132387 4631 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-bbff9c674-sjkrj" podUID="8a0e31f0-4c01-4e0a-b0ac-49f731ef26e3" containerName="route-controller-manager" containerID="cri-o://0f53650917609699a2516a6ef534d6cfb13b473fa2e01232fc5a5ed3a3775c11" gracePeriod=30 Nov 28 13:25:10 crc kubenswrapper[4631]: I1128 13:25:10.539813 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-bbff9c674-sjkrj" Nov 28 13:25:10 crc kubenswrapper[4631]: I1128 13:25:10.632878 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8a0e31f0-4c01-4e0a-b0ac-49f731ef26e3-serving-cert\") pod \"8a0e31f0-4c01-4e0a-b0ac-49f731ef26e3\" (UID: \"8a0e31f0-4c01-4e0a-b0ac-49f731ef26e3\") " Nov 28 13:25:10 crc kubenswrapper[4631]: I1128 13:25:10.633036 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sgv2h\" (UniqueName: \"kubernetes.io/projected/8a0e31f0-4c01-4e0a-b0ac-49f731ef26e3-kube-api-access-sgv2h\") pod \"8a0e31f0-4c01-4e0a-b0ac-49f731ef26e3\" (UID: \"8a0e31f0-4c01-4e0a-b0ac-49f731ef26e3\") " Nov 28 13:25:10 crc kubenswrapper[4631]: I1128 13:25:10.633096 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8a0e31f0-4c01-4e0a-b0ac-49f731ef26e3-client-ca\") pod \"8a0e31f0-4c01-4e0a-b0ac-49f731ef26e3\" (UID: \"8a0e31f0-4c01-4e0a-b0ac-49f731ef26e3\") " Nov 28 13:25:10 crc kubenswrapper[4631]: I1128 13:25:10.633151 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8a0e31f0-4c01-4e0a-b0ac-49f731ef26e3-config\") pod \"8a0e31f0-4c01-4e0a-b0ac-49f731ef26e3\" (UID: \"8a0e31f0-4c01-4e0a-b0ac-49f731ef26e3\") " Nov 28 13:25:10 crc kubenswrapper[4631]: I1128 13:25:10.634891 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8a0e31f0-4c01-4e0a-b0ac-49f731ef26e3-config" (OuterVolumeSpecName: "config") pod "8a0e31f0-4c01-4e0a-b0ac-49f731ef26e3" (UID: "8a0e31f0-4c01-4e0a-b0ac-49f731ef26e3"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 13:25:10 crc kubenswrapper[4631]: I1128 13:25:10.635378 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8a0e31f0-4c01-4e0a-b0ac-49f731ef26e3-client-ca" (OuterVolumeSpecName: "client-ca") pod "8a0e31f0-4c01-4e0a-b0ac-49f731ef26e3" (UID: "8a0e31f0-4c01-4e0a-b0ac-49f731ef26e3"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 13:25:10 crc kubenswrapper[4631]: I1128 13:25:10.657465 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8a0e31f0-4c01-4e0a-b0ac-49f731ef26e3-kube-api-access-sgv2h" (OuterVolumeSpecName: "kube-api-access-sgv2h") pod "8a0e31f0-4c01-4e0a-b0ac-49f731ef26e3" (UID: "8a0e31f0-4c01-4e0a-b0ac-49f731ef26e3"). InnerVolumeSpecName "kube-api-access-sgv2h". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:25:10 crc kubenswrapper[4631]: I1128 13:25:10.658242 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a0e31f0-4c01-4e0a-b0ac-49f731ef26e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8a0e31f0-4c01-4e0a-b0ac-49f731ef26e3" (UID: "8a0e31f0-4c01-4e0a-b0ac-49f731ef26e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:25:10 crc kubenswrapper[4631]: I1128 13:25:10.735860 4631 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8a0e31f0-4c01-4e0a-b0ac-49f731ef26e3-client-ca\") on node \"crc\" DevicePath \"\"" Nov 28 13:25:10 crc kubenswrapper[4631]: I1128 13:25:10.736353 4631 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8a0e31f0-4c01-4e0a-b0ac-49f731ef26e3-config\") on node \"crc\" DevicePath \"\"" Nov 28 13:25:10 crc kubenswrapper[4631]: I1128 13:25:10.736402 4631 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8a0e31f0-4c01-4e0a-b0ac-49f731ef26e3-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 28 13:25:10 crc kubenswrapper[4631]: I1128 13:25:10.736471 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sgv2h\" (UniqueName: \"kubernetes.io/projected/8a0e31f0-4c01-4e0a-b0ac-49f731ef26e3-kube-api-access-sgv2h\") on node \"crc\" DevicePath \"\"" Nov 28 13:25:10 crc kubenswrapper[4631]: I1128 13:25:10.853439 4631 generic.go:334] "Generic (PLEG): container finished" podID="8a0e31f0-4c01-4e0a-b0ac-49f731ef26e3" containerID="0f53650917609699a2516a6ef534d6cfb13b473fa2e01232fc5a5ed3a3775c11" exitCode=0 Nov 28 13:25:10 crc kubenswrapper[4631]: I1128 13:25:10.853558 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-bbff9c674-sjkrj" event={"ID":"8a0e31f0-4c01-4e0a-b0ac-49f731ef26e3","Type":"ContainerDied","Data":"0f53650917609699a2516a6ef534d6cfb13b473fa2e01232fc5a5ed3a3775c11"} Nov 28 13:25:10 crc kubenswrapper[4631]: I1128 13:25:10.853674 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-bbff9c674-sjkrj" event={"ID":"8a0e31f0-4c01-4e0a-b0ac-49f731ef26e3","Type":"ContainerDied","Data":"9c6ac2fdd76ad3893a951e30f6300d0f0ff63f2288e9c20a9ed438d708e327fd"} Nov 28 13:25:10 crc kubenswrapper[4631]: I1128 13:25:10.853705 4631 scope.go:117] "RemoveContainer" containerID="0f53650917609699a2516a6ef534d6cfb13b473fa2e01232fc5a5ed3a3775c11" Nov 28 13:25:10 crc kubenswrapper[4631]: I1128 13:25:10.853961 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-bbff9c674-sjkrj" Nov 28 13:25:10 crc kubenswrapper[4631]: I1128 13:25:10.880616 4631 scope.go:117] "RemoveContainer" containerID="0f53650917609699a2516a6ef534d6cfb13b473fa2e01232fc5a5ed3a3775c11" Nov 28 13:25:10 crc kubenswrapper[4631]: E1128 13:25:10.881402 4631 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0f53650917609699a2516a6ef534d6cfb13b473fa2e01232fc5a5ed3a3775c11\": container with ID starting with 0f53650917609699a2516a6ef534d6cfb13b473fa2e01232fc5a5ed3a3775c11 not found: ID does not exist" containerID="0f53650917609699a2516a6ef534d6cfb13b473fa2e01232fc5a5ed3a3775c11" Nov 28 13:25:10 crc kubenswrapper[4631]: I1128 13:25:10.881461 4631 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0f53650917609699a2516a6ef534d6cfb13b473fa2e01232fc5a5ed3a3775c11"} err="failed to get container status \"0f53650917609699a2516a6ef534d6cfb13b473fa2e01232fc5a5ed3a3775c11\": rpc error: code = NotFound desc = could not find container \"0f53650917609699a2516a6ef534d6cfb13b473fa2e01232fc5a5ed3a3775c11\": container with ID starting with 0f53650917609699a2516a6ef534d6cfb13b473fa2e01232fc5a5ed3a3775c11 not found: ID does not exist" Nov 28 13:25:10 crc kubenswrapper[4631]: I1128 13:25:10.906567 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-bbff9c674-sjkrj"] Nov 28 13:25:10 crc kubenswrapper[4631]: I1128 13:25:10.914234 4631 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-bbff9c674-sjkrj"] Nov 28 13:25:11 crc kubenswrapper[4631]: I1128 13:25:11.525259 4631 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8a0e31f0-4c01-4e0a-b0ac-49f731ef26e3" path="/var/lib/kubelet/pods/8a0e31f0-4c01-4e0a-b0ac-49f731ef26e3/volumes" Nov 28 13:25:12 crc kubenswrapper[4631]: I1128 13:25:12.008913 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-8469c66746-lszcn"] Nov 28 13:25:12 crc kubenswrapper[4631]: E1128 13:25:12.009130 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a0e31f0-4c01-4e0a-b0ac-49f731ef26e3" containerName="route-controller-manager" Nov 28 13:25:12 crc kubenswrapper[4631]: I1128 13:25:12.009144 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a0e31f0-4c01-4e0a-b0ac-49f731ef26e3" containerName="route-controller-manager" Nov 28 13:25:12 crc kubenswrapper[4631]: I1128 13:25:12.009245 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="8a0e31f0-4c01-4e0a-b0ac-49f731ef26e3" containerName="route-controller-manager" Nov 28 13:25:12 crc kubenswrapper[4631]: I1128 13:25:12.009690 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-8469c66746-lszcn" Nov 28 13:25:12 crc kubenswrapper[4631]: I1128 13:25:12.012433 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Nov 28 13:25:12 crc kubenswrapper[4631]: I1128 13:25:12.013017 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Nov 28 13:25:12 crc kubenswrapper[4631]: I1128 13:25:12.013191 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Nov 28 13:25:12 crc kubenswrapper[4631]: I1128 13:25:12.013372 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Nov 28 13:25:12 crc kubenswrapper[4631]: I1128 13:25:12.015258 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Nov 28 13:25:12 crc kubenswrapper[4631]: I1128 13:25:12.015528 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Nov 28 13:25:12 crc kubenswrapper[4631]: I1128 13:25:12.033729 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-8469c66746-lszcn"] Nov 28 13:25:12 crc kubenswrapper[4631]: I1128 13:25:12.056882 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xwp4c\" (UniqueName: \"kubernetes.io/projected/27afcdef-75dc-4f8c-8c8b-f0a80c15fd0c-kube-api-access-xwp4c\") pod \"route-controller-manager-8469c66746-lszcn\" (UID: \"27afcdef-75dc-4f8c-8c8b-f0a80c15fd0c\") " pod="openshift-route-controller-manager/route-controller-manager-8469c66746-lszcn" Nov 28 13:25:12 crc kubenswrapper[4631]: I1128 13:25:12.056955 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/27afcdef-75dc-4f8c-8c8b-f0a80c15fd0c-config\") pod \"route-controller-manager-8469c66746-lszcn\" (UID: \"27afcdef-75dc-4f8c-8c8b-f0a80c15fd0c\") " pod="openshift-route-controller-manager/route-controller-manager-8469c66746-lszcn" Nov 28 13:25:12 crc kubenswrapper[4631]: I1128 13:25:12.057001 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/27afcdef-75dc-4f8c-8c8b-f0a80c15fd0c-serving-cert\") pod \"route-controller-manager-8469c66746-lszcn\" (UID: \"27afcdef-75dc-4f8c-8c8b-f0a80c15fd0c\") " pod="openshift-route-controller-manager/route-controller-manager-8469c66746-lszcn" Nov 28 13:25:12 crc kubenswrapper[4631]: I1128 13:25:12.057037 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/27afcdef-75dc-4f8c-8c8b-f0a80c15fd0c-client-ca\") pod \"route-controller-manager-8469c66746-lszcn\" (UID: \"27afcdef-75dc-4f8c-8c8b-f0a80c15fd0c\") " pod="openshift-route-controller-manager/route-controller-manager-8469c66746-lszcn" Nov 28 13:25:12 crc kubenswrapper[4631]: I1128 13:25:12.159076 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/27afcdef-75dc-4f8c-8c8b-f0a80c15fd0c-config\") pod \"route-controller-manager-8469c66746-lszcn\" (UID: \"27afcdef-75dc-4f8c-8c8b-f0a80c15fd0c\") " pod="openshift-route-controller-manager/route-controller-manager-8469c66746-lszcn" Nov 28 13:25:12 crc kubenswrapper[4631]: I1128 13:25:12.160375 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/27afcdef-75dc-4f8c-8c8b-f0a80c15fd0c-config\") pod \"route-controller-manager-8469c66746-lszcn\" (UID: \"27afcdef-75dc-4f8c-8c8b-f0a80c15fd0c\") " pod="openshift-route-controller-manager/route-controller-manager-8469c66746-lszcn" Nov 28 13:25:12 crc kubenswrapper[4631]: I1128 13:25:12.160511 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/27afcdef-75dc-4f8c-8c8b-f0a80c15fd0c-serving-cert\") pod \"route-controller-manager-8469c66746-lszcn\" (UID: \"27afcdef-75dc-4f8c-8c8b-f0a80c15fd0c\") " pod="openshift-route-controller-manager/route-controller-manager-8469c66746-lszcn" Nov 28 13:25:12 crc kubenswrapper[4631]: I1128 13:25:12.161272 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/27afcdef-75dc-4f8c-8c8b-f0a80c15fd0c-client-ca\") pod \"route-controller-manager-8469c66746-lszcn\" (UID: \"27afcdef-75dc-4f8c-8c8b-f0a80c15fd0c\") " pod="openshift-route-controller-manager/route-controller-manager-8469c66746-lszcn" Nov 28 13:25:12 crc kubenswrapper[4631]: I1128 13:25:12.161435 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xwp4c\" (UniqueName: \"kubernetes.io/projected/27afcdef-75dc-4f8c-8c8b-f0a80c15fd0c-kube-api-access-xwp4c\") pod \"route-controller-manager-8469c66746-lszcn\" (UID: \"27afcdef-75dc-4f8c-8c8b-f0a80c15fd0c\") " pod="openshift-route-controller-manager/route-controller-manager-8469c66746-lszcn" Nov 28 13:25:12 crc kubenswrapper[4631]: I1128 13:25:12.162497 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/27afcdef-75dc-4f8c-8c8b-f0a80c15fd0c-client-ca\") pod \"route-controller-manager-8469c66746-lszcn\" (UID: \"27afcdef-75dc-4f8c-8c8b-f0a80c15fd0c\") " pod="openshift-route-controller-manager/route-controller-manager-8469c66746-lszcn" Nov 28 13:25:12 crc kubenswrapper[4631]: I1128 13:25:12.169616 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/27afcdef-75dc-4f8c-8c8b-f0a80c15fd0c-serving-cert\") pod \"route-controller-manager-8469c66746-lszcn\" (UID: \"27afcdef-75dc-4f8c-8c8b-f0a80c15fd0c\") " pod="openshift-route-controller-manager/route-controller-manager-8469c66746-lszcn" Nov 28 13:25:12 crc kubenswrapper[4631]: I1128 13:25:12.193056 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xwp4c\" (UniqueName: \"kubernetes.io/projected/27afcdef-75dc-4f8c-8c8b-f0a80c15fd0c-kube-api-access-xwp4c\") pod \"route-controller-manager-8469c66746-lszcn\" (UID: \"27afcdef-75dc-4f8c-8c8b-f0a80c15fd0c\") " pod="openshift-route-controller-manager/route-controller-manager-8469c66746-lszcn" Nov 28 13:25:12 crc kubenswrapper[4631]: I1128 13:25:12.326894 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-8469c66746-lszcn" Nov 28 13:25:12 crc kubenswrapper[4631]: I1128 13:25:12.783723 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-8469c66746-lszcn"] Nov 28 13:25:12 crc kubenswrapper[4631]: I1128 13:25:12.871396 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-8469c66746-lszcn" event={"ID":"27afcdef-75dc-4f8c-8c8b-f0a80c15fd0c","Type":"ContainerStarted","Data":"4c215ad6b9ad922fa6c51a44f0773ac4cef587b77d7ad61eb7ee026962e3a063"} Nov 28 13:25:13 crc kubenswrapper[4631]: I1128 13:25:13.880593 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-8469c66746-lszcn" event={"ID":"27afcdef-75dc-4f8c-8c8b-f0a80c15fd0c","Type":"ContainerStarted","Data":"b55ffd1197faa0f80ab14cd6c3abb550e8132a02d310990c2c1e1ba9076916ef"} Nov 28 13:25:13 crc kubenswrapper[4631]: I1128 13:25:13.880968 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-8469c66746-lszcn" Nov 28 13:25:13 crc kubenswrapper[4631]: I1128 13:25:13.886119 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-8469c66746-lszcn" Nov 28 13:25:13 crc kubenswrapper[4631]: I1128 13:25:13.900150 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-8469c66746-lszcn" podStartSLOduration=3.900122064 podStartE2EDuration="3.900122064s" podCreationTimestamp="2025-11-28 13:25:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 13:25:13.89994637 +0000 UTC m=+270.707249714" watchObservedRunningTime="2025-11-28 13:25:13.900122064 +0000 UTC m=+270.707425408" Nov 28 13:25:43 crc kubenswrapper[4631]: I1128 13:25:43.361688 4631 cert_rotation.go:91] certificate rotation detected, shutting down client connections to start using new credentials Nov 28 13:26:49 crc kubenswrapper[4631]: I1128 13:26:49.635425 4631 patch_prober.go:28] interesting pod/machine-config-daemon-47sxw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 28 13:26:49 crc kubenswrapper[4631]: I1128 13:26:49.636476 4631 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" podUID="8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 28 13:27:19 crc kubenswrapper[4631]: I1128 13:27:19.635223 4631 patch_prober.go:28] interesting pod/machine-config-daemon-47sxw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 28 13:27:19 crc kubenswrapper[4631]: I1128 13:27:19.638116 4631 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" podUID="8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 28 13:27:49 crc kubenswrapper[4631]: I1128 13:27:49.635081 4631 patch_prober.go:28] interesting pod/machine-config-daemon-47sxw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 28 13:27:49 crc kubenswrapper[4631]: I1128 13:27:49.636385 4631 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" podUID="8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 28 13:27:49 crc kubenswrapper[4631]: I1128 13:27:49.636696 4631 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" Nov 28 13:27:49 crc kubenswrapper[4631]: I1128 13:27:49.637914 4631 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"b04ff78ff7998dc8957c0acf1246dd6adabf1e8af15924c3953943c9bdb3150a"} pod="openshift-machine-config-operator/machine-config-daemon-47sxw" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 28 13:27:49 crc kubenswrapper[4631]: I1128 13:27:49.638033 4631 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" podUID="8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1" containerName="machine-config-daemon" containerID="cri-o://b04ff78ff7998dc8957c0acf1246dd6adabf1e8af15924c3953943c9bdb3150a" gracePeriod=600 Nov 28 13:27:49 crc kubenswrapper[4631]: I1128 13:27:49.991562 4631 generic.go:334] "Generic (PLEG): container finished" podID="8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1" containerID="b04ff78ff7998dc8957c0acf1246dd6adabf1e8af15924c3953943c9bdb3150a" exitCode=0 Nov 28 13:27:49 crc kubenswrapper[4631]: I1128 13:27:49.991668 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" event={"ID":"8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1","Type":"ContainerDied","Data":"b04ff78ff7998dc8957c0acf1246dd6adabf1e8af15924c3953943c9bdb3150a"} Nov 28 13:27:49 crc kubenswrapper[4631]: I1128 13:27:49.991965 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" event={"ID":"8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1","Type":"ContainerStarted","Data":"42d567fa8bd6c073cd89c5129f91adc5f5feefc7933a184e745232605d6012b2"} Nov 28 13:27:49 crc kubenswrapper[4631]: I1128 13:27:49.992847 4631 scope.go:117] "RemoveContainer" containerID="7f613562449a8f5ebf17cd881967c47d44fb0296171712af40f42263f69f8136" Nov 28 13:28:52 crc kubenswrapper[4631]: I1128 13:28:52.437730 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-rslr5"] Nov 28 13:28:52 crc kubenswrapper[4631]: I1128 13:28:52.438979 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-rslr5" Nov 28 13:28:52 crc kubenswrapper[4631]: W1128 13:28:52.443070 4631 reflector.go:561] object-"cert-manager"/"cert-manager-cainjector-dockercfg-kgjxk": failed to list *v1.Secret: secrets "cert-manager-cainjector-dockercfg-kgjxk" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "cert-manager": no relationship found between node 'crc' and this object Nov 28 13:28:52 crc kubenswrapper[4631]: E1128 13:28:52.443112 4631 reflector.go:158] "Unhandled Error" err="object-\"cert-manager\"/\"cert-manager-cainjector-dockercfg-kgjxk\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"cert-manager-cainjector-dockercfg-kgjxk\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"cert-manager\": no relationship found between node 'crc' and this object" logger="UnhandledError" Nov 28 13:28:52 crc kubenswrapper[4631]: W1128 13:28:52.443167 4631 reflector.go:561] object-"cert-manager"/"openshift-service-ca.crt": failed to list *v1.ConfigMap: configmaps "openshift-service-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "cert-manager": no relationship found between node 'crc' and this object Nov 28 13:28:52 crc kubenswrapper[4631]: E1128 13:28:52.443180 4631 reflector.go:158] "Unhandled Error" err="object-\"cert-manager\"/\"openshift-service-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"openshift-service-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"cert-manager\": no relationship found between node 'crc' and this object" logger="UnhandledError" Nov 28 13:28:52 crc kubenswrapper[4631]: W1128 13:28:52.443601 4631 reflector.go:561] object-"cert-manager"/"kube-root-ca.crt": failed to list *v1.ConfigMap: configmaps "kube-root-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "cert-manager": no relationship found between node 'crc' and this object Nov 28 13:28:52 crc kubenswrapper[4631]: E1128 13:28:52.443668 4631 reflector.go:158] "Unhandled Error" err="object-\"cert-manager\"/\"kube-root-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"kube-root-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"cert-manager\": no relationship found between node 'crc' and this object" logger="UnhandledError" Nov 28 13:28:52 crc kubenswrapper[4631]: I1128 13:28:52.463776 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-rslr5"] Nov 28 13:28:52 crc kubenswrapper[4631]: I1128 13:28:52.474302 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-5b446d88c5-kf7cv"] Nov 28 13:28:52 crc kubenswrapper[4631]: I1128 13:28:52.475145 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-kf7cv" Nov 28 13:28:52 crc kubenswrapper[4631]: I1128 13:28:52.478351 4631 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-52dmb" Nov 28 13:28:52 crc kubenswrapper[4631]: I1128 13:28:52.509774 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-l6tzn"] Nov 28 13:28:52 crc kubenswrapper[4631]: I1128 13:28:52.511342 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-l6tzn" Nov 28 13:28:52 crc kubenswrapper[4631]: I1128 13:28:52.517695 4631 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-5cnbn" Nov 28 13:28:52 crc kubenswrapper[4631]: I1128 13:28:52.520493 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-kf7cv"] Nov 28 13:28:52 crc kubenswrapper[4631]: I1128 13:28:52.526803 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m5dpz\" (UniqueName: \"kubernetes.io/projected/6591b4ca-08f0-4346-aed6-8e9d1caf2b7f-kube-api-access-m5dpz\") pod \"cert-manager-cainjector-7f985d654d-rslr5\" (UID: \"6591b4ca-08f0-4346-aed6-8e9d1caf2b7f\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-rslr5" Nov 28 13:28:52 crc kubenswrapper[4631]: I1128 13:28:52.526853 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hmntn\" (UniqueName: \"kubernetes.io/projected/4e46e751-ddac-4bc4-b463-d7cf9b772871-kube-api-access-hmntn\") pod \"cert-manager-5b446d88c5-kf7cv\" (UID: \"4e46e751-ddac-4bc4-b463-d7cf9b772871\") " pod="cert-manager/cert-manager-5b446d88c5-kf7cv" Nov 28 13:28:52 crc kubenswrapper[4631]: I1128 13:28:52.526882 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h274j\" (UniqueName: \"kubernetes.io/projected/11540749-0bb3-4c94-9d22-12a889ac10ce-kube-api-access-h274j\") pod \"cert-manager-webhook-5655c58dd6-l6tzn\" (UID: \"11540749-0bb3-4c94-9d22-12a889ac10ce\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-l6tzn" Nov 28 13:28:52 crc kubenswrapper[4631]: I1128 13:28:52.535344 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-l6tzn"] Nov 28 13:28:52 crc kubenswrapper[4631]: I1128 13:28:52.632405 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hmntn\" (UniqueName: \"kubernetes.io/projected/4e46e751-ddac-4bc4-b463-d7cf9b772871-kube-api-access-hmntn\") pod \"cert-manager-5b446d88c5-kf7cv\" (UID: \"4e46e751-ddac-4bc4-b463-d7cf9b772871\") " pod="cert-manager/cert-manager-5b446d88c5-kf7cv" Nov 28 13:28:52 crc kubenswrapper[4631]: I1128 13:28:52.632484 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h274j\" (UniqueName: \"kubernetes.io/projected/11540749-0bb3-4c94-9d22-12a889ac10ce-kube-api-access-h274j\") pod \"cert-manager-webhook-5655c58dd6-l6tzn\" (UID: \"11540749-0bb3-4c94-9d22-12a889ac10ce\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-l6tzn" Nov 28 13:28:52 crc kubenswrapper[4631]: I1128 13:28:52.632553 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m5dpz\" (UniqueName: \"kubernetes.io/projected/6591b4ca-08f0-4346-aed6-8e9d1caf2b7f-kube-api-access-m5dpz\") pod \"cert-manager-cainjector-7f985d654d-rslr5\" (UID: \"6591b4ca-08f0-4346-aed6-8e9d1caf2b7f\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-rslr5" Nov 28 13:28:53 crc kubenswrapper[4631]: I1128 13:28:53.339928 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Nov 28 13:28:53 crc kubenswrapper[4631]: I1128 13:28:53.355257 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Nov 28 13:28:53 crc kubenswrapper[4631]: I1128 13:28:53.365100 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hmntn\" (UniqueName: \"kubernetes.io/projected/4e46e751-ddac-4bc4-b463-d7cf9b772871-kube-api-access-hmntn\") pod \"cert-manager-5b446d88c5-kf7cv\" (UID: \"4e46e751-ddac-4bc4-b463-d7cf9b772871\") " pod="cert-manager/cert-manager-5b446d88c5-kf7cv" Nov 28 13:28:53 crc kubenswrapper[4631]: I1128 13:28:53.366665 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m5dpz\" (UniqueName: \"kubernetes.io/projected/6591b4ca-08f0-4346-aed6-8e9d1caf2b7f-kube-api-access-m5dpz\") pod \"cert-manager-cainjector-7f985d654d-rslr5\" (UID: \"6591b4ca-08f0-4346-aed6-8e9d1caf2b7f\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-rslr5" Nov 28 13:28:53 crc kubenswrapper[4631]: I1128 13:28:53.368121 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h274j\" (UniqueName: \"kubernetes.io/projected/11540749-0bb3-4c94-9d22-12a889ac10ce-kube-api-access-h274j\") pod \"cert-manager-webhook-5655c58dd6-l6tzn\" (UID: \"11540749-0bb3-4c94-9d22-12a889ac10ce\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-l6tzn" Nov 28 13:28:53 crc kubenswrapper[4631]: I1128 13:28:53.402912 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-kf7cv" Nov 28 13:28:53 crc kubenswrapper[4631]: I1128 13:28:53.437045 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-l6tzn" Nov 28 13:28:53 crc kubenswrapper[4631]: I1128 13:28:53.720204 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-l6tzn"] Nov 28 13:28:53 crc kubenswrapper[4631]: I1128 13:28:53.731819 4631 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 28 13:28:53 crc kubenswrapper[4631]: I1128 13:28:53.867848 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-kf7cv"] Nov 28 13:28:53 crc kubenswrapper[4631]: W1128 13:28:53.871736 4631 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4e46e751_ddac_4bc4_b463_d7cf9b772871.slice/crio-a0e04f79a039daa0271bd0b62478b1a4a5375633adc576a673226211b7b99763 WatchSource:0}: Error finding container a0e04f79a039daa0271bd0b62478b1a4a5375633adc576a673226211b7b99763: Status 404 returned error can't find the container with id a0e04f79a039daa0271bd0b62478b1a4a5375633adc576a673226211b7b99763 Nov 28 13:28:53 crc kubenswrapper[4631]: I1128 13:28:53.945038 4631 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-kgjxk" Nov 28 13:28:53 crc kubenswrapper[4631]: I1128 13:28:53.945249 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-rslr5" Nov 28 13:28:54 crc kubenswrapper[4631]: I1128 13:28:54.169238 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-rslr5"] Nov 28 13:28:54 crc kubenswrapper[4631]: W1128 13:28:54.176327 4631 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6591b4ca_08f0_4346_aed6_8e9d1caf2b7f.slice/crio-e37ac5dbc44bfb89d11bcdcc49bd7431e32b92544e6fe634148cd1bb22c7f420 WatchSource:0}: Error finding container e37ac5dbc44bfb89d11bcdcc49bd7431e32b92544e6fe634148cd1bb22c7f420: Status 404 returned error can't find the container with id e37ac5dbc44bfb89d11bcdcc49bd7431e32b92544e6fe634148cd1bb22c7f420 Nov 28 13:28:54 crc kubenswrapper[4631]: I1128 13:28:54.437012 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-l6tzn" event={"ID":"11540749-0bb3-4c94-9d22-12a889ac10ce","Type":"ContainerStarted","Data":"55d8341bea45d4b9d24a09ac0ccfa70e741ace7d1367ce2732b7861b15379600"} Nov 28 13:28:54 crc kubenswrapper[4631]: I1128 13:28:54.438087 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-kf7cv" event={"ID":"4e46e751-ddac-4bc4-b463-d7cf9b772871","Type":"ContainerStarted","Data":"a0e04f79a039daa0271bd0b62478b1a4a5375633adc576a673226211b7b99763"} Nov 28 13:28:54 crc kubenswrapper[4631]: I1128 13:28:54.439235 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-rslr5" event={"ID":"6591b4ca-08f0-4346-aed6-8e9d1caf2b7f","Type":"ContainerStarted","Data":"e37ac5dbc44bfb89d11bcdcc49bd7431e32b92544e6fe634148cd1bb22c7f420"} Nov 28 13:28:56 crc kubenswrapper[4631]: I1128 13:28:56.454330 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-l6tzn" event={"ID":"11540749-0bb3-4c94-9d22-12a889ac10ce","Type":"ContainerStarted","Data":"12955713024fab1c7bf8ab86720f28fa38176685f4cd9175a9eae7df87ba0a01"} Nov 28 13:28:56 crc kubenswrapper[4631]: I1128 13:28:56.454515 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-5655c58dd6-l6tzn" Nov 28 13:28:56 crc kubenswrapper[4631]: I1128 13:28:56.478085 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-5655c58dd6-l6tzn" podStartSLOduration=2.139209487 podStartE2EDuration="4.478048365s" podCreationTimestamp="2025-11-28 13:28:52 +0000 UTC" firstStartedPulling="2025-11-28 13:28:53.731565235 +0000 UTC m=+490.538868579" lastFinishedPulling="2025-11-28 13:28:56.070404073 +0000 UTC m=+492.877707457" observedRunningTime="2025-11-28 13:28:56.471685076 +0000 UTC m=+493.278988430" watchObservedRunningTime="2025-11-28 13:28:56.478048365 +0000 UTC m=+493.285351709" Nov 28 13:28:59 crc kubenswrapper[4631]: I1128 13:28:59.476165 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-kf7cv" event={"ID":"4e46e751-ddac-4bc4-b463-d7cf9b772871","Type":"ContainerStarted","Data":"9ae7c493a08fb6ce3e01b51db1f1f8a23e5907cbecf9f9e8b070787f16f7ecb6"} Nov 28 13:28:59 crc kubenswrapper[4631]: I1128 13:28:59.478211 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-rslr5" event={"ID":"6591b4ca-08f0-4346-aed6-8e9d1caf2b7f","Type":"ContainerStarted","Data":"4a75c0f30413d75a770f65d585429c688b4863a0cbe3f250e9a573a6ada12a12"} Nov 28 13:28:59 crc kubenswrapper[4631]: I1128 13:28:59.501748 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-5b446d88c5-kf7cv" podStartSLOduration=2.677541162 podStartE2EDuration="7.501717769s" podCreationTimestamp="2025-11-28 13:28:52 +0000 UTC" firstStartedPulling="2025-11-28 13:28:53.873481199 +0000 UTC m=+490.680784553" lastFinishedPulling="2025-11-28 13:28:58.697657806 +0000 UTC m=+495.504961160" observedRunningTime="2025-11-28 13:28:59.495667228 +0000 UTC m=+496.302970592" watchObservedRunningTime="2025-11-28 13:28:59.501717769 +0000 UTC m=+496.309021113" Nov 28 13:28:59 crc kubenswrapper[4631]: I1128 13:28:59.522423 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-7f985d654d-rslr5" podStartSLOduration=3.025909394 podStartE2EDuration="7.522377615s" podCreationTimestamp="2025-11-28 13:28:52 +0000 UTC" firstStartedPulling="2025-11-28 13:28:54.179939404 +0000 UTC m=+490.987242768" lastFinishedPulling="2025-11-28 13:28:58.676407655 +0000 UTC m=+495.483710989" observedRunningTime="2025-11-28 13:28:59.521133554 +0000 UTC m=+496.328436898" watchObservedRunningTime="2025-11-28 13:28:59.522377615 +0000 UTC m=+496.329680989" Nov 28 13:29:02 crc kubenswrapper[4631]: I1128 13:29:02.802941 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-268tw"] Nov 28 13:29:02 crc kubenswrapper[4631]: I1128 13:29:02.804081 4631 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-268tw" podUID="a205f475-340d-4ec5-a860-d4a5c2ceb39e" containerName="ovn-controller" containerID="cri-o://e571a03b3c52184c8fa59e1ab6fff8d9118f114dd6c156638ba36fb7a0d2fe40" gracePeriod=30 Nov 28 13:29:02 crc kubenswrapper[4631]: I1128 13:29:02.804204 4631 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-268tw" podUID="a205f475-340d-4ec5-a860-d4a5c2ceb39e" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://016163e216d642018075fa0e9ab61fa630fa9718030246855fd44a8104c3d6e6" gracePeriod=30 Nov 28 13:29:02 crc kubenswrapper[4631]: I1128 13:29:02.804279 4631 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-268tw" podUID="a205f475-340d-4ec5-a860-d4a5c2ceb39e" containerName="kube-rbac-proxy-node" containerID="cri-o://e5085439ece35186f2692a8cf5be0b011fbf3f212f4f772b674c70649486c89b" gracePeriod=30 Nov 28 13:29:02 crc kubenswrapper[4631]: I1128 13:29:02.804319 4631 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-268tw" podUID="a205f475-340d-4ec5-a860-d4a5c2ceb39e" containerName="ovn-acl-logging" containerID="cri-o://95aeff4871c229569b0c99ff9a6cd17c63f10c1c89d9f04c533dce699c89ce31" gracePeriod=30 Nov 28 13:29:02 crc kubenswrapper[4631]: I1128 13:29:02.804686 4631 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-268tw" podUID="a205f475-340d-4ec5-a860-d4a5c2ceb39e" containerName="northd" containerID="cri-o://794ab47dc4660d76accb0456995b602cd5cdfc08616f63427947aba9e725eaf7" gracePeriod=30 Nov 28 13:29:02 crc kubenswrapper[4631]: I1128 13:29:02.804803 4631 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-268tw" podUID="a205f475-340d-4ec5-a860-d4a5c2ceb39e" containerName="sbdb" containerID="cri-o://9c674448c309477eb36de0eb0312fde682c3248fb66ed1d1d4100dc114fda786" gracePeriod=30 Nov 28 13:29:02 crc kubenswrapper[4631]: I1128 13:29:02.804847 4631 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-268tw" podUID="a205f475-340d-4ec5-a860-d4a5c2ceb39e" containerName="nbdb" containerID="cri-o://0ab2d69688bf305aeb7ece5c432d87adcb06eaca6c90ef5aa4e73b7fe477246b" gracePeriod=30 Nov 28 13:29:02 crc kubenswrapper[4631]: I1128 13:29:02.857889 4631 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-268tw" podUID="a205f475-340d-4ec5-a860-d4a5c2ceb39e" containerName="ovnkube-controller" containerID="cri-o://da4a7ce2b9fcb91df9dab6ba86046a998e310a5cc62a8eefcd98ea01a26f8db1" gracePeriod=30 Nov 28 13:29:03 crc kubenswrapper[4631]: E1128 13:29:03.030712 4631 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of da4a7ce2b9fcb91df9dab6ba86046a998e310a5cc62a8eefcd98ea01a26f8db1 is running failed: container process not found" containerID="da4a7ce2b9fcb91df9dab6ba86046a998e310a5cc62a8eefcd98ea01a26f8db1" cmd=["/bin/bash","-c","#!/bin/bash\ntest -f /etc/cni/net.d/10-ovn-kubernetes.conf\n"] Nov 28 13:29:03 crc kubenswrapper[4631]: E1128 13:29:03.031140 4631 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of da4a7ce2b9fcb91df9dab6ba86046a998e310a5cc62a8eefcd98ea01a26f8db1 is running failed: container process not found" containerID="da4a7ce2b9fcb91df9dab6ba86046a998e310a5cc62a8eefcd98ea01a26f8db1" cmd=["/bin/bash","-c","#!/bin/bash\ntest -f /etc/cni/net.d/10-ovn-kubernetes.conf\n"] Nov 28 13:29:03 crc kubenswrapper[4631]: E1128 13:29:03.031370 4631 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of da4a7ce2b9fcb91df9dab6ba86046a998e310a5cc62a8eefcd98ea01a26f8db1 is running failed: container process not found" containerID="da4a7ce2b9fcb91df9dab6ba86046a998e310a5cc62a8eefcd98ea01a26f8db1" cmd=["/bin/bash","-c","#!/bin/bash\ntest -f /etc/cni/net.d/10-ovn-kubernetes.conf\n"] Nov 28 13:29:03 crc kubenswrapper[4631]: E1128 13:29:03.031409 4631 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of da4a7ce2b9fcb91df9dab6ba86046a998e310a5cc62a8eefcd98ea01a26f8db1 is running failed: container process not found" probeType="Readiness" pod="openshift-ovn-kubernetes/ovnkube-node-268tw" podUID="a205f475-340d-4ec5-a860-d4a5c2ceb39e" containerName="ovnkube-controller" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.188369 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-268tw_a205f475-340d-4ec5-a860-d4a5c2ceb39e/ovn-acl-logging/0.log" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.188829 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-268tw_a205f475-340d-4ec5-a860-d4a5c2ceb39e/ovn-controller/0.log" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.189206 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-268tw" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.202402 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/a205f475-340d-4ec5-a860-d4a5c2ceb39e-ovn-node-metrics-cert\") pod \"a205f475-340d-4ec5-a860-d4a5c2ceb39e\" (UID: \"a205f475-340d-4ec5-a860-d4a5c2ceb39e\") " Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.202747 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a205f475-340d-4ec5-a860-d4a5c2ceb39e-etc-openvswitch\") pod \"a205f475-340d-4ec5-a860-d4a5c2ceb39e\" (UID: \"a205f475-340d-4ec5-a860-d4a5c2ceb39e\") " Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.202768 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/a205f475-340d-4ec5-a860-d4a5c2ceb39e-host-run-netns\") pod \"a205f475-340d-4ec5-a860-d4a5c2ceb39e\" (UID: \"a205f475-340d-4ec5-a860-d4a5c2ceb39e\") " Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.202794 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/a205f475-340d-4ec5-a860-d4a5c2ceb39e-ovnkube-script-lib\") pod \"a205f475-340d-4ec5-a860-d4a5c2ceb39e\" (UID: \"a205f475-340d-4ec5-a860-d4a5c2ceb39e\") " Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.202812 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/a205f475-340d-4ec5-a860-d4a5c2ceb39e-systemd-units\") pod \"a205f475-340d-4ec5-a860-d4a5c2ceb39e\" (UID: \"a205f475-340d-4ec5-a860-d4a5c2ceb39e\") " Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.202830 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/a205f475-340d-4ec5-a860-d4a5c2ceb39e-host-kubelet\") pod \"a205f475-340d-4ec5-a860-d4a5c2ceb39e\" (UID: \"a205f475-340d-4ec5-a860-d4a5c2ceb39e\") " Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.202848 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/a205f475-340d-4ec5-a860-d4a5c2ceb39e-host-cni-netd\") pod \"a205f475-340d-4ec5-a860-d4a5c2ceb39e\" (UID: \"a205f475-340d-4ec5-a860-d4a5c2ceb39e\") " Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.202870 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a205f475-340d-4ec5-a860-d4a5c2ceb39e-run-openvswitch\") pod \"a205f475-340d-4ec5-a860-d4a5c2ceb39e\" (UID: \"a205f475-340d-4ec5-a860-d4a5c2ceb39e\") " Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.202892 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/a205f475-340d-4ec5-a860-d4a5c2ceb39e-host-cni-bin\") pod \"a205f475-340d-4ec5-a860-d4a5c2ceb39e\" (UID: \"a205f475-340d-4ec5-a860-d4a5c2ceb39e\") " Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.202920 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/a205f475-340d-4ec5-a860-d4a5c2ceb39e-host-run-ovn-kubernetes\") pod \"a205f475-340d-4ec5-a860-d4a5c2ceb39e\" (UID: \"a205f475-340d-4ec5-a860-d4a5c2ceb39e\") " Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.202912 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a205f475-340d-4ec5-a860-d4a5c2ceb39e-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "a205f475-340d-4ec5-a860-d4a5c2ceb39e" (UID: "a205f475-340d-4ec5-a860-d4a5c2ceb39e"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.202951 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/a205f475-340d-4ec5-a860-d4a5c2ceb39e-host-var-lib-cni-networks-ovn-kubernetes\") pod \"a205f475-340d-4ec5-a860-d4a5c2ceb39e\" (UID: \"a205f475-340d-4ec5-a860-d4a5c2ceb39e\") " Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.202981 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/a205f475-340d-4ec5-a860-d4a5c2ceb39e-host-slash\") pod \"a205f475-340d-4ec5-a860-d4a5c2ceb39e\" (UID: \"a205f475-340d-4ec5-a860-d4a5c2ceb39e\") " Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.202987 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a205f475-340d-4ec5-a860-d4a5c2ceb39e-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "a205f475-340d-4ec5-a860-d4a5c2ceb39e" (UID: "a205f475-340d-4ec5-a860-d4a5c2ceb39e"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.203012 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a205f475-340d-4ec5-a860-d4a5c2ceb39e-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "a205f475-340d-4ec5-a860-d4a5c2ceb39e" (UID: "a205f475-340d-4ec5-a860-d4a5c2ceb39e"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.203011 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/a205f475-340d-4ec5-a860-d4a5c2ceb39e-env-overrides\") pod \"a205f475-340d-4ec5-a860-d4a5c2ceb39e\" (UID: \"a205f475-340d-4ec5-a860-d4a5c2ceb39e\") " Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.203082 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/a205f475-340d-4ec5-a860-d4a5c2ceb39e-log-socket\") pod \"a205f475-340d-4ec5-a860-d4a5c2ceb39e\" (UID: \"a205f475-340d-4ec5-a860-d4a5c2ceb39e\") " Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.203122 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a205f475-340d-4ec5-a860-d4a5c2ceb39e-var-lib-openvswitch\") pod \"a205f475-340d-4ec5-a860-d4a5c2ceb39e\" (UID: \"a205f475-340d-4ec5-a860-d4a5c2ceb39e\") " Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.203152 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/a205f475-340d-4ec5-a860-d4a5c2ceb39e-ovnkube-config\") pod \"a205f475-340d-4ec5-a860-d4a5c2ceb39e\" (UID: \"a205f475-340d-4ec5-a860-d4a5c2ceb39e\") " Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.203222 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a205f475-340d-4ec5-a860-d4a5c2ceb39e-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "a205f475-340d-4ec5-a860-d4a5c2ceb39e" (UID: "a205f475-340d-4ec5-a860-d4a5c2ceb39e"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.203260 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a205f475-340d-4ec5-a860-d4a5c2ceb39e-log-socket" (OuterVolumeSpecName: "log-socket") pod "a205f475-340d-4ec5-a860-d4a5c2ceb39e" (UID: "a205f475-340d-4ec5-a860-d4a5c2ceb39e"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.203437 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a205f475-340d-4ec5-a860-d4a5c2ceb39e-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "a205f475-340d-4ec5-a860-d4a5c2ceb39e" (UID: "a205f475-340d-4ec5-a860-d4a5c2ceb39e"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.203477 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a205f475-340d-4ec5-a860-d4a5c2ceb39e-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "a205f475-340d-4ec5-a860-d4a5c2ceb39e" (UID: "a205f475-340d-4ec5-a860-d4a5c2ceb39e"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.203493 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a205f475-340d-4ec5-a860-d4a5c2ceb39e-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "a205f475-340d-4ec5-a860-d4a5c2ceb39e" (UID: "a205f475-340d-4ec5-a860-d4a5c2ceb39e"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.203502 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a205f475-340d-4ec5-a860-d4a5c2ceb39e-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "a205f475-340d-4ec5-a860-d4a5c2ceb39e" (UID: "a205f475-340d-4ec5-a860-d4a5c2ceb39e"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.203523 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a205f475-340d-4ec5-a860-d4a5c2ceb39e-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "a205f475-340d-4ec5-a860-d4a5c2ceb39e" (UID: "a205f475-340d-4ec5-a860-d4a5c2ceb39e"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.203548 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a205f475-340d-4ec5-a860-d4a5c2ceb39e-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "a205f475-340d-4ec5-a860-d4a5c2ceb39e" (UID: "a205f475-340d-4ec5-a860-d4a5c2ceb39e"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.203568 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a205f475-340d-4ec5-a860-d4a5c2ceb39e-host-slash" (OuterVolumeSpecName: "host-slash") pod "a205f475-340d-4ec5-a860-d4a5c2ceb39e" (UID: "a205f475-340d-4ec5-a860-d4a5c2ceb39e"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.203667 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a205f475-340d-4ec5-a860-d4a5c2ceb39e-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "a205f475-340d-4ec5-a860-d4a5c2ceb39e" (UID: "a205f475-340d-4ec5-a860-d4a5c2ceb39e"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.203728 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a205f475-340d-4ec5-a860-d4a5c2ceb39e-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "a205f475-340d-4ec5-a860-d4a5c2ceb39e" (UID: "a205f475-340d-4ec5-a860-d4a5c2ceb39e"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.203750 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a205f475-340d-4ec5-a860-d4a5c2ceb39e-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "a205f475-340d-4ec5-a860-d4a5c2ceb39e" (UID: "a205f475-340d-4ec5-a860-d4a5c2ceb39e"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.203784 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/a205f475-340d-4ec5-a860-d4a5c2ceb39e-run-ovn\") pod \"a205f475-340d-4ec5-a860-d4a5c2ceb39e\" (UID: \"a205f475-340d-4ec5-a860-d4a5c2ceb39e\") " Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.203814 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/a205f475-340d-4ec5-a860-d4a5c2ceb39e-run-systemd\") pod \"a205f475-340d-4ec5-a860-d4a5c2ceb39e\" (UID: \"a205f475-340d-4ec5-a860-d4a5c2ceb39e\") " Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.203743 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a205f475-340d-4ec5-a860-d4a5c2ceb39e-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "a205f475-340d-4ec5-a860-d4a5c2ceb39e" (UID: "a205f475-340d-4ec5-a860-d4a5c2ceb39e"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.204461 4631 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/a205f475-340d-4ec5-a860-d4a5c2ceb39e-run-ovn\") on node \"crc\" DevicePath \"\"" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.204486 4631 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a205f475-340d-4ec5-a860-d4a5c2ceb39e-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.204502 4631 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/a205f475-340d-4ec5-a860-d4a5c2ceb39e-host-run-netns\") on node \"crc\" DevicePath \"\"" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.204515 4631 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/a205f475-340d-4ec5-a860-d4a5c2ceb39e-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.204529 4631 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/a205f475-340d-4ec5-a860-d4a5c2ceb39e-systemd-units\") on node \"crc\" DevicePath \"\"" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.204540 4631 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/a205f475-340d-4ec5-a860-d4a5c2ceb39e-host-kubelet\") on node \"crc\" DevicePath \"\"" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.204552 4631 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/a205f475-340d-4ec5-a860-d4a5c2ceb39e-host-cni-netd\") on node \"crc\" DevicePath \"\"" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.204564 4631 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a205f475-340d-4ec5-a860-d4a5c2ceb39e-run-openvswitch\") on node \"crc\" DevicePath \"\"" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.204578 4631 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/a205f475-340d-4ec5-a860-d4a5c2ceb39e-host-cni-bin\") on node \"crc\" DevicePath \"\"" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.204592 4631 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/a205f475-340d-4ec5-a860-d4a5c2ceb39e-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.204607 4631 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/a205f475-340d-4ec5-a860-d4a5c2ceb39e-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.204619 4631 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/a205f475-340d-4ec5-a860-d4a5c2ceb39e-host-slash\") on node \"crc\" DevicePath \"\"" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.204632 4631 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/a205f475-340d-4ec5-a860-d4a5c2ceb39e-env-overrides\") on node \"crc\" DevicePath \"\"" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.204643 4631 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/a205f475-340d-4ec5-a860-d4a5c2ceb39e-log-socket\") on node \"crc\" DevicePath \"\"" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.204656 4631 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a205f475-340d-4ec5-a860-d4a5c2ceb39e-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.204666 4631 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/a205f475-340d-4ec5-a860-d4a5c2ceb39e-ovnkube-config\") on node \"crc\" DevicePath \"\"" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.210546 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a205f475-340d-4ec5-a860-d4a5c2ceb39e-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "a205f475-340d-4ec5-a860-d4a5c2ceb39e" (UID: "a205f475-340d-4ec5-a860-d4a5c2ceb39e"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.222742 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a205f475-340d-4ec5-a860-d4a5c2ceb39e-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "a205f475-340d-4ec5-a860-d4a5c2ceb39e" (UID: "a205f475-340d-4ec5-a860-d4a5c2ceb39e"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.262004 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-wjsdr"] Nov 28 13:29:03 crc kubenswrapper[4631]: E1128 13:29:03.262375 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a205f475-340d-4ec5-a860-d4a5c2ceb39e" containerName="nbdb" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.262401 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="a205f475-340d-4ec5-a860-d4a5c2ceb39e" containerName="nbdb" Nov 28 13:29:03 crc kubenswrapper[4631]: E1128 13:29:03.262414 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a205f475-340d-4ec5-a860-d4a5c2ceb39e" containerName="ovnkube-controller" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.262429 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="a205f475-340d-4ec5-a860-d4a5c2ceb39e" containerName="ovnkube-controller" Nov 28 13:29:03 crc kubenswrapper[4631]: E1128 13:29:03.262440 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a205f475-340d-4ec5-a860-d4a5c2ceb39e" containerName="sbdb" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.262449 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="a205f475-340d-4ec5-a860-d4a5c2ceb39e" containerName="sbdb" Nov 28 13:29:03 crc kubenswrapper[4631]: E1128 13:29:03.262467 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a205f475-340d-4ec5-a860-d4a5c2ceb39e" containerName="kube-rbac-proxy-node" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.262477 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="a205f475-340d-4ec5-a860-d4a5c2ceb39e" containerName="kube-rbac-proxy-node" Nov 28 13:29:03 crc kubenswrapper[4631]: E1128 13:29:03.262487 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a205f475-340d-4ec5-a860-d4a5c2ceb39e" containerName="kubecfg-setup" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.262496 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="a205f475-340d-4ec5-a860-d4a5c2ceb39e" containerName="kubecfg-setup" Nov 28 13:29:03 crc kubenswrapper[4631]: E1128 13:29:03.262511 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a205f475-340d-4ec5-a860-d4a5c2ceb39e" containerName="ovn-controller" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.262521 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="a205f475-340d-4ec5-a860-d4a5c2ceb39e" containerName="ovn-controller" Nov 28 13:29:03 crc kubenswrapper[4631]: E1128 13:29:03.262533 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a205f475-340d-4ec5-a860-d4a5c2ceb39e" containerName="northd" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.262542 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="a205f475-340d-4ec5-a860-d4a5c2ceb39e" containerName="northd" Nov 28 13:29:03 crc kubenswrapper[4631]: E1128 13:29:03.262559 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a205f475-340d-4ec5-a860-d4a5c2ceb39e" containerName="ovn-acl-logging" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.262567 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="a205f475-340d-4ec5-a860-d4a5c2ceb39e" containerName="ovn-acl-logging" Nov 28 13:29:03 crc kubenswrapper[4631]: E1128 13:29:03.262590 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a205f475-340d-4ec5-a860-d4a5c2ceb39e" containerName="kube-rbac-proxy-ovn-metrics" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.262599 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="a205f475-340d-4ec5-a860-d4a5c2ceb39e" containerName="kube-rbac-proxy-ovn-metrics" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.262773 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="a205f475-340d-4ec5-a860-d4a5c2ceb39e" containerName="ovn-controller" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.262790 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="a205f475-340d-4ec5-a860-d4a5c2ceb39e" containerName="ovnkube-controller" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.262808 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="a205f475-340d-4ec5-a860-d4a5c2ceb39e" containerName="ovn-acl-logging" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.262821 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="a205f475-340d-4ec5-a860-d4a5c2ceb39e" containerName="sbdb" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.262842 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="a205f475-340d-4ec5-a860-d4a5c2ceb39e" containerName="kube-rbac-proxy-ovn-metrics" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.262855 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="a205f475-340d-4ec5-a860-d4a5c2ceb39e" containerName="northd" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.262864 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="a205f475-340d-4ec5-a860-d4a5c2ceb39e" containerName="nbdb" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.262875 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="a205f475-340d-4ec5-a860-d4a5c2ceb39e" containerName="kube-rbac-proxy-node" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.265489 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-wjsdr" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.305224 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qdzgl\" (UniqueName: \"kubernetes.io/projected/a205f475-340d-4ec5-a860-d4a5c2ceb39e-kube-api-access-qdzgl\") pod \"a205f475-340d-4ec5-a860-d4a5c2ceb39e\" (UID: \"a205f475-340d-4ec5-a860-d4a5c2ceb39e\") " Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.305272 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/a205f475-340d-4ec5-a860-d4a5c2ceb39e-node-log\") pod \"a205f475-340d-4ec5-a860-d4a5c2ceb39e\" (UID: \"a205f475-340d-4ec5-a860-d4a5c2ceb39e\") " Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.305451 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/0d38965f-6676-4e8e-9a4b-da5364a20e4e-ovnkube-script-lib\") pod \"ovnkube-node-wjsdr\" (UID: \"0d38965f-6676-4e8e-9a4b-da5364a20e4e\") " pod="openshift-ovn-kubernetes/ovnkube-node-wjsdr" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.305476 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/0d38965f-6676-4e8e-9a4b-da5364a20e4e-ovn-node-metrics-cert\") pod \"ovnkube-node-wjsdr\" (UID: \"0d38965f-6676-4e8e-9a4b-da5364a20e4e\") " pod="openshift-ovn-kubernetes/ovnkube-node-wjsdr" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.305496 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/0d38965f-6676-4e8e-9a4b-da5364a20e4e-host-cni-bin\") pod \"ovnkube-node-wjsdr\" (UID: \"0d38965f-6676-4e8e-9a4b-da5364a20e4e\") " pod="openshift-ovn-kubernetes/ovnkube-node-wjsdr" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.305519 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/0d38965f-6676-4e8e-9a4b-da5364a20e4e-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-wjsdr\" (UID: \"0d38965f-6676-4e8e-9a4b-da5364a20e4e\") " pod="openshift-ovn-kubernetes/ovnkube-node-wjsdr" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.305547 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/0d38965f-6676-4e8e-9a4b-da5364a20e4e-env-overrides\") pod \"ovnkube-node-wjsdr\" (UID: \"0d38965f-6676-4e8e-9a4b-da5364a20e4e\") " pod="openshift-ovn-kubernetes/ovnkube-node-wjsdr" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.305572 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/0d38965f-6676-4e8e-9a4b-da5364a20e4e-log-socket\") pod \"ovnkube-node-wjsdr\" (UID: \"0d38965f-6676-4e8e-9a4b-da5364a20e4e\") " pod="openshift-ovn-kubernetes/ovnkube-node-wjsdr" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.305594 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/0d38965f-6676-4e8e-9a4b-da5364a20e4e-var-lib-openvswitch\") pod \"ovnkube-node-wjsdr\" (UID: \"0d38965f-6676-4e8e-9a4b-da5364a20e4e\") " pod="openshift-ovn-kubernetes/ovnkube-node-wjsdr" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.305610 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/0d38965f-6676-4e8e-9a4b-da5364a20e4e-run-openvswitch\") pod \"ovnkube-node-wjsdr\" (UID: \"0d38965f-6676-4e8e-9a4b-da5364a20e4e\") " pod="openshift-ovn-kubernetes/ovnkube-node-wjsdr" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.305633 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/0d38965f-6676-4e8e-9a4b-da5364a20e4e-run-systemd\") pod \"ovnkube-node-wjsdr\" (UID: \"0d38965f-6676-4e8e-9a4b-da5364a20e4e\") " pod="openshift-ovn-kubernetes/ovnkube-node-wjsdr" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.305651 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/0d38965f-6676-4e8e-9a4b-da5364a20e4e-node-log\") pod \"ovnkube-node-wjsdr\" (UID: \"0d38965f-6676-4e8e-9a4b-da5364a20e4e\") " pod="openshift-ovn-kubernetes/ovnkube-node-wjsdr" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.305671 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jlp4j\" (UniqueName: \"kubernetes.io/projected/0d38965f-6676-4e8e-9a4b-da5364a20e4e-kube-api-access-jlp4j\") pod \"ovnkube-node-wjsdr\" (UID: \"0d38965f-6676-4e8e-9a4b-da5364a20e4e\") " pod="openshift-ovn-kubernetes/ovnkube-node-wjsdr" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.305691 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/0d38965f-6676-4e8e-9a4b-da5364a20e4e-host-run-ovn-kubernetes\") pod \"ovnkube-node-wjsdr\" (UID: \"0d38965f-6676-4e8e-9a4b-da5364a20e4e\") " pod="openshift-ovn-kubernetes/ovnkube-node-wjsdr" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.305712 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/0d38965f-6676-4e8e-9a4b-da5364a20e4e-systemd-units\") pod \"ovnkube-node-wjsdr\" (UID: \"0d38965f-6676-4e8e-9a4b-da5364a20e4e\") " pod="openshift-ovn-kubernetes/ovnkube-node-wjsdr" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.305734 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/0d38965f-6676-4e8e-9a4b-da5364a20e4e-host-run-netns\") pod \"ovnkube-node-wjsdr\" (UID: \"0d38965f-6676-4e8e-9a4b-da5364a20e4e\") " pod="openshift-ovn-kubernetes/ovnkube-node-wjsdr" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.305760 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/0d38965f-6676-4e8e-9a4b-da5364a20e4e-host-kubelet\") pod \"ovnkube-node-wjsdr\" (UID: \"0d38965f-6676-4e8e-9a4b-da5364a20e4e\") " pod="openshift-ovn-kubernetes/ovnkube-node-wjsdr" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.305781 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/0d38965f-6676-4e8e-9a4b-da5364a20e4e-etc-openvswitch\") pod \"ovnkube-node-wjsdr\" (UID: \"0d38965f-6676-4e8e-9a4b-da5364a20e4e\") " pod="openshift-ovn-kubernetes/ovnkube-node-wjsdr" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.305806 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/0d38965f-6676-4e8e-9a4b-da5364a20e4e-run-ovn\") pod \"ovnkube-node-wjsdr\" (UID: \"0d38965f-6676-4e8e-9a4b-da5364a20e4e\") " pod="openshift-ovn-kubernetes/ovnkube-node-wjsdr" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.305829 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/0d38965f-6676-4e8e-9a4b-da5364a20e4e-ovnkube-config\") pod \"ovnkube-node-wjsdr\" (UID: \"0d38965f-6676-4e8e-9a4b-da5364a20e4e\") " pod="openshift-ovn-kubernetes/ovnkube-node-wjsdr" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.305851 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/0d38965f-6676-4e8e-9a4b-da5364a20e4e-host-cni-netd\") pod \"ovnkube-node-wjsdr\" (UID: \"0d38965f-6676-4e8e-9a4b-da5364a20e4e\") " pod="openshift-ovn-kubernetes/ovnkube-node-wjsdr" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.305870 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/0d38965f-6676-4e8e-9a4b-da5364a20e4e-host-slash\") pod \"ovnkube-node-wjsdr\" (UID: \"0d38965f-6676-4e8e-9a4b-da5364a20e4e\") " pod="openshift-ovn-kubernetes/ovnkube-node-wjsdr" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.305908 4631 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/a205f475-340d-4ec5-a860-d4a5c2ceb39e-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.305938 4631 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/a205f475-340d-4ec5-a860-d4a5c2ceb39e-run-systemd\") on node \"crc\" DevicePath \"\"" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.306672 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a205f475-340d-4ec5-a860-d4a5c2ceb39e-node-log" (OuterVolumeSpecName: "node-log") pod "a205f475-340d-4ec5-a860-d4a5c2ceb39e" (UID: "a205f475-340d-4ec5-a860-d4a5c2ceb39e"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.318844 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a205f475-340d-4ec5-a860-d4a5c2ceb39e-kube-api-access-qdzgl" (OuterVolumeSpecName: "kube-api-access-qdzgl") pod "a205f475-340d-4ec5-a860-d4a5c2ceb39e" (UID: "a205f475-340d-4ec5-a860-d4a5c2ceb39e"). InnerVolumeSpecName "kube-api-access-qdzgl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.407196 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/0d38965f-6676-4e8e-9a4b-da5364a20e4e-run-ovn\") pod \"ovnkube-node-wjsdr\" (UID: \"0d38965f-6676-4e8e-9a4b-da5364a20e4e\") " pod="openshift-ovn-kubernetes/ovnkube-node-wjsdr" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.407270 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/0d38965f-6676-4e8e-9a4b-da5364a20e4e-ovnkube-config\") pod \"ovnkube-node-wjsdr\" (UID: \"0d38965f-6676-4e8e-9a4b-da5364a20e4e\") " pod="openshift-ovn-kubernetes/ovnkube-node-wjsdr" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.407321 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/0d38965f-6676-4e8e-9a4b-da5364a20e4e-host-cni-netd\") pod \"ovnkube-node-wjsdr\" (UID: \"0d38965f-6676-4e8e-9a4b-da5364a20e4e\") " pod="openshift-ovn-kubernetes/ovnkube-node-wjsdr" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.407345 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/0d38965f-6676-4e8e-9a4b-da5364a20e4e-host-slash\") pod \"ovnkube-node-wjsdr\" (UID: \"0d38965f-6676-4e8e-9a4b-da5364a20e4e\") " pod="openshift-ovn-kubernetes/ovnkube-node-wjsdr" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.407370 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/0d38965f-6676-4e8e-9a4b-da5364a20e4e-ovnkube-script-lib\") pod \"ovnkube-node-wjsdr\" (UID: \"0d38965f-6676-4e8e-9a4b-da5364a20e4e\") " pod="openshift-ovn-kubernetes/ovnkube-node-wjsdr" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.407396 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/0d38965f-6676-4e8e-9a4b-da5364a20e4e-ovn-node-metrics-cert\") pod \"ovnkube-node-wjsdr\" (UID: \"0d38965f-6676-4e8e-9a4b-da5364a20e4e\") " pod="openshift-ovn-kubernetes/ovnkube-node-wjsdr" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.407422 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/0d38965f-6676-4e8e-9a4b-da5364a20e4e-host-cni-bin\") pod \"ovnkube-node-wjsdr\" (UID: \"0d38965f-6676-4e8e-9a4b-da5364a20e4e\") " pod="openshift-ovn-kubernetes/ovnkube-node-wjsdr" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.407445 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/0d38965f-6676-4e8e-9a4b-da5364a20e4e-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-wjsdr\" (UID: \"0d38965f-6676-4e8e-9a4b-da5364a20e4e\") " pod="openshift-ovn-kubernetes/ovnkube-node-wjsdr" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.407483 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/0d38965f-6676-4e8e-9a4b-da5364a20e4e-env-overrides\") pod \"ovnkube-node-wjsdr\" (UID: \"0d38965f-6676-4e8e-9a4b-da5364a20e4e\") " pod="openshift-ovn-kubernetes/ovnkube-node-wjsdr" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.407510 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/0d38965f-6676-4e8e-9a4b-da5364a20e4e-log-socket\") pod \"ovnkube-node-wjsdr\" (UID: \"0d38965f-6676-4e8e-9a4b-da5364a20e4e\") " pod="openshift-ovn-kubernetes/ovnkube-node-wjsdr" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.407538 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/0d38965f-6676-4e8e-9a4b-da5364a20e4e-var-lib-openvswitch\") pod \"ovnkube-node-wjsdr\" (UID: \"0d38965f-6676-4e8e-9a4b-da5364a20e4e\") " pod="openshift-ovn-kubernetes/ovnkube-node-wjsdr" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.407565 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/0d38965f-6676-4e8e-9a4b-da5364a20e4e-run-openvswitch\") pod \"ovnkube-node-wjsdr\" (UID: \"0d38965f-6676-4e8e-9a4b-da5364a20e4e\") " pod="openshift-ovn-kubernetes/ovnkube-node-wjsdr" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.407594 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/0d38965f-6676-4e8e-9a4b-da5364a20e4e-run-systemd\") pod \"ovnkube-node-wjsdr\" (UID: \"0d38965f-6676-4e8e-9a4b-da5364a20e4e\") " pod="openshift-ovn-kubernetes/ovnkube-node-wjsdr" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.407617 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/0d38965f-6676-4e8e-9a4b-da5364a20e4e-node-log\") pod \"ovnkube-node-wjsdr\" (UID: \"0d38965f-6676-4e8e-9a4b-da5364a20e4e\") " pod="openshift-ovn-kubernetes/ovnkube-node-wjsdr" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.407638 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/0d38965f-6676-4e8e-9a4b-da5364a20e4e-host-run-ovn-kubernetes\") pod \"ovnkube-node-wjsdr\" (UID: \"0d38965f-6676-4e8e-9a4b-da5364a20e4e\") " pod="openshift-ovn-kubernetes/ovnkube-node-wjsdr" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.407659 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jlp4j\" (UniqueName: \"kubernetes.io/projected/0d38965f-6676-4e8e-9a4b-da5364a20e4e-kube-api-access-jlp4j\") pod \"ovnkube-node-wjsdr\" (UID: \"0d38965f-6676-4e8e-9a4b-da5364a20e4e\") " pod="openshift-ovn-kubernetes/ovnkube-node-wjsdr" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.407694 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/0d38965f-6676-4e8e-9a4b-da5364a20e4e-systemd-units\") pod \"ovnkube-node-wjsdr\" (UID: \"0d38965f-6676-4e8e-9a4b-da5364a20e4e\") " pod="openshift-ovn-kubernetes/ovnkube-node-wjsdr" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.407717 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/0d38965f-6676-4e8e-9a4b-da5364a20e4e-host-run-netns\") pod \"ovnkube-node-wjsdr\" (UID: \"0d38965f-6676-4e8e-9a4b-da5364a20e4e\") " pod="openshift-ovn-kubernetes/ovnkube-node-wjsdr" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.407749 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/0d38965f-6676-4e8e-9a4b-da5364a20e4e-host-kubelet\") pod \"ovnkube-node-wjsdr\" (UID: \"0d38965f-6676-4e8e-9a4b-da5364a20e4e\") " pod="openshift-ovn-kubernetes/ovnkube-node-wjsdr" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.407772 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/0d38965f-6676-4e8e-9a4b-da5364a20e4e-etc-openvswitch\") pod \"ovnkube-node-wjsdr\" (UID: \"0d38965f-6676-4e8e-9a4b-da5364a20e4e\") " pod="openshift-ovn-kubernetes/ovnkube-node-wjsdr" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.407825 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qdzgl\" (UniqueName: \"kubernetes.io/projected/a205f475-340d-4ec5-a860-d4a5c2ceb39e-kube-api-access-qdzgl\") on node \"crc\" DevicePath \"\"" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.407840 4631 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/a205f475-340d-4ec5-a860-d4a5c2ceb39e-node-log\") on node \"crc\" DevicePath \"\"" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.407896 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/0d38965f-6676-4e8e-9a4b-da5364a20e4e-etc-openvswitch\") pod \"ovnkube-node-wjsdr\" (UID: \"0d38965f-6676-4e8e-9a4b-da5364a20e4e\") " pod="openshift-ovn-kubernetes/ovnkube-node-wjsdr" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.407951 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/0d38965f-6676-4e8e-9a4b-da5364a20e4e-run-ovn\") pod \"ovnkube-node-wjsdr\" (UID: \"0d38965f-6676-4e8e-9a4b-da5364a20e4e\") " pod="openshift-ovn-kubernetes/ovnkube-node-wjsdr" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.408960 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/0d38965f-6676-4e8e-9a4b-da5364a20e4e-ovnkube-config\") pod \"ovnkube-node-wjsdr\" (UID: \"0d38965f-6676-4e8e-9a4b-da5364a20e4e\") " pod="openshift-ovn-kubernetes/ovnkube-node-wjsdr" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.409039 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/0d38965f-6676-4e8e-9a4b-da5364a20e4e-host-cni-netd\") pod \"ovnkube-node-wjsdr\" (UID: \"0d38965f-6676-4e8e-9a4b-da5364a20e4e\") " pod="openshift-ovn-kubernetes/ovnkube-node-wjsdr" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.409077 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/0d38965f-6676-4e8e-9a4b-da5364a20e4e-host-slash\") pod \"ovnkube-node-wjsdr\" (UID: \"0d38965f-6676-4e8e-9a4b-da5364a20e4e\") " pod="openshift-ovn-kubernetes/ovnkube-node-wjsdr" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.409418 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/0d38965f-6676-4e8e-9a4b-da5364a20e4e-run-openvswitch\") pod \"ovnkube-node-wjsdr\" (UID: \"0d38965f-6676-4e8e-9a4b-da5364a20e4e\") " pod="openshift-ovn-kubernetes/ovnkube-node-wjsdr" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.409575 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/0d38965f-6676-4e8e-9a4b-da5364a20e4e-systemd-units\") pod \"ovnkube-node-wjsdr\" (UID: \"0d38965f-6676-4e8e-9a4b-da5364a20e4e\") " pod="openshift-ovn-kubernetes/ovnkube-node-wjsdr" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.409616 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/0d38965f-6676-4e8e-9a4b-da5364a20e4e-host-kubelet\") pod \"ovnkube-node-wjsdr\" (UID: \"0d38965f-6676-4e8e-9a4b-da5364a20e4e\") " pod="openshift-ovn-kubernetes/ovnkube-node-wjsdr" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.409608 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/0d38965f-6676-4e8e-9a4b-da5364a20e4e-host-run-netns\") pod \"ovnkube-node-wjsdr\" (UID: \"0d38965f-6676-4e8e-9a4b-da5364a20e4e\") " pod="openshift-ovn-kubernetes/ovnkube-node-wjsdr" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.409627 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/0d38965f-6676-4e8e-9a4b-da5364a20e4e-node-log\") pod \"ovnkube-node-wjsdr\" (UID: \"0d38965f-6676-4e8e-9a4b-da5364a20e4e\") " pod="openshift-ovn-kubernetes/ovnkube-node-wjsdr" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.409675 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/0d38965f-6676-4e8e-9a4b-da5364a20e4e-run-systemd\") pod \"ovnkube-node-wjsdr\" (UID: \"0d38965f-6676-4e8e-9a4b-da5364a20e4e\") " pod="openshift-ovn-kubernetes/ovnkube-node-wjsdr" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.409676 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/0d38965f-6676-4e8e-9a4b-da5364a20e4e-ovnkube-script-lib\") pod \"ovnkube-node-wjsdr\" (UID: \"0d38965f-6676-4e8e-9a4b-da5364a20e4e\") " pod="openshift-ovn-kubernetes/ovnkube-node-wjsdr" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.409657 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/0d38965f-6676-4e8e-9a4b-da5364a20e4e-host-cni-bin\") pod \"ovnkube-node-wjsdr\" (UID: \"0d38965f-6676-4e8e-9a4b-da5364a20e4e\") " pod="openshift-ovn-kubernetes/ovnkube-node-wjsdr" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.409714 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/0d38965f-6676-4e8e-9a4b-da5364a20e4e-host-run-ovn-kubernetes\") pod \"ovnkube-node-wjsdr\" (UID: \"0d38965f-6676-4e8e-9a4b-da5364a20e4e\") " pod="openshift-ovn-kubernetes/ovnkube-node-wjsdr" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.409719 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/0d38965f-6676-4e8e-9a4b-da5364a20e4e-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-wjsdr\" (UID: \"0d38965f-6676-4e8e-9a4b-da5364a20e4e\") " pod="openshift-ovn-kubernetes/ovnkube-node-wjsdr" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.409735 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/0d38965f-6676-4e8e-9a4b-da5364a20e4e-log-socket\") pod \"ovnkube-node-wjsdr\" (UID: \"0d38965f-6676-4e8e-9a4b-da5364a20e4e\") " pod="openshift-ovn-kubernetes/ovnkube-node-wjsdr" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.409748 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/0d38965f-6676-4e8e-9a4b-da5364a20e4e-var-lib-openvswitch\") pod \"ovnkube-node-wjsdr\" (UID: \"0d38965f-6676-4e8e-9a4b-da5364a20e4e\") " pod="openshift-ovn-kubernetes/ovnkube-node-wjsdr" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.410215 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/0d38965f-6676-4e8e-9a4b-da5364a20e4e-env-overrides\") pod \"ovnkube-node-wjsdr\" (UID: \"0d38965f-6676-4e8e-9a4b-da5364a20e4e\") " pod="openshift-ovn-kubernetes/ovnkube-node-wjsdr" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.414098 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/0d38965f-6676-4e8e-9a4b-da5364a20e4e-ovn-node-metrics-cert\") pod \"ovnkube-node-wjsdr\" (UID: \"0d38965f-6676-4e8e-9a4b-da5364a20e4e\") " pod="openshift-ovn-kubernetes/ovnkube-node-wjsdr" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.425830 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jlp4j\" (UniqueName: \"kubernetes.io/projected/0d38965f-6676-4e8e-9a4b-da5364a20e4e-kube-api-access-jlp4j\") pod \"ovnkube-node-wjsdr\" (UID: \"0d38965f-6676-4e8e-9a4b-da5364a20e4e\") " pod="openshift-ovn-kubernetes/ovnkube-node-wjsdr" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.441281 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-5655c58dd6-l6tzn" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.522895 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-j8d66_9a346894-644f-4359-baa5-23bb2d0acc5f/kube-multus/0.log" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.522947 4631 generic.go:334] "Generic (PLEG): container finished" podID="9a346894-644f-4359-baa5-23bb2d0acc5f" containerID="1e499b5d5f702b05bb24f616fb7d5ba1799d9b03e06c9efbfbbdcfd86e00e753" exitCode=2 Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.523004 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-j8d66" event={"ID":"9a346894-644f-4359-baa5-23bb2d0acc5f","Type":"ContainerDied","Data":"1e499b5d5f702b05bb24f616fb7d5ba1799d9b03e06c9efbfbbdcfd86e00e753"} Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.523394 4631 scope.go:117] "RemoveContainer" containerID="1e499b5d5f702b05bb24f616fb7d5ba1799d9b03e06c9efbfbbdcfd86e00e753" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.531927 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-268tw_a205f475-340d-4ec5-a860-d4a5c2ceb39e/ovn-acl-logging/0.log" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.532724 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-268tw_a205f475-340d-4ec5-a860-d4a5c2ceb39e/ovn-controller/0.log" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.533431 4631 generic.go:334] "Generic (PLEG): container finished" podID="a205f475-340d-4ec5-a860-d4a5c2ceb39e" containerID="da4a7ce2b9fcb91df9dab6ba86046a998e310a5cc62a8eefcd98ea01a26f8db1" exitCode=0 Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.533477 4631 generic.go:334] "Generic (PLEG): container finished" podID="a205f475-340d-4ec5-a860-d4a5c2ceb39e" containerID="9c674448c309477eb36de0eb0312fde682c3248fb66ed1d1d4100dc114fda786" exitCode=0 Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.533487 4631 generic.go:334] "Generic (PLEG): container finished" podID="a205f475-340d-4ec5-a860-d4a5c2ceb39e" containerID="0ab2d69688bf305aeb7ece5c432d87adcb06eaca6c90ef5aa4e73b7fe477246b" exitCode=0 Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.533495 4631 generic.go:334] "Generic (PLEG): container finished" podID="a205f475-340d-4ec5-a860-d4a5c2ceb39e" containerID="794ab47dc4660d76accb0456995b602cd5cdfc08616f63427947aba9e725eaf7" exitCode=0 Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.533506 4631 generic.go:334] "Generic (PLEG): container finished" podID="a205f475-340d-4ec5-a860-d4a5c2ceb39e" containerID="016163e216d642018075fa0e9ab61fa630fa9718030246855fd44a8104c3d6e6" exitCode=0 Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.533534 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-268tw" event={"ID":"a205f475-340d-4ec5-a860-d4a5c2ceb39e","Type":"ContainerDied","Data":"da4a7ce2b9fcb91df9dab6ba86046a998e310a5cc62a8eefcd98ea01a26f8db1"} Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.533587 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-268tw" event={"ID":"a205f475-340d-4ec5-a860-d4a5c2ceb39e","Type":"ContainerDied","Data":"9c674448c309477eb36de0eb0312fde682c3248fb66ed1d1d4100dc114fda786"} Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.533603 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-268tw" event={"ID":"a205f475-340d-4ec5-a860-d4a5c2ceb39e","Type":"ContainerDied","Data":"0ab2d69688bf305aeb7ece5c432d87adcb06eaca6c90ef5aa4e73b7fe477246b"} Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.533617 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-268tw" event={"ID":"a205f475-340d-4ec5-a860-d4a5c2ceb39e","Type":"ContainerDied","Data":"794ab47dc4660d76accb0456995b602cd5cdfc08616f63427947aba9e725eaf7"} Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.533636 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-268tw" event={"ID":"a205f475-340d-4ec5-a860-d4a5c2ceb39e","Type":"ContainerDied","Data":"016163e216d642018075fa0e9ab61fa630fa9718030246855fd44a8104c3d6e6"} Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.533639 4631 scope.go:117] "RemoveContainer" containerID="da4a7ce2b9fcb91df9dab6ba86046a998e310a5cc62a8eefcd98ea01a26f8db1" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.533648 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-268tw" event={"ID":"a205f475-340d-4ec5-a860-d4a5c2ceb39e","Type":"ContainerDied","Data":"e5085439ece35186f2692a8cf5be0b011fbf3f212f4f772b674c70649486c89b"} Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.533662 4631 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"95aeff4871c229569b0c99ff9a6cd17c63f10c1c89d9f04c533dce699c89ce31"} Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.533680 4631 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e571a03b3c52184c8fa59e1ab6fff8d9118f114dd6c156638ba36fb7a0d2fe40"} Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.533686 4631 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"542662c37344400bc1d05230e709812104866735429d37ed0d7894050e7a54e8"} Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.533552 4631 generic.go:334] "Generic (PLEG): container finished" podID="a205f475-340d-4ec5-a860-d4a5c2ceb39e" containerID="e5085439ece35186f2692a8cf5be0b011fbf3f212f4f772b674c70649486c89b" exitCode=0 Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.533713 4631 generic.go:334] "Generic (PLEG): container finished" podID="a205f475-340d-4ec5-a860-d4a5c2ceb39e" containerID="95aeff4871c229569b0c99ff9a6cd17c63f10c1c89d9f04c533dce699c89ce31" exitCode=143 Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.533728 4631 generic.go:334] "Generic (PLEG): container finished" podID="a205f475-340d-4ec5-a860-d4a5c2ceb39e" containerID="e571a03b3c52184c8fa59e1ab6fff8d9118f114dd6c156638ba36fb7a0d2fe40" exitCode=143 Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.533746 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-268tw" event={"ID":"a205f475-340d-4ec5-a860-d4a5c2ceb39e","Type":"ContainerDied","Data":"95aeff4871c229569b0c99ff9a6cd17c63f10c1c89d9f04c533dce699c89ce31"} Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.533760 4631 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"da4a7ce2b9fcb91df9dab6ba86046a998e310a5cc62a8eefcd98ea01a26f8db1"} Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.533767 4631 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9c674448c309477eb36de0eb0312fde682c3248fb66ed1d1d4100dc114fda786"} Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.533774 4631 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"0ab2d69688bf305aeb7ece5c432d87adcb06eaca6c90ef5aa4e73b7fe477246b"} Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.533779 4631 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"794ab47dc4660d76accb0456995b602cd5cdfc08616f63427947aba9e725eaf7"} Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.533785 4631 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"016163e216d642018075fa0e9ab61fa630fa9718030246855fd44a8104c3d6e6"} Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.533791 4631 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e5085439ece35186f2692a8cf5be0b011fbf3f212f4f772b674c70649486c89b"} Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.533493 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-268tw" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.533797 4631 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"95aeff4871c229569b0c99ff9a6cd17c63f10c1c89d9f04c533dce699c89ce31"} Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.533941 4631 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e571a03b3c52184c8fa59e1ab6fff8d9118f114dd6c156638ba36fb7a0d2fe40"} Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.533949 4631 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"542662c37344400bc1d05230e709812104866735429d37ed0d7894050e7a54e8"} Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.533959 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-268tw" event={"ID":"a205f475-340d-4ec5-a860-d4a5c2ceb39e","Type":"ContainerDied","Data":"e571a03b3c52184c8fa59e1ab6fff8d9118f114dd6c156638ba36fb7a0d2fe40"} Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.533969 4631 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"da4a7ce2b9fcb91df9dab6ba86046a998e310a5cc62a8eefcd98ea01a26f8db1"} Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.533976 4631 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9c674448c309477eb36de0eb0312fde682c3248fb66ed1d1d4100dc114fda786"} Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.533985 4631 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"0ab2d69688bf305aeb7ece5c432d87adcb06eaca6c90ef5aa4e73b7fe477246b"} Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.533992 4631 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"794ab47dc4660d76accb0456995b602cd5cdfc08616f63427947aba9e725eaf7"} Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.533998 4631 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"016163e216d642018075fa0e9ab61fa630fa9718030246855fd44a8104c3d6e6"} Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.534005 4631 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e5085439ece35186f2692a8cf5be0b011fbf3f212f4f772b674c70649486c89b"} Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.534014 4631 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"95aeff4871c229569b0c99ff9a6cd17c63f10c1c89d9f04c533dce699c89ce31"} Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.534020 4631 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e571a03b3c52184c8fa59e1ab6fff8d9118f114dd6c156638ba36fb7a0d2fe40"} Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.534026 4631 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"542662c37344400bc1d05230e709812104866735429d37ed0d7894050e7a54e8"} Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.534034 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-268tw" event={"ID":"a205f475-340d-4ec5-a860-d4a5c2ceb39e","Type":"ContainerDied","Data":"11131235ce93081bf60d5d43d90a0fe910b89daf18e38b53128dcd441a1c0194"} Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.534042 4631 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"da4a7ce2b9fcb91df9dab6ba86046a998e310a5cc62a8eefcd98ea01a26f8db1"} Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.534066 4631 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9c674448c309477eb36de0eb0312fde682c3248fb66ed1d1d4100dc114fda786"} Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.534073 4631 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"0ab2d69688bf305aeb7ece5c432d87adcb06eaca6c90ef5aa4e73b7fe477246b"} Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.534079 4631 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"794ab47dc4660d76accb0456995b602cd5cdfc08616f63427947aba9e725eaf7"} Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.534084 4631 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"016163e216d642018075fa0e9ab61fa630fa9718030246855fd44a8104c3d6e6"} Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.534091 4631 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e5085439ece35186f2692a8cf5be0b011fbf3f212f4f772b674c70649486c89b"} Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.534097 4631 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"95aeff4871c229569b0c99ff9a6cd17c63f10c1c89d9f04c533dce699c89ce31"} Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.534103 4631 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e571a03b3c52184c8fa59e1ab6fff8d9118f114dd6c156638ba36fb7a0d2fe40"} Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.534109 4631 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"542662c37344400bc1d05230e709812104866735429d37ed0d7894050e7a54e8"} Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.551252 4631 scope.go:117] "RemoveContainer" containerID="9c674448c309477eb36de0eb0312fde682c3248fb66ed1d1d4100dc114fda786" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.575061 4631 scope.go:117] "RemoveContainer" containerID="0ab2d69688bf305aeb7ece5c432d87adcb06eaca6c90ef5aa4e73b7fe477246b" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.579580 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-wjsdr" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.601384 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-268tw"] Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.613772 4631 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-268tw"] Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.627610 4631 scope.go:117] "RemoveContainer" containerID="794ab47dc4660d76accb0456995b602cd5cdfc08616f63427947aba9e725eaf7" Nov 28 13:29:03 crc kubenswrapper[4631]: W1128 13:29:03.632965 4631 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0d38965f_6676_4e8e_9a4b_da5364a20e4e.slice/crio-9694b775be14852475a382143843c7b34d5a5047d8a2f24dd09d0f11c5f23fd9 WatchSource:0}: Error finding container 9694b775be14852475a382143843c7b34d5a5047d8a2f24dd09d0f11c5f23fd9: Status 404 returned error can't find the container with id 9694b775be14852475a382143843c7b34d5a5047d8a2f24dd09d0f11c5f23fd9 Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.650927 4631 scope.go:117] "RemoveContainer" containerID="016163e216d642018075fa0e9ab61fa630fa9718030246855fd44a8104c3d6e6" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.671070 4631 scope.go:117] "RemoveContainer" containerID="e5085439ece35186f2692a8cf5be0b011fbf3f212f4f772b674c70649486c89b" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.687847 4631 scope.go:117] "RemoveContainer" containerID="95aeff4871c229569b0c99ff9a6cd17c63f10c1c89d9f04c533dce699c89ce31" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.703191 4631 scope.go:117] "RemoveContainer" containerID="e571a03b3c52184c8fa59e1ab6fff8d9118f114dd6c156638ba36fb7a0d2fe40" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.730760 4631 scope.go:117] "RemoveContainer" containerID="542662c37344400bc1d05230e709812104866735429d37ed0d7894050e7a54e8" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.749643 4631 scope.go:117] "RemoveContainer" containerID="da4a7ce2b9fcb91df9dab6ba86046a998e310a5cc62a8eefcd98ea01a26f8db1" Nov 28 13:29:03 crc kubenswrapper[4631]: E1128 13:29:03.750258 4631 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"da4a7ce2b9fcb91df9dab6ba86046a998e310a5cc62a8eefcd98ea01a26f8db1\": container with ID starting with da4a7ce2b9fcb91df9dab6ba86046a998e310a5cc62a8eefcd98ea01a26f8db1 not found: ID does not exist" containerID="da4a7ce2b9fcb91df9dab6ba86046a998e310a5cc62a8eefcd98ea01a26f8db1" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.750376 4631 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"da4a7ce2b9fcb91df9dab6ba86046a998e310a5cc62a8eefcd98ea01a26f8db1"} err="failed to get container status \"da4a7ce2b9fcb91df9dab6ba86046a998e310a5cc62a8eefcd98ea01a26f8db1\": rpc error: code = NotFound desc = could not find container \"da4a7ce2b9fcb91df9dab6ba86046a998e310a5cc62a8eefcd98ea01a26f8db1\": container with ID starting with da4a7ce2b9fcb91df9dab6ba86046a998e310a5cc62a8eefcd98ea01a26f8db1 not found: ID does not exist" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.750454 4631 scope.go:117] "RemoveContainer" containerID="9c674448c309477eb36de0eb0312fde682c3248fb66ed1d1d4100dc114fda786" Nov 28 13:29:03 crc kubenswrapper[4631]: E1128 13:29:03.751255 4631 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9c674448c309477eb36de0eb0312fde682c3248fb66ed1d1d4100dc114fda786\": container with ID starting with 9c674448c309477eb36de0eb0312fde682c3248fb66ed1d1d4100dc114fda786 not found: ID does not exist" containerID="9c674448c309477eb36de0eb0312fde682c3248fb66ed1d1d4100dc114fda786" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.751338 4631 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9c674448c309477eb36de0eb0312fde682c3248fb66ed1d1d4100dc114fda786"} err="failed to get container status \"9c674448c309477eb36de0eb0312fde682c3248fb66ed1d1d4100dc114fda786\": rpc error: code = NotFound desc = could not find container \"9c674448c309477eb36de0eb0312fde682c3248fb66ed1d1d4100dc114fda786\": container with ID starting with 9c674448c309477eb36de0eb0312fde682c3248fb66ed1d1d4100dc114fda786 not found: ID does not exist" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.751385 4631 scope.go:117] "RemoveContainer" containerID="0ab2d69688bf305aeb7ece5c432d87adcb06eaca6c90ef5aa4e73b7fe477246b" Nov 28 13:29:03 crc kubenswrapper[4631]: E1128 13:29:03.751871 4631 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0ab2d69688bf305aeb7ece5c432d87adcb06eaca6c90ef5aa4e73b7fe477246b\": container with ID starting with 0ab2d69688bf305aeb7ece5c432d87adcb06eaca6c90ef5aa4e73b7fe477246b not found: ID does not exist" containerID="0ab2d69688bf305aeb7ece5c432d87adcb06eaca6c90ef5aa4e73b7fe477246b" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.751950 4631 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0ab2d69688bf305aeb7ece5c432d87adcb06eaca6c90ef5aa4e73b7fe477246b"} err="failed to get container status \"0ab2d69688bf305aeb7ece5c432d87adcb06eaca6c90ef5aa4e73b7fe477246b\": rpc error: code = NotFound desc = could not find container \"0ab2d69688bf305aeb7ece5c432d87adcb06eaca6c90ef5aa4e73b7fe477246b\": container with ID starting with 0ab2d69688bf305aeb7ece5c432d87adcb06eaca6c90ef5aa4e73b7fe477246b not found: ID does not exist" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.751987 4631 scope.go:117] "RemoveContainer" containerID="794ab47dc4660d76accb0456995b602cd5cdfc08616f63427947aba9e725eaf7" Nov 28 13:29:03 crc kubenswrapper[4631]: E1128 13:29:03.752598 4631 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"794ab47dc4660d76accb0456995b602cd5cdfc08616f63427947aba9e725eaf7\": container with ID starting with 794ab47dc4660d76accb0456995b602cd5cdfc08616f63427947aba9e725eaf7 not found: ID does not exist" containerID="794ab47dc4660d76accb0456995b602cd5cdfc08616f63427947aba9e725eaf7" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.752634 4631 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"794ab47dc4660d76accb0456995b602cd5cdfc08616f63427947aba9e725eaf7"} err="failed to get container status \"794ab47dc4660d76accb0456995b602cd5cdfc08616f63427947aba9e725eaf7\": rpc error: code = NotFound desc = could not find container \"794ab47dc4660d76accb0456995b602cd5cdfc08616f63427947aba9e725eaf7\": container with ID starting with 794ab47dc4660d76accb0456995b602cd5cdfc08616f63427947aba9e725eaf7 not found: ID does not exist" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.752651 4631 scope.go:117] "RemoveContainer" containerID="016163e216d642018075fa0e9ab61fa630fa9718030246855fd44a8104c3d6e6" Nov 28 13:29:03 crc kubenswrapper[4631]: E1128 13:29:03.753034 4631 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"016163e216d642018075fa0e9ab61fa630fa9718030246855fd44a8104c3d6e6\": container with ID starting with 016163e216d642018075fa0e9ab61fa630fa9718030246855fd44a8104c3d6e6 not found: ID does not exist" containerID="016163e216d642018075fa0e9ab61fa630fa9718030246855fd44a8104c3d6e6" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.753114 4631 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"016163e216d642018075fa0e9ab61fa630fa9718030246855fd44a8104c3d6e6"} err="failed to get container status \"016163e216d642018075fa0e9ab61fa630fa9718030246855fd44a8104c3d6e6\": rpc error: code = NotFound desc = could not find container \"016163e216d642018075fa0e9ab61fa630fa9718030246855fd44a8104c3d6e6\": container with ID starting with 016163e216d642018075fa0e9ab61fa630fa9718030246855fd44a8104c3d6e6 not found: ID does not exist" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.753181 4631 scope.go:117] "RemoveContainer" containerID="e5085439ece35186f2692a8cf5be0b011fbf3f212f4f772b674c70649486c89b" Nov 28 13:29:03 crc kubenswrapper[4631]: E1128 13:29:03.753609 4631 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e5085439ece35186f2692a8cf5be0b011fbf3f212f4f772b674c70649486c89b\": container with ID starting with e5085439ece35186f2692a8cf5be0b011fbf3f212f4f772b674c70649486c89b not found: ID does not exist" containerID="e5085439ece35186f2692a8cf5be0b011fbf3f212f4f772b674c70649486c89b" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.753645 4631 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e5085439ece35186f2692a8cf5be0b011fbf3f212f4f772b674c70649486c89b"} err="failed to get container status \"e5085439ece35186f2692a8cf5be0b011fbf3f212f4f772b674c70649486c89b\": rpc error: code = NotFound desc = could not find container \"e5085439ece35186f2692a8cf5be0b011fbf3f212f4f772b674c70649486c89b\": container with ID starting with e5085439ece35186f2692a8cf5be0b011fbf3f212f4f772b674c70649486c89b not found: ID does not exist" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.753667 4631 scope.go:117] "RemoveContainer" containerID="95aeff4871c229569b0c99ff9a6cd17c63f10c1c89d9f04c533dce699c89ce31" Nov 28 13:29:03 crc kubenswrapper[4631]: E1128 13:29:03.754094 4631 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"95aeff4871c229569b0c99ff9a6cd17c63f10c1c89d9f04c533dce699c89ce31\": container with ID starting with 95aeff4871c229569b0c99ff9a6cd17c63f10c1c89d9f04c533dce699c89ce31 not found: ID does not exist" containerID="95aeff4871c229569b0c99ff9a6cd17c63f10c1c89d9f04c533dce699c89ce31" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.754171 4631 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"95aeff4871c229569b0c99ff9a6cd17c63f10c1c89d9f04c533dce699c89ce31"} err="failed to get container status \"95aeff4871c229569b0c99ff9a6cd17c63f10c1c89d9f04c533dce699c89ce31\": rpc error: code = NotFound desc = could not find container \"95aeff4871c229569b0c99ff9a6cd17c63f10c1c89d9f04c533dce699c89ce31\": container with ID starting with 95aeff4871c229569b0c99ff9a6cd17c63f10c1c89d9f04c533dce699c89ce31 not found: ID does not exist" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.754228 4631 scope.go:117] "RemoveContainer" containerID="e571a03b3c52184c8fa59e1ab6fff8d9118f114dd6c156638ba36fb7a0d2fe40" Nov 28 13:29:03 crc kubenswrapper[4631]: E1128 13:29:03.754748 4631 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e571a03b3c52184c8fa59e1ab6fff8d9118f114dd6c156638ba36fb7a0d2fe40\": container with ID starting with e571a03b3c52184c8fa59e1ab6fff8d9118f114dd6c156638ba36fb7a0d2fe40 not found: ID does not exist" containerID="e571a03b3c52184c8fa59e1ab6fff8d9118f114dd6c156638ba36fb7a0d2fe40" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.754779 4631 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e571a03b3c52184c8fa59e1ab6fff8d9118f114dd6c156638ba36fb7a0d2fe40"} err="failed to get container status \"e571a03b3c52184c8fa59e1ab6fff8d9118f114dd6c156638ba36fb7a0d2fe40\": rpc error: code = NotFound desc = could not find container \"e571a03b3c52184c8fa59e1ab6fff8d9118f114dd6c156638ba36fb7a0d2fe40\": container with ID starting with e571a03b3c52184c8fa59e1ab6fff8d9118f114dd6c156638ba36fb7a0d2fe40 not found: ID does not exist" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.754801 4631 scope.go:117] "RemoveContainer" containerID="542662c37344400bc1d05230e709812104866735429d37ed0d7894050e7a54e8" Nov 28 13:29:03 crc kubenswrapper[4631]: E1128 13:29:03.755169 4631 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"542662c37344400bc1d05230e709812104866735429d37ed0d7894050e7a54e8\": container with ID starting with 542662c37344400bc1d05230e709812104866735429d37ed0d7894050e7a54e8 not found: ID does not exist" containerID="542662c37344400bc1d05230e709812104866735429d37ed0d7894050e7a54e8" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.755249 4631 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"542662c37344400bc1d05230e709812104866735429d37ed0d7894050e7a54e8"} err="failed to get container status \"542662c37344400bc1d05230e709812104866735429d37ed0d7894050e7a54e8\": rpc error: code = NotFound desc = could not find container \"542662c37344400bc1d05230e709812104866735429d37ed0d7894050e7a54e8\": container with ID starting with 542662c37344400bc1d05230e709812104866735429d37ed0d7894050e7a54e8 not found: ID does not exist" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.755371 4631 scope.go:117] "RemoveContainer" containerID="da4a7ce2b9fcb91df9dab6ba86046a998e310a5cc62a8eefcd98ea01a26f8db1" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.755786 4631 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"da4a7ce2b9fcb91df9dab6ba86046a998e310a5cc62a8eefcd98ea01a26f8db1"} err="failed to get container status \"da4a7ce2b9fcb91df9dab6ba86046a998e310a5cc62a8eefcd98ea01a26f8db1\": rpc error: code = NotFound desc = could not find container \"da4a7ce2b9fcb91df9dab6ba86046a998e310a5cc62a8eefcd98ea01a26f8db1\": container with ID starting with da4a7ce2b9fcb91df9dab6ba86046a998e310a5cc62a8eefcd98ea01a26f8db1 not found: ID does not exist" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.755863 4631 scope.go:117] "RemoveContainer" containerID="9c674448c309477eb36de0eb0312fde682c3248fb66ed1d1d4100dc114fda786" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.756223 4631 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9c674448c309477eb36de0eb0312fde682c3248fb66ed1d1d4100dc114fda786"} err="failed to get container status \"9c674448c309477eb36de0eb0312fde682c3248fb66ed1d1d4100dc114fda786\": rpc error: code = NotFound desc = could not find container \"9c674448c309477eb36de0eb0312fde682c3248fb66ed1d1d4100dc114fda786\": container with ID starting with 9c674448c309477eb36de0eb0312fde682c3248fb66ed1d1d4100dc114fda786 not found: ID does not exist" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.756261 4631 scope.go:117] "RemoveContainer" containerID="0ab2d69688bf305aeb7ece5c432d87adcb06eaca6c90ef5aa4e73b7fe477246b" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.756737 4631 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0ab2d69688bf305aeb7ece5c432d87adcb06eaca6c90ef5aa4e73b7fe477246b"} err="failed to get container status \"0ab2d69688bf305aeb7ece5c432d87adcb06eaca6c90ef5aa4e73b7fe477246b\": rpc error: code = NotFound desc = could not find container \"0ab2d69688bf305aeb7ece5c432d87adcb06eaca6c90ef5aa4e73b7fe477246b\": container with ID starting with 0ab2d69688bf305aeb7ece5c432d87adcb06eaca6c90ef5aa4e73b7fe477246b not found: ID does not exist" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.756830 4631 scope.go:117] "RemoveContainer" containerID="794ab47dc4660d76accb0456995b602cd5cdfc08616f63427947aba9e725eaf7" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.757371 4631 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"794ab47dc4660d76accb0456995b602cd5cdfc08616f63427947aba9e725eaf7"} err="failed to get container status \"794ab47dc4660d76accb0456995b602cd5cdfc08616f63427947aba9e725eaf7\": rpc error: code = NotFound desc = could not find container \"794ab47dc4660d76accb0456995b602cd5cdfc08616f63427947aba9e725eaf7\": container with ID starting with 794ab47dc4660d76accb0456995b602cd5cdfc08616f63427947aba9e725eaf7 not found: ID does not exist" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.759477 4631 scope.go:117] "RemoveContainer" containerID="016163e216d642018075fa0e9ab61fa630fa9718030246855fd44a8104c3d6e6" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.760156 4631 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"016163e216d642018075fa0e9ab61fa630fa9718030246855fd44a8104c3d6e6"} err="failed to get container status \"016163e216d642018075fa0e9ab61fa630fa9718030246855fd44a8104c3d6e6\": rpc error: code = NotFound desc = could not find container \"016163e216d642018075fa0e9ab61fa630fa9718030246855fd44a8104c3d6e6\": container with ID starting with 016163e216d642018075fa0e9ab61fa630fa9718030246855fd44a8104c3d6e6 not found: ID does not exist" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.760235 4631 scope.go:117] "RemoveContainer" containerID="e5085439ece35186f2692a8cf5be0b011fbf3f212f4f772b674c70649486c89b" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.760688 4631 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e5085439ece35186f2692a8cf5be0b011fbf3f212f4f772b674c70649486c89b"} err="failed to get container status \"e5085439ece35186f2692a8cf5be0b011fbf3f212f4f772b674c70649486c89b\": rpc error: code = NotFound desc = could not find container \"e5085439ece35186f2692a8cf5be0b011fbf3f212f4f772b674c70649486c89b\": container with ID starting with e5085439ece35186f2692a8cf5be0b011fbf3f212f4f772b674c70649486c89b not found: ID does not exist" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.760730 4631 scope.go:117] "RemoveContainer" containerID="95aeff4871c229569b0c99ff9a6cd17c63f10c1c89d9f04c533dce699c89ce31" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.762035 4631 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"95aeff4871c229569b0c99ff9a6cd17c63f10c1c89d9f04c533dce699c89ce31"} err="failed to get container status \"95aeff4871c229569b0c99ff9a6cd17c63f10c1c89d9f04c533dce699c89ce31\": rpc error: code = NotFound desc = could not find container \"95aeff4871c229569b0c99ff9a6cd17c63f10c1c89d9f04c533dce699c89ce31\": container with ID starting with 95aeff4871c229569b0c99ff9a6cd17c63f10c1c89d9f04c533dce699c89ce31 not found: ID does not exist" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.762092 4631 scope.go:117] "RemoveContainer" containerID="e571a03b3c52184c8fa59e1ab6fff8d9118f114dd6c156638ba36fb7a0d2fe40" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.762909 4631 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e571a03b3c52184c8fa59e1ab6fff8d9118f114dd6c156638ba36fb7a0d2fe40"} err="failed to get container status \"e571a03b3c52184c8fa59e1ab6fff8d9118f114dd6c156638ba36fb7a0d2fe40\": rpc error: code = NotFound desc = could not find container \"e571a03b3c52184c8fa59e1ab6fff8d9118f114dd6c156638ba36fb7a0d2fe40\": container with ID starting with e571a03b3c52184c8fa59e1ab6fff8d9118f114dd6c156638ba36fb7a0d2fe40 not found: ID does not exist" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.762939 4631 scope.go:117] "RemoveContainer" containerID="542662c37344400bc1d05230e709812104866735429d37ed0d7894050e7a54e8" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.763432 4631 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"542662c37344400bc1d05230e709812104866735429d37ed0d7894050e7a54e8"} err="failed to get container status \"542662c37344400bc1d05230e709812104866735429d37ed0d7894050e7a54e8\": rpc error: code = NotFound desc = could not find container \"542662c37344400bc1d05230e709812104866735429d37ed0d7894050e7a54e8\": container with ID starting with 542662c37344400bc1d05230e709812104866735429d37ed0d7894050e7a54e8 not found: ID does not exist" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.763491 4631 scope.go:117] "RemoveContainer" containerID="da4a7ce2b9fcb91df9dab6ba86046a998e310a5cc62a8eefcd98ea01a26f8db1" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.763907 4631 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"da4a7ce2b9fcb91df9dab6ba86046a998e310a5cc62a8eefcd98ea01a26f8db1"} err="failed to get container status \"da4a7ce2b9fcb91df9dab6ba86046a998e310a5cc62a8eefcd98ea01a26f8db1\": rpc error: code = NotFound desc = could not find container \"da4a7ce2b9fcb91df9dab6ba86046a998e310a5cc62a8eefcd98ea01a26f8db1\": container with ID starting with da4a7ce2b9fcb91df9dab6ba86046a998e310a5cc62a8eefcd98ea01a26f8db1 not found: ID does not exist" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.763937 4631 scope.go:117] "RemoveContainer" containerID="9c674448c309477eb36de0eb0312fde682c3248fb66ed1d1d4100dc114fda786" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.764346 4631 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9c674448c309477eb36de0eb0312fde682c3248fb66ed1d1d4100dc114fda786"} err="failed to get container status \"9c674448c309477eb36de0eb0312fde682c3248fb66ed1d1d4100dc114fda786\": rpc error: code = NotFound desc = could not find container \"9c674448c309477eb36de0eb0312fde682c3248fb66ed1d1d4100dc114fda786\": container with ID starting with 9c674448c309477eb36de0eb0312fde682c3248fb66ed1d1d4100dc114fda786 not found: ID does not exist" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.764402 4631 scope.go:117] "RemoveContainer" containerID="0ab2d69688bf305aeb7ece5c432d87adcb06eaca6c90ef5aa4e73b7fe477246b" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.764751 4631 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0ab2d69688bf305aeb7ece5c432d87adcb06eaca6c90ef5aa4e73b7fe477246b"} err="failed to get container status \"0ab2d69688bf305aeb7ece5c432d87adcb06eaca6c90ef5aa4e73b7fe477246b\": rpc error: code = NotFound desc = could not find container \"0ab2d69688bf305aeb7ece5c432d87adcb06eaca6c90ef5aa4e73b7fe477246b\": container with ID starting with 0ab2d69688bf305aeb7ece5c432d87adcb06eaca6c90ef5aa4e73b7fe477246b not found: ID does not exist" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.764776 4631 scope.go:117] "RemoveContainer" containerID="794ab47dc4660d76accb0456995b602cd5cdfc08616f63427947aba9e725eaf7" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.765386 4631 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"794ab47dc4660d76accb0456995b602cd5cdfc08616f63427947aba9e725eaf7"} err="failed to get container status \"794ab47dc4660d76accb0456995b602cd5cdfc08616f63427947aba9e725eaf7\": rpc error: code = NotFound desc = could not find container \"794ab47dc4660d76accb0456995b602cd5cdfc08616f63427947aba9e725eaf7\": container with ID starting with 794ab47dc4660d76accb0456995b602cd5cdfc08616f63427947aba9e725eaf7 not found: ID does not exist" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.765422 4631 scope.go:117] "RemoveContainer" containerID="016163e216d642018075fa0e9ab61fa630fa9718030246855fd44a8104c3d6e6" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.765770 4631 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"016163e216d642018075fa0e9ab61fa630fa9718030246855fd44a8104c3d6e6"} err="failed to get container status \"016163e216d642018075fa0e9ab61fa630fa9718030246855fd44a8104c3d6e6\": rpc error: code = NotFound desc = could not find container \"016163e216d642018075fa0e9ab61fa630fa9718030246855fd44a8104c3d6e6\": container with ID starting with 016163e216d642018075fa0e9ab61fa630fa9718030246855fd44a8104c3d6e6 not found: ID does not exist" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.765794 4631 scope.go:117] "RemoveContainer" containerID="e5085439ece35186f2692a8cf5be0b011fbf3f212f4f772b674c70649486c89b" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.766190 4631 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e5085439ece35186f2692a8cf5be0b011fbf3f212f4f772b674c70649486c89b"} err="failed to get container status \"e5085439ece35186f2692a8cf5be0b011fbf3f212f4f772b674c70649486c89b\": rpc error: code = NotFound desc = could not find container \"e5085439ece35186f2692a8cf5be0b011fbf3f212f4f772b674c70649486c89b\": container with ID starting with e5085439ece35186f2692a8cf5be0b011fbf3f212f4f772b674c70649486c89b not found: ID does not exist" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.766217 4631 scope.go:117] "RemoveContainer" containerID="95aeff4871c229569b0c99ff9a6cd17c63f10c1c89d9f04c533dce699c89ce31" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.766830 4631 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"95aeff4871c229569b0c99ff9a6cd17c63f10c1c89d9f04c533dce699c89ce31"} err="failed to get container status \"95aeff4871c229569b0c99ff9a6cd17c63f10c1c89d9f04c533dce699c89ce31\": rpc error: code = NotFound desc = could not find container \"95aeff4871c229569b0c99ff9a6cd17c63f10c1c89d9f04c533dce699c89ce31\": container with ID starting with 95aeff4871c229569b0c99ff9a6cd17c63f10c1c89d9f04c533dce699c89ce31 not found: ID does not exist" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.766861 4631 scope.go:117] "RemoveContainer" containerID="e571a03b3c52184c8fa59e1ab6fff8d9118f114dd6c156638ba36fb7a0d2fe40" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.768193 4631 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e571a03b3c52184c8fa59e1ab6fff8d9118f114dd6c156638ba36fb7a0d2fe40"} err="failed to get container status \"e571a03b3c52184c8fa59e1ab6fff8d9118f114dd6c156638ba36fb7a0d2fe40\": rpc error: code = NotFound desc = could not find container \"e571a03b3c52184c8fa59e1ab6fff8d9118f114dd6c156638ba36fb7a0d2fe40\": container with ID starting with e571a03b3c52184c8fa59e1ab6fff8d9118f114dd6c156638ba36fb7a0d2fe40 not found: ID does not exist" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.768223 4631 scope.go:117] "RemoveContainer" containerID="542662c37344400bc1d05230e709812104866735429d37ed0d7894050e7a54e8" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.782082 4631 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"542662c37344400bc1d05230e709812104866735429d37ed0d7894050e7a54e8"} err="failed to get container status \"542662c37344400bc1d05230e709812104866735429d37ed0d7894050e7a54e8\": rpc error: code = NotFound desc = could not find container \"542662c37344400bc1d05230e709812104866735429d37ed0d7894050e7a54e8\": container with ID starting with 542662c37344400bc1d05230e709812104866735429d37ed0d7894050e7a54e8 not found: ID does not exist" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.782177 4631 scope.go:117] "RemoveContainer" containerID="da4a7ce2b9fcb91df9dab6ba86046a998e310a5cc62a8eefcd98ea01a26f8db1" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.788492 4631 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"da4a7ce2b9fcb91df9dab6ba86046a998e310a5cc62a8eefcd98ea01a26f8db1"} err="failed to get container status \"da4a7ce2b9fcb91df9dab6ba86046a998e310a5cc62a8eefcd98ea01a26f8db1\": rpc error: code = NotFound desc = could not find container \"da4a7ce2b9fcb91df9dab6ba86046a998e310a5cc62a8eefcd98ea01a26f8db1\": container with ID starting with da4a7ce2b9fcb91df9dab6ba86046a998e310a5cc62a8eefcd98ea01a26f8db1 not found: ID does not exist" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.788569 4631 scope.go:117] "RemoveContainer" containerID="9c674448c309477eb36de0eb0312fde682c3248fb66ed1d1d4100dc114fda786" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.790045 4631 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9c674448c309477eb36de0eb0312fde682c3248fb66ed1d1d4100dc114fda786"} err="failed to get container status \"9c674448c309477eb36de0eb0312fde682c3248fb66ed1d1d4100dc114fda786\": rpc error: code = NotFound desc = could not find container \"9c674448c309477eb36de0eb0312fde682c3248fb66ed1d1d4100dc114fda786\": container with ID starting with 9c674448c309477eb36de0eb0312fde682c3248fb66ed1d1d4100dc114fda786 not found: ID does not exist" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.790090 4631 scope.go:117] "RemoveContainer" containerID="0ab2d69688bf305aeb7ece5c432d87adcb06eaca6c90ef5aa4e73b7fe477246b" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.791063 4631 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0ab2d69688bf305aeb7ece5c432d87adcb06eaca6c90ef5aa4e73b7fe477246b"} err="failed to get container status \"0ab2d69688bf305aeb7ece5c432d87adcb06eaca6c90ef5aa4e73b7fe477246b\": rpc error: code = NotFound desc = could not find container \"0ab2d69688bf305aeb7ece5c432d87adcb06eaca6c90ef5aa4e73b7fe477246b\": container with ID starting with 0ab2d69688bf305aeb7ece5c432d87adcb06eaca6c90ef5aa4e73b7fe477246b not found: ID does not exist" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.791132 4631 scope.go:117] "RemoveContainer" containerID="794ab47dc4660d76accb0456995b602cd5cdfc08616f63427947aba9e725eaf7" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.791644 4631 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"794ab47dc4660d76accb0456995b602cd5cdfc08616f63427947aba9e725eaf7"} err="failed to get container status \"794ab47dc4660d76accb0456995b602cd5cdfc08616f63427947aba9e725eaf7\": rpc error: code = NotFound desc = could not find container \"794ab47dc4660d76accb0456995b602cd5cdfc08616f63427947aba9e725eaf7\": container with ID starting with 794ab47dc4660d76accb0456995b602cd5cdfc08616f63427947aba9e725eaf7 not found: ID does not exist" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.792032 4631 scope.go:117] "RemoveContainer" containerID="016163e216d642018075fa0e9ab61fa630fa9718030246855fd44a8104c3d6e6" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.792948 4631 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"016163e216d642018075fa0e9ab61fa630fa9718030246855fd44a8104c3d6e6"} err="failed to get container status \"016163e216d642018075fa0e9ab61fa630fa9718030246855fd44a8104c3d6e6\": rpc error: code = NotFound desc = could not find container \"016163e216d642018075fa0e9ab61fa630fa9718030246855fd44a8104c3d6e6\": container with ID starting with 016163e216d642018075fa0e9ab61fa630fa9718030246855fd44a8104c3d6e6 not found: ID does not exist" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.792972 4631 scope.go:117] "RemoveContainer" containerID="e5085439ece35186f2692a8cf5be0b011fbf3f212f4f772b674c70649486c89b" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.793345 4631 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e5085439ece35186f2692a8cf5be0b011fbf3f212f4f772b674c70649486c89b"} err="failed to get container status \"e5085439ece35186f2692a8cf5be0b011fbf3f212f4f772b674c70649486c89b\": rpc error: code = NotFound desc = could not find container \"e5085439ece35186f2692a8cf5be0b011fbf3f212f4f772b674c70649486c89b\": container with ID starting with e5085439ece35186f2692a8cf5be0b011fbf3f212f4f772b674c70649486c89b not found: ID does not exist" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.793368 4631 scope.go:117] "RemoveContainer" containerID="95aeff4871c229569b0c99ff9a6cd17c63f10c1c89d9f04c533dce699c89ce31" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.793887 4631 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"95aeff4871c229569b0c99ff9a6cd17c63f10c1c89d9f04c533dce699c89ce31"} err="failed to get container status \"95aeff4871c229569b0c99ff9a6cd17c63f10c1c89d9f04c533dce699c89ce31\": rpc error: code = NotFound desc = could not find container \"95aeff4871c229569b0c99ff9a6cd17c63f10c1c89d9f04c533dce699c89ce31\": container with ID starting with 95aeff4871c229569b0c99ff9a6cd17c63f10c1c89d9f04c533dce699c89ce31 not found: ID does not exist" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.793914 4631 scope.go:117] "RemoveContainer" containerID="e571a03b3c52184c8fa59e1ab6fff8d9118f114dd6c156638ba36fb7a0d2fe40" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.794481 4631 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e571a03b3c52184c8fa59e1ab6fff8d9118f114dd6c156638ba36fb7a0d2fe40"} err="failed to get container status \"e571a03b3c52184c8fa59e1ab6fff8d9118f114dd6c156638ba36fb7a0d2fe40\": rpc error: code = NotFound desc = could not find container \"e571a03b3c52184c8fa59e1ab6fff8d9118f114dd6c156638ba36fb7a0d2fe40\": container with ID starting with e571a03b3c52184c8fa59e1ab6fff8d9118f114dd6c156638ba36fb7a0d2fe40 not found: ID does not exist" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.794519 4631 scope.go:117] "RemoveContainer" containerID="542662c37344400bc1d05230e709812104866735429d37ed0d7894050e7a54e8" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.794765 4631 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"542662c37344400bc1d05230e709812104866735429d37ed0d7894050e7a54e8"} err="failed to get container status \"542662c37344400bc1d05230e709812104866735429d37ed0d7894050e7a54e8\": rpc error: code = NotFound desc = could not find container \"542662c37344400bc1d05230e709812104866735429d37ed0d7894050e7a54e8\": container with ID starting with 542662c37344400bc1d05230e709812104866735429d37ed0d7894050e7a54e8 not found: ID does not exist" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.794791 4631 scope.go:117] "RemoveContainer" containerID="da4a7ce2b9fcb91df9dab6ba86046a998e310a5cc62a8eefcd98ea01a26f8db1" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.795311 4631 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"da4a7ce2b9fcb91df9dab6ba86046a998e310a5cc62a8eefcd98ea01a26f8db1"} err="failed to get container status \"da4a7ce2b9fcb91df9dab6ba86046a998e310a5cc62a8eefcd98ea01a26f8db1\": rpc error: code = NotFound desc = could not find container \"da4a7ce2b9fcb91df9dab6ba86046a998e310a5cc62a8eefcd98ea01a26f8db1\": container with ID starting with da4a7ce2b9fcb91df9dab6ba86046a998e310a5cc62a8eefcd98ea01a26f8db1 not found: ID does not exist" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.795342 4631 scope.go:117] "RemoveContainer" containerID="9c674448c309477eb36de0eb0312fde682c3248fb66ed1d1d4100dc114fda786" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.795740 4631 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9c674448c309477eb36de0eb0312fde682c3248fb66ed1d1d4100dc114fda786"} err="failed to get container status \"9c674448c309477eb36de0eb0312fde682c3248fb66ed1d1d4100dc114fda786\": rpc error: code = NotFound desc = could not find container \"9c674448c309477eb36de0eb0312fde682c3248fb66ed1d1d4100dc114fda786\": container with ID starting with 9c674448c309477eb36de0eb0312fde682c3248fb66ed1d1d4100dc114fda786 not found: ID does not exist" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.795765 4631 scope.go:117] "RemoveContainer" containerID="0ab2d69688bf305aeb7ece5c432d87adcb06eaca6c90ef5aa4e73b7fe477246b" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.796126 4631 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0ab2d69688bf305aeb7ece5c432d87adcb06eaca6c90ef5aa4e73b7fe477246b"} err="failed to get container status \"0ab2d69688bf305aeb7ece5c432d87adcb06eaca6c90ef5aa4e73b7fe477246b\": rpc error: code = NotFound desc = could not find container \"0ab2d69688bf305aeb7ece5c432d87adcb06eaca6c90ef5aa4e73b7fe477246b\": container with ID starting with 0ab2d69688bf305aeb7ece5c432d87adcb06eaca6c90ef5aa4e73b7fe477246b not found: ID does not exist" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.796153 4631 scope.go:117] "RemoveContainer" containerID="794ab47dc4660d76accb0456995b602cd5cdfc08616f63427947aba9e725eaf7" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.796654 4631 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"794ab47dc4660d76accb0456995b602cd5cdfc08616f63427947aba9e725eaf7"} err="failed to get container status \"794ab47dc4660d76accb0456995b602cd5cdfc08616f63427947aba9e725eaf7\": rpc error: code = NotFound desc = could not find container \"794ab47dc4660d76accb0456995b602cd5cdfc08616f63427947aba9e725eaf7\": container with ID starting with 794ab47dc4660d76accb0456995b602cd5cdfc08616f63427947aba9e725eaf7 not found: ID does not exist" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.796702 4631 scope.go:117] "RemoveContainer" containerID="016163e216d642018075fa0e9ab61fa630fa9718030246855fd44a8104c3d6e6" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.797143 4631 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"016163e216d642018075fa0e9ab61fa630fa9718030246855fd44a8104c3d6e6"} err="failed to get container status \"016163e216d642018075fa0e9ab61fa630fa9718030246855fd44a8104c3d6e6\": rpc error: code = NotFound desc = could not find container \"016163e216d642018075fa0e9ab61fa630fa9718030246855fd44a8104c3d6e6\": container with ID starting with 016163e216d642018075fa0e9ab61fa630fa9718030246855fd44a8104c3d6e6 not found: ID does not exist" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.797180 4631 scope.go:117] "RemoveContainer" containerID="e5085439ece35186f2692a8cf5be0b011fbf3f212f4f772b674c70649486c89b" Nov 28 13:29:03 crc kubenswrapper[4631]: I1128 13:29:03.797652 4631 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e5085439ece35186f2692a8cf5be0b011fbf3f212f4f772b674c70649486c89b"} err="failed to get container status \"e5085439ece35186f2692a8cf5be0b011fbf3f212f4f772b674c70649486c89b\": rpc error: code = NotFound desc = could not find container \"e5085439ece35186f2692a8cf5be0b011fbf3f212f4f772b674c70649486c89b\": container with ID starting with e5085439ece35186f2692a8cf5be0b011fbf3f212f4f772b674c70649486c89b not found: ID does not exist" Nov 28 13:29:04 crc kubenswrapper[4631]: I1128 13:29:04.551592 4631 generic.go:334] "Generic (PLEG): container finished" podID="0d38965f-6676-4e8e-9a4b-da5364a20e4e" containerID="625f8e85da34b850ca0d10b1a00280877886ae38b3340bb899178d9456dfca56" exitCode=0 Nov 28 13:29:04 crc kubenswrapper[4631]: I1128 13:29:04.552028 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-wjsdr" event={"ID":"0d38965f-6676-4e8e-9a4b-da5364a20e4e","Type":"ContainerDied","Data":"625f8e85da34b850ca0d10b1a00280877886ae38b3340bb899178d9456dfca56"} Nov 28 13:29:04 crc kubenswrapper[4631]: I1128 13:29:04.552073 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-wjsdr" event={"ID":"0d38965f-6676-4e8e-9a4b-da5364a20e4e","Type":"ContainerStarted","Data":"9694b775be14852475a382143843c7b34d5a5047d8a2f24dd09d0f11c5f23fd9"} Nov 28 13:29:04 crc kubenswrapper[4631]: I1128 13:29:04.564809 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-j8d66_9a346894-644f-4359-baa5-23bb2d0acc5f/kube-multus/0.log" Nov 28 13:29:04 crc kubenswrapper[4631]: I1128 13:29:04.564947 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-j8d66" event={"ID":"9a346894-644f-4359-baa5-23bb2d0acc5f","Type":"ContainerStarted","Data":"5d79461c03b8db15cfe7de2575dd03794043cb5e6995ddcd4561a22d756b02ea"} Nov 28 13:29:05 crc kubenswrapper[4631]: I1128 13:29:05.526606 4631 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a205f475-340d-4ec5-a860-d4a5c2ceb39e" path="/var/lib/kubelet/pods/a205f475-340d-4ec5-a860-d4a5c2ceb39e/volumes" Nov 28 13:29:05 crc kubenswrapper[4631]: I1128 13:29:05.576448 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-wjsdr" event={"ID":"0d38965f-6676-4e8e-9a4b-da5364a20e4e","Type":"ContainerStarted","Data":"68d688e7c18c3f61de9fc09fb79b703badc039f7d0c563fbe53aabc9811f5406"} Nov 28 13:29:05 crc kubenswrapper[4631]: I1128 13:29:05.577419 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-wjsdr" event={"ID":"0d38965f-6676-4e8e-9a4b-da5364a20e4e","Type":"ContainerStarted","Data":"7c3e9097d2923f3f7ddeddb6fed70e395869a8c6700f5651a716f7291cdea670"} Nov 28 13:29:05 crc kubenswrapper[4631]: I1128 13:29:05.577474 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-wjsdr" event={"ID":"0d38965f-6676-4e8e-9a4b-da5364a20e4e","Type":"ContainerStarted","Data":"702cea94e26672a580ddead129a7e9c0ea9ac275bd08369f4ea6bac5705b14c0"} Nov 28 13:29:05 crc kubenswrapper[4631]: I1128 13:29:05.577487 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-wjsdr" event={"ID":"0d38965f-6676-4e8e-9a4b-da5364a20e4e","Type":"ContainerStarted","Data":"4042d868073426bfb5a86a1259f442fc8b97ac86702842b0ff97b16a58899042"} Nov 28 13:29:05 crc kubenswrapper[4631]: I1128 13:29:05.577500 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-wjsdr" event={"ID":"0d38965f-6676-4e8e-9a4b-da5364a20e4e","Type":"ContainerStarted","Data":"f2ee85634a49f0b52b0edf310e6b46211b7d634139befa8410091f93ae5c1d74"} Nov 28 13:29:05 crc kubenswrapper[4631]: I1128 13:29:05.577512 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-wjsdr" event={"ID":"0d38965f-6676-4e8e-9a4b-da5364a20e4e","Type":"ContainerStarted","Data":"626584714f70513267e0671dd2c01e2f477e323a686904c2679b9736b52f78b7"} Nov 28 13:29:08 crc kubenswrapper[4631]: I1128 13:29:08.602728 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-wjsdr" event={"ID":"0d38965f-6676-4e8e-9a4b-da5364a20e4e","Type":"ContainerStarted","Data":"38a824d7dc9c933a84d6ee360342f3802600aecbbead980a4f9d30887ac6ac28"} Nov 28 13:29:11 crc kubenswrapper[4631]: I1128 13:29:11.634990 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-wjsdr" event={"ID":"0d38965f-6676-4e8e-9a4b-da5364a20e4e","Type":"ContainerStarted","Data":"11e5b46f8119b98a47e27a88c890b811aec42a10d96605ae913537c5aa50727b"} Nov 28 13:29:11 crc kubenswrapper[4631]: I1128 13:29:11.638458 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-wjsdr" Nov 28 13:29:11 crc kubenswrapper[4631]: I1128 13:29:11.638834 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-wjsdr" Nov 28 13:29:11 crc kubenswrapper[4631]: I1128 13:29:11.639025 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-wjsdr" Nov 28 13:29:11 crc kubenswrapper[4631]: I1128 13:29:11.683581 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-wjsdr" Nov 28 13:29:11 crc kubenswrapper[4631]: I1128 13:29:11.689995 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-wjsdr" Nov 28 13:29:11 crc kubenswrapper[4631]: I1128 13:29:11.692855 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-wjsdr" podStartSLOduration=8.692830104 podStartE2EDuration="8.692830104s" podCreationTimestamp="2025-11-28 13:29:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 13:29:11.690134546 +0000 UTC m=+508.497437920" watchObservedRunningTime="2025-11-28 13:29:11.692830104 +0000 UTC m=+508.500133458" Nov 28 13:29:33 crc kubenswrapper[4631]: I1128 13:29:33.612048 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-wjsdr" Nov 28 13:29:44 crc kubenswrapper[4631]: I1128 13:29:44.624525 4631 scope.go:117] "RemoveContainer" containerID="1c53926191bf67e68a0e0230740de428c280f79550fd964cf4bef72a592a0f57" Nov 28 13:29:45 crc kubenswrapper[4631]: I1128 13:29:45.032655 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f6xhhl"] Nov 28 13:29:45 crc kubenswrapper[4631]: I1128 13:29:45.034817 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f6xhhl" Nov 28 13:29:45 crc kubenswrapper[4631]: I1128 13:29:45.036414 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Nov 28 13:29:45 crc kubenswrapper[4631]: I1128 13:29:45.046440 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f6xhhl"] Nov 28 13:29:45 crc kubenswrapper[4631]: I1128 13:29:45.071180 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l764c\" (UniqueName: \"kubernetes.io/projected/86f665f5-dfec-46bd-91aa-b07aa5803661-kube-api-access-l764c\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f6xhhl\" (UID: \"86f665f5-dfec-46bd-91aa-b07aa5803661\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f6xhhl" Nov 28 13:29:45 crc kubenswrapper[4631]: I1128 13:29:45.071241 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/86f665f5-dfec-46bd-91aa-b07aa5803661-bundle\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f6xhhl\" (UID: \"86f665f5-dfec-46bd-91aa-b07aa5803661\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f6xhhl" Nov 28 13:29:45 crc kubenswrapper[4631]: I1128 13:29:45.071399 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/86f665f5-dfec-46bd-91aa-b07aa5803661-util\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f6xhhl\" (UID: \"86f665f5-dfec-46bd-91aa-b07aa5803661\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f6xhhl" Nov 28 13:29:45 crc kubenswrapper[4631]: I1128 13:29:45.172786 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l764c\" (UniqueName: \"kubernetes.io/projected/86f665f5-dfec-46bd-91aa-b07aa5803661-kube-api-access-l764c\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f6xhhl\" (UID: \"86f665f5-dfec-46bd-91aa-b07aa5803661\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f6xhhl" Nov 28 13:29:45 crc kubenswrapper[4631]: I1128 13:29:45.172865 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/86f665f5-dfec-46bd-91aa-b07aa5803661-bundle\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f6xhhl\" (UID: \"86f665f5-dfec-46bd-91aa-b07aa5803661\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f6xhhl" Nov 28 13:29:45 crc kubenswrapper[4631]: I1128 13:29:45.172929 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/86f665f5-dfec-46bd-91aa-b07aa5803661-util\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f6xhhl\" (UID: \"86f665f5-dfec-46bd-91aa-b07aa5803661\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f6xhhl" Nov 28 13:29:45 crc kubenswrapper[4631]: I1128 13:29:45.173694 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/86f665f5-dfec-46bd-91aa-b07aa5803661-util\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f6xhhl\" (UID: \"86f665f5-dfec-46bd-91aa-b07aa5803661\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f6xhhl" Nov 28 13:29:45 crc kubenswrapper[4631]: I1128 13:29:45.174069 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/86f665f5-dfec-46bd-91aa-b07aa5803661-bundle\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f6xhhl\" (UID: \"86f665f5-dfec-46bd-91aa-b07aa5803661\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f6xhhl" Nov 28 13:29:45 crc kubenswrapper[4631]: I1128 13:29:45.203628 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l764c\" (UniqueName: \"kubernetes.io/projected/86f665f5-dfec-46bd-91aa-b07aa5803661-kube-api-access-l764c\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f6xhhl\" (UID: \"86f665f5-dfec-46bd-91aa-b07aa5803661\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f6xhhl" Nov 28 13:29:45 crc kubenswrapper[4631]: I1128 13:29:45.350658 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f6xhhl" Nov 28 13:29:45 crc kubenswrapper[4631]: I1128 13:29:45.627556 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f6xhhl"] Nov 28 13:29:45 crc kubenswrapper[4631]: I1128 13:29:45.864234 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f6xhhl" event={"ID":"86f665f5-dfec-46bd-91aa-b07aa5803661","Type":"ContainerStarted","Data":"726b86d4e20e02b06c6c5c4923f88267ff50402a93a2550733979c3cde0399cc"} Nov 28 13:29:45 crc kubenswrapper[4631]: I1128 13:29:45.864301 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f6xhhl" event={"ID":"86f665f5-dfec-46bd-91aa-b07aa5803661","Type":"ContainerStarted","Data":"c38b08f5e1efa7e6c16f7667b90bb125cf8618d53aa1ed3a0df1153bb5f5f4d2"} Nov 28 13:29:46 crc kubenswrapper[4631]: I1128 13:29:46.873597 4631 generic.go:334] "Generic (PLEG): container finished" podID="86f665f5-dfec-46bd-91aa-b07aa5803661" containerID="726b86d4e20e02b06c6c5c4923f88267ff50402a93a2550733979c3cde0399cc" exitCode=0 Nov 28 13:29:46 crc kubenswrapper[4631]: I1128 13:29:46.873681 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f6xhhl" event={"ID":"86f665f5-dfec-46bd-91aa-b07aa5803661","Type":"ContainerDied","Data":"726b86d4e20e02b06c6c5c4923f88267ff50402a93a2550733979c3cde0399cc"} Nov 28 13:29:48 crc kubenswrapper[4631]: I1128 13:29:48.887756 4631 generic.go:334] "Generic (PLEG): container finished" podID="86f665f5-dfec-46bd-91aa-b07aa5803661" containerID="50f909bdb2c904541d206775de6577e1768bdb41b6269b3bebf4512d490daf24" exitCode=0 Nov 28 13:29:48 crc kubenswrapper[4631]: I1128 13:29:48.887823 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f6xhhl" event={"ID":"86f665f5-dfec-46bd-91aa-b07aa5803661","Type":"ContainerDied","Data":"50f909bdb2c904541d206775de6577e1768bdb41b6269b3bebf4512d490daf24"} Nov 28 13:29:49 crc kubenswrapper[4631]: I1128 13:29:49.635373 4631 patch_prober.go:28] interesting pod/machine-config-daemon-47sxw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 28 13:29:49 crc kubenswrapper[4631]: I1128 13:29:49.636044 4631 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" podUID="8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 28 13:29:49 crc kubenswrapper[4631]: I1128 13:29:49.896079 4631 generic.go:334] "Generic (PLEG): container finished" podID="86f665f5-dfec-46bd-91aa-b07aa5803661" containerID="a7efc18ec3c5f2558536a87a416d42f6c895b8f335eca63ded97a2b41f38a944" exitCode=0 Nov 28 13:29:49 crc kubenswrapper[4631]: I1128 13:29:49.896210 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f6xhhl" event={"ID":"86f665f5-dfec-46bd-91aa-b07aa5803661","Type":"ContainerDied","Data":"a7efc18ec3c5f2558536a87a416d42f6c895b8f335eca63ded97a2b41f38a944"} Nov 28 13:29:51 crc kubenswrapper[4631]: I1128 13:29:51.196062 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f6xhhl" Nov 28 13:29:51 crc kubenswrapper[4631]: I1128 13:29:51.387793 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/86f665f5-dfec-46bd-91aa-b07aa5803661-util\") pod \"86f665f5-dfec-46bd-91aa-b07aa5803661\" (UID: \"86f665f5-dfec-46bd-91aa-b07aa5803661\") " Nov 28 13:29:51 crc kubenswrapper[4631]: I1128 13:29:51.387903 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l764c\" (UniqueName: \"kubernetes.io/projected/86f665f5-dfec-46bd-91aa-b07aa5803661-kube-api-access-l764c\") pod \"86f665f5-dfec-46bd-91aa-b07aa5803661\" (UID: \"86f665f5-dfec-46bd-91aa-b07aa5803661\") " Nov 28 13:29:51 crc kubenswrapper[4631]: I1128 13:29:51.388022 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/86f665f5-dfec-46bd-91aa-b07aa5803661-bundle\") pod \"86f665f5-dfec-46bd-91aa-b07aa5803661\" (UID: \"86f665f5-dfec-46bd-91aa-b07aa5803661\") " Nov 28 13:29:51 crc kubenswrapper[4631]: I1128 13:29:51.388803 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/86f665f5-dfec-46bd-91aa-b07aa5803661-bundle" (OuterVolumeSpecName: "bundle") pod "86f665f5-dfec-46bd-91aa-b07aa5803661" (UID: "86f665f5-dfec-46bd-91aa-b07aa5803661"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 13:29:51 crc kubenswrapper[4631]: I1128 13:29:51.394626 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/86f665f5-dfec-46bd-91aa-b07aa5803661-kube-api-access-l764c" (OuterVolumeSpecName: "kube-api-access-l764c") pod "86f665f5-dfec-46bd-91aa-b07aa5803661" (UID: "86f665f5-dfec-46bd-91aa-b07aa5803661"). InnerVolumeSpecName "kube-api-access-l764c". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:29:51 crc kubenswrapper[4631]: I1128 13:29:51.489160 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l764c\" (UniqueName: \"kubernetes.io/projected/86f665f5-dfec-46bd-91aa-b07aa5803661-kube-api-access-l764c\") on node \"crc\" DevicePath \"\"" Nov 28 13:29:51 crc kubenswrapper[4631]: I1128 13:29:51.489206 4631 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/86f665f5-dfec-46bd-91aa-b07aa5803661-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 13:29:51 crc kubenswrapper[4631]: I1128 13:29:51.489131 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/86f665f5-dfec-46bd-91aa-b07aa5803661-util" (OuterVolumeSpecName: "util") pod "86f665f5-dfec-46bd-91aa-b07aa5803661" (UID: "86f665f5-dfec-46bd-91aa-b07aa5803661"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 13:29:51 crc kubenswrapper[4631]: I1128 13:29:51.590121 4631 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/86f665f5-dfec-46bd-91aa-b07aa5803661-util\") on node \"crc\" DevicePath \"\"" Nov 28 13:29:51 crc kubenswrapper[4631]: I1128 13:29:51.915107 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f6xhhl" event={"ID":"86f665f5-dfec-46bd-91aa-b07aa5803661","Type":"ContainerDied","Data":"c38b08f5e1efa7e6c16f7667b90bb125cf8618d53aa1ed3a0df1153bb5f5f4d2"} Nov 28 13:29:51 crc kubenswrapper[4631]: I1128 13:29:51.915174 4631 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c38b08f5e1efa7e6c16f7667b90bb125cf8618d53aa1ed3a0df1153bb5f5f4d2" Nov 28 13:29:51 crc kubenswrapper[4631]: I1128 13:29:51.915454 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f6xhhl" Nov 28 13:29:56 crc kubenswrapper[4631]: I1128 13:29:56.566408 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-5b5b58f5c8-8swhv"] Nov 28 13:29:56 crc kubenswrapper[4631]: E1128 13:29:56.567365 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="86f665f5-dfec-46bd-91aa-b07aa5803661" containerName="extract" Nov 28 13:29:56 crc kubenswrapper[4631]: I1128 13:29:56.567379 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="86f665f5-dfec-46bd-91aa-b07aa5803661" containerName="extract" Nov 28 13:29:56 crc kubenswrapper[4631]: E1128 13:29:56.567403 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="86f665f5-dfec-46bd-91aa-b07aa5803661" containerName="pull" Nov 28 13:29:56 crc kubenswrapper[4631]: I1128 13:29:56.567410 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="86f665f5-dfec-46bd-91aa-b07aa5803661" containerName="pull" Nov 28 13:29:56 crc kubenswrapper[4631]: E1128 13:29:56.567421 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="86f665f5-dfec-46bd-91aa-b07aa5803661" containerName="util" Nov 28 13:29:56 crc kubenswrapper[4631]: I1128 13:29:56.567427 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="86f665f5-dfec-46bd-91aa-b07aa5803661" containerName="util" Nov 28 13:29:56 crc kubenswrapper[4631]: I1128 13:29:56.567533 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="86f665f5-dfec-46bd-91aa-b07aa5803661" containerName="extract" Nov 28 13:29:56 crc kubenswrapper[4631]: I1128 13:29:56.567963 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-8swhv" Nov 28 13:29:56 crc kubenswrapper[4631]: I1128 13:29:56.573326 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Nov 28 13:29:56 crc kubenswrapper[4631]: I1128 13:29:56.606353 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-nvj8m" Nov 28 13:29:56 crc kubenswrapper[4631]: I1128 13:29:56.606385 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Nov 28 13:29:56 crc kubenswrapper[4631]: I1128 13:29:56.611124 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-5b5b58f5c8-8swhv"] Nov 28 13:29:56 crc kubenswrapper[4631]: I1128 13:29:56.709480 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-62vbt\" (UniqueName: \"kubernetes.io/projected/9124a25d-f057-4244-bd4d-a612b7a4a01b-kube-api-access-62vbt\") pod \"nmstate-operator-5b5b58f5c8-8swhv\" (UID: \"9124a25d-f057-4244-bd4d-a612b7a4a01b\") " pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-8swhv" Nov 28 13:29:56 crc kubenswrapper[4631]: I1128 13:29:56.811085 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-62vbt\" (UniqueName: \"kubernetes.io/projected/9124a25d-f057-4244-bd4d-a612b7a4a01b-kube-api-access-62vbt\") pod \"nmstate-operator-5b5b58f5c8-8swhv\" (UID: \"9124a25d-f057-4244-bd4d-a612b7a4a01b\") " pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-8swhv" Nov 28 13:29:56 crc kubenswrapper[4631]: I1128 13:29:56.843958 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-62vbt\" (UniqueName: \"kubernetes.io/projected/9124a25d-f057-4244-bd4d-a612b7a4a01b-kube-api-access-62vbt\") pod \"nmstate-operator-5b5b58f5c8-8swhv\" (UID: \"9124a25d-f057-4244-bd4d-a612b7a4a01b\") " pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-8swhv" Nov 28 13:29:56 crc kubenswrapper[4631]: I1128 13:29:56.920697 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-8swhv" Nov 28 13:29:57 crc kubenswrapper[4631]: I1128 13:29:57.157137 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-5b5b58f5c8-8swhv"] Nov 28 13:29:57 crc kubenswrapper[4631]: I1128 13:29:57.960975 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-8swhv" event={"ID":"9124a25d-f057-4244-bd4d-a612b7a4a01b","Type":"ContainerStarted","Data":"d975e8e1abea1f2f5e76a30819e35ad89465f720c46f00a1dab0711e571ee9a1"} Nov 28 13:30:00 crc kubenswrapper[4631]: I1128 13:30:00.151381 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29405610-wbcgx"] Nov 28 13:30:00 crc kubenswrapper[4631]: I1128 13:30:00.152971 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29405610-wbcgx" Nov 28 13:30:00 crc kubenswrapper[4631]: I1128 13:30:00.156420 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 28 13:30:00 crc kubenswrapper[4631]: I1128 13:30:00.156475 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 28 13:30:00 crc kubenswrapper[4631]: I1128 13:30:00.158885 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/aed94628-7bd5-4f59-add9-81a80b78a133-config-volume\") pod \"collect-profiles-29405610-wbcgx\" (UID: \"aed94628-7bd5-4f59-add9-81a80b78a133\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405610-wbcgx" Nov 28 13:30:00 crc kubenswrapper[4631]: I1128 13:30:00.159018 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/aed94628-7bd5-4f59-add9-81a80b78a133-secret-volume\") pod \"collect-profiles-29405610-wbcgx\" (UID: \"aed94628-7bd5-4f59-add9-81a80b78a133\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405610-wbcgx" Nov 28 13:30:00 crc kubenswrapper[4631]: I1128 13:30:00.159086 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-trhq5\" (UniqueName: \"kubernetes.io/projected/aed94628-7bd5-4f59-add9-81a80b78a133-kube-api-access-trhq5\") pod \"collect-profiles-29405610-wbcgx\" (UID: \"aed94628-7bd5-4f59-add9-81a80b78a133\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405610-wbcgx" Nov 28 13:30:00 crc kubenswrapper[4631]: I1128 13:30:00.164506 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29405610-wbcgx"] Nov 28 13:30:00 crc kubenswrapper[4631]: I1128 13:30:00.262662 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-trhq5\" (UniqueName: \"kubernetes.io/projected/aed94628-7bd5-4f59-add9-81a80b78a133-kube-api-access-trhq5\") pod \"collect-profiles-29405610-wbcgx\" (UID: \"aed94628-7bd5-4f59-add9-81a80b78a133\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405610-wbcgx" Nov 28 13:30:00 crc kubenswrapper[4631]: I1128 13:30:00.262901 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/aed94628-7bd5-4f59-add9-81a80b78a133-config-volume\") pod \"collect-profiles-29405610-wbcgx\" (UID: \"aed94628-7bd5-4f59-add9-81a80b78a133\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405610-wbcgx" Nov 28 13:30:00 crc kubenswrapper[4631]: I1128 13:30:00.263120 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/aed94628-7bd5-4f59-add9-81a80b78a133-secret-volume\") pod \"collect-profiles-29405610-wbcgx\" (UID: \"aed94628-7bd5-4f59-add9-81a80b78a133\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405610-wbcgx" Nov 28 13:30:00 crc kubenswrapper[4631]: I1128 13:30:00.265726 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/aed94628-7bd5-4f59-add9-81a80b78a133-config-volume\") pod \"collect-profiles-29405610-wbcgx\" (UID: \"aed94628-7bd5-4f59-add9-81a80b78a133\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405610-wbcgx" Nov 28 13:30:00 crc kubenswrapper[4631]: I1128 13:30:00.281049 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/aed94628-7bd5-4f59-add9-81a80b78a133-secret-volume\") pod \"collect-profiles-29405610-wbcgx\" (UID: \"aed94628-7bd5-4f59-add9-81a80b78a133\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405610-wbcgx" Nov 28 13:30:00 crc kubenswrapper[4631]: I1128 13:30:00.284333 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-trhq5\" (UniqueName: \"kubernetes.io/projected/aed94628-7bd5-4f59-add9-81a80b78a133-kube-api-access-trhq5\") pod \"collect-profiles-29405610-wbcgx\" (UID: \"aed94628-7bd5-4f59-add9-81a80b78a133\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405610-wbcgx" Nov 28 13:30:00 crc kubenswrapper[4631]: I1128 13:30:00.478233 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29405610-wbcgx" Nov 28 13:30:00 crc kubenswrapper[4631]: I1128 13:30:00.700659 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29405610-wbcgx"] Nov 28 13:30:00 crc kubenswrapper[4631]: I1128 13:30:00.979302 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29405610-wbcgx" event={"ID":"aed94628-7bd5-4f59-add9-81a80b78a133","Type":"ContainerStarted","Data":"8645aad4175aa8cba8b2690f5c3464b2bccdd8a0134b163b25c95b78477ae78b"} Nov 28 13:30:00 crc kubenswrapper[4631]: I1128 13:30:00.979391 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29405610-wbcgx" event={"ID":"aed94628-7bd5-4f59-add9-81a80b78a133","Type":"ContainerStarted","Data":"228a7bfb7432df31d232d889595b74a2ba72e64f183b6848c504c894d7faf124"} Nov 28 13:30:00 crc kubenswrapper[4631]: I1128 13:30:00.980763 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-8swhv" event={"ID":"9124a25d-f057-4244-bd4d-a612b7a4a01b","Type":"ContainerStarted","Data":"4257ff3e08a226744c6f55212444293b469c43e02d1a40efbc9189f258cbc4be"} Nov 28 13:30:00 crc kubenswrapper[4631]: I1128 13:30:00.997574 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29405610-wbcgx" podStartSLOduration=0.997543343 podStartE2EDuration="997.543343ms" podCreationTimestamp="2025-11-28 13:30:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 13:30:00.993969454 +0000 UTC m=+557.801272808" watchObservedRunningTime="2025-11-28 13:30:00.997543343 +0000 UTC m=+557.804846687" Nov 28 13:30:01 crc kubenswrapper[4631]: I1128 13:30:01.015013 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-8swhv" podStartSLOduration=1.7430461990000001 podStartE2EDuration="5.014988048s" podCreationTimestamp="2025-11-28 13:29:56 +0000 UTC" firstStartedPulling="2025-11-28 13:29:57.169456866 +0000 UTC m=+553.976760200" lastFinishedPulling="2025-11-28 13:30:00.441398705 +0000 UTC m=+557.248702049" observedRunningTime="2025-11-28 13:30:01.014136397 +0000 UTC m=+557.821439741" watchObservedRunningTime="2025-11-28 13:30:01.014988048 +0000 UTC m=+557.822291392" Nov 28 13:30:01 crc kubenswrapper[4631]: I1128 13:30:01.988518 4631 generic.go:334] "Generic (PLEG): container finished" podID="aed94628-7bd5-4f59-add9-81a80b78a133" containerID="8645aad4175aa8cba8b2690f5c3464b2bccdd8a0134b163b25c95b78477ae78b" exitCode=0 Nov 28 13:30:01 crc kubenswrapper[4631]: I1128 13:30:01.988625 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29405610-wbcgx" event={"ID":"aed94628-7bd5-4f59-add9-81a80b78a133","Type":"ContainerDied","Data":"8645aad4175aa8cba8b2690f5c3464b2bccdd8a0134b163b25c95b78477ae78b"} Nov 28 13:30:03 crc kubenswrapper[4631]: I1128 13:30:03.250903 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29405610-wbcgx" Nov 28 13:30:03 crc kubenswrapper[4631]: I1128 13:30:03.314058 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/aed94628-7bd5-4f59-add9-81a80b78a133-secret-volume\") pod \"aed94628-7bd5-4f59-add9-81a80b78a133\" (UID: \"aed94628-7bd5-4f59-add9-81a80b78a133\") " Nov 28 13:30:03 crc kubenswrapper[4631]: I1128 13:30:03.314214 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/aed94628-7bd5-4f59-add9-81a80b78a133-config-volume\") pod \"aed94628-7bd5-4f59-add9-81a80b78a133\" (UID: \"aed94628-7bd5-4f59-add9-81a80b78a133\") " Nov 28 13:30:03 crc kubenswrapper[4631]: I1128 13:30:03.314955 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aed94628-7bd5-4f59-add9-81a80b78a133-config-volume" (OuterVolumeSpecName: "config-volume") pod "aed94628-7bd5-4f59-add9-81a80b78a133" (UID: "aed94628-7bd5-4f59-add9-81a80b78a133"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 13:30:03 crc kubenswrapper[4631]: I1128 13:30:03.315138 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-trhq5\" (UniqueName: \"kubernetes.io/projected/aed94628-7bd5-4f59-add9-81a80b78a133-kube-api-access-trhq5\") pod \"aed94628-7bd5-4f59-add9-81a80b78a133\" (UID: \"aed94628-7bd5-4f59-add9-81a80b78a133\") " Nov 28 13:30:03 crc kubenswrapper[4631]: I1128 13:30:03.315831 4631 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/aed94628-7bd5-4f59-add9-81a80b78a133-config-volume\") on node \"crc\" DevicePath \"\"" Nov 28 13:30:03 crc kubenswrapper[4631]: I1128 13:30:03.320886 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aed94628-7bd5-4f59-add9-81a80b78a133-kube-api-access-trhq5" (OuterVolumeSpecName: "kube-api-access-trhq5") pod "aed94628-7bd5-4f59-add9-81a80b78a133" (UID: "aed94628-7bd5-4f59-add9-81a80b78a133"). InnerVolumeSpecName "kube-api-access-trhq5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:30:03 crc kubenswrapper[4631]: I1128 13:30:03.320886 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aed94628-7bd5-4f59-add9-81a80b78a133-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "aed94628-7bd5-4f59-add9-81a80b78a133" (UID: "aed94628-7bd5-4f59-add9-81a80b78a133"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:30:03 crc kubenswrapper[4631]: I1128 13:30:03.419453 4631 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/aed94628-7bd5-4f59-add9-81a80b78a133-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 28 13:30:03 crc kubenswrapper[4631]: I1128 13:30:03.419597 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-trhq5\" (UniqueName: \"kubernetes.io/projected/aed94628-7bd5-4f59-add9-81a80b78a133-kube-api-access-trhq5\") on node \"crc\" DevicePath \"\"" Nov 28 13:30:04 crc kubenswrapper[4631]: I1128 13:30:04.002417 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29405610-wbcgx" event={"ID":"aed94628-7bd5-4f59-add9-81a80b78a133","Type":"ContainerDied","Data":"228a7bfb7432df31d232d889595b74a2ba72e64f183b6848c504c894d7faf124"} Nov 28 13:30:04 crc kubenswrapper[4631]: I1128 13:30:04.002483 4631 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="228a7bfb7432df31d232d889595b74a2ba72e64f183b6848c504c894d7faf124" Nov 28 13:30:04 crc kubenswrapper[4631]: I1128 13:30:04.002889 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29405610-wbcgx" Nov 28 13:30:05 crc kubenswrapper[4631]: I1128 13:30:05.402142 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-7f946cbc9-z7fvn"] Nov 28 13:30:05 crc kubenswrapper[4631]: E1128 13:30:05.403046 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aed94628-7bd5-4f59-add9-81a80b78a133" containerName="collect-profiles" Nov 28 13:30:05 crc kubenswrapper[4631]: I1128 13:30:05.403071 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="aed94628-7bd5-4f59-add9-81a80b78a133" containerName="collect-profiles" Nov 28 13:30:05 crc kubenswrapper[4631]: I1128 13:30:05.403215 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="aed94628-7bd5-4f59-add9-81a80b78a133" containerName="collect-profiles" Nov 28 13:30:05 crc kubenswrapper[4631]: I1128 13:30:05.404110 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-z7fvn" Nov 28 13:30:05 crc kubenswrapper[4631]: I1128 13:30:05.406059 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-xq4g2" Nov 28 13:30:05 crc kubenswrapper[4631]: I1128 13:30:05.413825 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-5f6d4c5ccb-b2ssz"] Nov 28 13:30:05 crc kubenswrapper[4631]: I1128 13:30:05.414732 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-b2ssz" Nov 28 13:30:05 crc kubenswrapper[4631]: I1128 13:30:05.419797 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Nov 28 13:30:05 crc kubenswrapper[4631]: I1128 13:30:05.489735 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-9t5sn"] Nov 28 13:30:05 crc kubenswrapper[4631]: I1128 13:30:05.490644 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-9t5sn" Nov 28 13:30:05 crc kubenswrapper[4631]: I1128 13:30:05.510649 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-7f946cbc9-z7fvn"] Nov 28 13:30:05 crc kubenswrapper[4631]: I1128 13:30:05.527786 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-5f6d4c5ccb-b2ssz"] Nov 28 13:30:05 crc kubenswrapper[4631]: I1128 13:30:05.546332 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pvgqp\" (UniqueName: \"kubernetes.io/projected/f4c0815f-e4ee-4be8-855a-9bf949bf91a0-kube-api-access-pvgqp\") pod \"nmstate-webhook-5f6d4c5ccb-b2ssz\" (UID: \"f4c0815f-e4ee-4be8-855a-9bf949bf91a0\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-b2ssz" Nov 28 13:30:05 crc kubenswrapper[4631]: I1128 13:30:05.546393 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/f4c0815f-e4ee-4be8-855a-9bf949bf91a0-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-b2ssz\" (UID: \"f4c0815f-e4ee-4be8-855a-9bf949bf91a0\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-b2ssz" Nov 28 13:30:05 crc kubenswrapper[4631]: I1128 13:30:05.546437 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vhmvh\" (UniqueName: \"kubernetes.io/projected/40f6a5ae-0fc9-4af6-a001-667ce456f18a-kube-api-access-vhmvh\") pod \"nmstate-metrics-7f946cbc9-z7fvn\" (UID: \"40f6a5ae-0fc9-4af6-a001-667ce456f18a\") " pod="openshift-nmstate/nmstate-metrics-7f946cbc9-z7fvn" Nov 28 13:30:05 crc kubenswrapper[4631]: I1128 13:30:05.622019 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7fbb5f6569-k9dxt"] Nov 28 13:30:05 crc kubenswrapper[4631]: I1128 13:30:05.622975 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-k9dxt" Nov 28 13:30:05 crc kubenswrapper[4631]: I1128 13:30:05.625404 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Nov 28 13:30:05 crc kubenswrapper[4631]: I1128 13:30:05.625556 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Nov 28 13:30:05 crc kubenswrapper[4631]: I1128 13:30:05.626182 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-qdqnw" Nov 28 13:30:05 crc kubenswrapper[4631]: I1128 13:30:05.644182 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7fbb5f6569-k9dxt"] Nov 28 13:30:05 crc kubenswrapper[4631]: I1128 13:30:05.647366 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/f4c0815f-e4ee-4be8-855a-9bf949bf91a0-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-b2ssz\" (UID: \"f4c0815f-e4ee-4be8-855a-9bf949bf91a0\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-b2ssz" Nov 28 13:30:05 crc kubenswrapper[4631]: I1128 13:30:05.647448 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7k99p\" (UniqueName: \"kubernetes.io/projected/6dd14cda-0401-4840-8e08-aebcb536ffb9-kube-api-access-7k99p\") pod \"nmstate-handler-9t5sn\" (UID: \"6dd14cda-0401-4840-8e08-aebcb536ffb9\") " pod="openshift-nmstate/nmstate-handler-9t5sn" Nov 28 13:30:05 crc kubenswrapper[4631]: I1128 13:30:05.647482 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vhmvh\" (UniqueName: \"kubernetes.io/projected/40f6a5ae-0fc9-4af6-a001-667ce456f18a-kube-api-access-vhmvh\") pod \"nmstate-metrics-7f946cbc9-z7fvn\" (UID: \"40f6a5ae-0fc9-4af6-a001-667ce456f18a\") " pod="openshift-nmstate/nmstate-metrics-7f946cbc9-z7fvn" Nov 28 13:30:05 crc kubenswrapper[4631]: I1128 13:30:05.647512 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/6dd14cda-0401-4840-8e08-aebcb536ffb9-nmstate-lock\") pod \"nmstate-handler-9t5sn\" (UID: \"6dd14cda-0401-4840-8e08-aebcb536ffb9\") " pod="openshift-nmstate/nmstate-handler-9t5sn" Nov 28 13:30:05 crc kubenswrapper[4631]: I1128 13:30:05.647545 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/6dd14cda-0401-4840-8e08-aebcb536ffb9-ovs-socket\") pod \"nmstate-handler-9t5sn\" (UID: \"6dd14cda-0401-4840-8e08-aebcb536ffb9\") " pod="openshift-nmstate/nmstate-handler-9t5sn" Nov 28 13:30:05 crc kubenswrapper[4631]: I1128 13:30:05.647875 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pvgqp\" (UniqueName: \"kubernetes.io/projected/f4c0815f-e4ee-4be8-855a-9bf949bf91a0-kube-api-access-pvgqp\") pod \"nmstate-webhook-5f6d4c5ccb-b2ssz\" (UID: \"f4c0815f-e4ee-4be8-855a-9bf949bf91a0\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-b2ssz" Nov 28 13:30:05 crc kubenswrapper[4631]: I1128 13:30:05.647942 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/6dd14cda-0401-4840-8e08-aebcb536ffb9-dbus-socket\") pod \"nmstate-handler-9t5sn\" (UID: \"6dd14cda-0401-4840-8e08-aebcb536ffb9\") " pod="openshift-nmstate/nmstate-handler-9t5sn" Nov 28 13:30:05 crc kubenswrapper[4631]: I1128 13:30:05.654565 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/f4c0815f-e4ee-4be8-855a-9bf949bf91a0-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-b2ssz\" (UID: \"f4c0815f-e4ee-4be8-855a-9bf949bf91a0\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-b2ssz" Nov 28 13:30:05 crc kubenswrapper[4631]: I1128 13:30:05.692351 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pvgqp\" (UniqueName: \"kubernetes.io/projected/f4c0815f-e4ee-4be8-855a-9bf949bf91a0-kube-api-access-pvgqp\") pod \"nmstate-webhook-5f6d4c5ccb-b2ssz\" (UID: \"f4c0815f-e4ee-4be8-855a-9bf949bf91a0\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-b2ssz" Nov 28 13:30:05 crc kubenswrapper[4631]: I1128 13:30:05.693508 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vhmvh\" (UniqueName: \"kubernetes.io/projected/40f6a5ae-0fc9-4af6-a001-667ce456f18a-kube-api-access-vhmvh\") pod \"nmstate-metrics-7f946cbc9-z7fvn\" (UID: \"40f6a5ae-0fc9-4af6-a001-667ce456f18a\") " pod="openshift-nmstate/nmstate-metrics-7f946cbc9-z7fvn" Nov 28 13:30:05 crc kubenswrapper[4631]: I1128 13:30:05.722866 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-z7fvn" Nov 28 13:30:05 crc kubenswrapper[4631]: I1128 13:30:05.731446 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-b2ssz" Nov 28 13:30:05 crc kubenswrapper[4631]: I1128 13:30:05.749932 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/6dd14cda-0401-4840-8e08-aebcb536ffb9-dbus-socket\") pod \"nmstate-handler-9t5sn\" (UID: \"6dd14cda-0401-4840-8e08-aebcb536ffb9\") " pod="openshift-nmstate/nmstate-handler-9t5sn" Nov 28 13:30:05 crc kubenswrapper[4631]: I1128 13:30:05.749996 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/61c15dca-2c0d-4c3f-b63c-6a6b7be6ad99-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-k9dxt\" (UID: \"61c15dca-2c0d-4c3f-b63c-6a6b7be6ad99\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-k9dxt" Nov 28 13:30:05 crc kubenswrapper[4631]: I1128 13:30:05.750154 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7k99p\" (UniqueName: \"kubernetes.io/projected/6dd14cda-0401-4840-8e08-aebcb536ffb9-kube-api-access-7k99p\") pod \"nmstate-handler-9t5sn\" (UID: \"6dd14cda-0401-4840-8e08-aebcb536ffb9\") " pod="openshift-nmstate/nmstate-handler-9t5sn" Nov 28 13:30:05 crc kubenswrapper[4631]: I1128 13:30:05.750216 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/61c15dca-2c0d-4c3f-b63c-6a6b7be6ad99-nginx-conf\") pod \"nmstate-console-plugin-7fbb5f6569-k9dxt\" (UID: \"61c15dca-2c0d-4c3f-b63c-6a6b7be6ad99\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-k9dxt" Nov 28 13:30:05 crc kubenswrapper[4631]: I1128 13:30:05.750249 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xxmdz\" (UniqueName: \"kubernetes.io/projected/61c15dca-2c0d-4c3f-b63c-6a6b7be6ad99-kube-api-access-xxmdz\") pod \"nmstate-console-plugin-7fbb5f6569-k9dxt\" (UID: \"61c15dca-2c0d-4c3f-b63c-6a6b7be6ad99\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-k9dxt" Nov 28 13:30:05 crc kubenswrapper[4631]: I1128 13:30:05.750271 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/6dd14cda-0401-4840-8e08-aebcb536ffb9-nmstate-lock\") pod \"nmstate-handler-9t5sn\" (UID: \"6dd14cda-0401-4840-8e08-aebcb536ffb9\") " pod="openshift-nmstate/nmstate-handler-9t5sn" Nov 28 13:30:05 crc kubenswrapper[4631]: I1128 13:30:05.752568 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/6dd14cda-0401-4840-8e08-aebcb536ffb9-dbus-socket\") pod \"nmstate-handler-9t5sn\" (UID: \"6dd14cda-0401-4840-8e08-aebcb536ffb9\") " pod="openshift-nmstate/nmstate-handler-9t5sn" Nov 28 13:30:05 crc kubenswrapper[4631]: I1128 13:30:05.756493 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/6dd14cda-0401-4840-8e08-aebcb536ffb9-nmstate-lock\") pod \"nmstate-handler-9t5sn\" (UID: \"6dd14cda-0401-4840-8e08-aebcb536ffb9\") " pod="openshift-nmstate/nmstate-handler-9t5sn" Nov 28 13:30:05 crc kubenswrapper[4631]: I1128 13:30:05.756626 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/6dd14cda-0401-4840-8e08-aebcb536ffb9-ovs-socket\") pod \"nmstate-handler-9t5sn\" (UID: \"6dd14cda-0401-4840-8e08-aebcb536ffb9\") " pod="openshift-nmstate/nmstate-handler-9t5sn" Nov 28 13:30:05 crc kubenswrapper[4631]: I1128 13:30:05.756794 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/6dd14cda-0401-4840-8e08-aebcb536ffb9-ovs-socket\") pod \"nmstate-handler-9t5sn\" (UID: \"6dd14cda-0401-4840-8e08-aebcb536ffb9\") " pod="openshift-nmstate/nmstate-handler-9t5sn" Nov 28 13:30:05 crc kubenswrapper[4631]: I1128 13:30:05.784772 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7k99p\" (UniqueName: \"kubernetes.io/projected/6dd14cda-0401-4840-8e08-aebcb536ffb9-kube-api-access-7k99p\") pod \"nmstate-handler-9t5sn\" (UID: \"6dd14cda-0401-4840-8e08-aebcb536ffb9\") " pod="openshift-nmstate/nmstate-handler-9t5sn" Nov 28 13:30:05 crc kubenswrapper[4631]: I1128 13:30:05.807262 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-9t5sn" Nov 28 13:30:05 crc kubenswrapper[4631]: I1128 13:30:05.859439 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/61c15dca-2c0d-4c3f-b63c-6a6b7be6ad99-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-k9dxt\" (UID: \"61c15dca-2c0d-4c3f-b63c-6a6b7be6ad99\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-k9dxt" Nov 28 13:30:05 crc kubenswrapper[4631]: I1128 13:30:05.859718 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/61c15dca-2c0d-4c3f-b63c-6a6b7be6ad99-nginx-conf\") pod \"nmstate-console-plugin-7fbb5f6569-k9dxt\" (UID: \"61c15dca-2c0d-4c3f-b63c-6a6b7be6ad99\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-k9dxt" Nov 28 13:30:05 crc kubenswrapper[4631]: I1128 13:30:05.859834 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xxmdz\" (UniqueName: \"kubernetes.io/projected/61c15dca-2c0d-4c3f-b63c-6a6b7be6ad99-kube-api-access-xxmdz\") pod \"nmstate-console-plugin-7fbb5f6569-k9dxt\" (UID: \"61c15dca-2c0d-4c3f-b63c-6a6b7be6ad99\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-k9dxt" Nov 28 13:30:05 crc kubenswrapper[4631]: E1128 13:30:05.859747 4631 secret.go:188] Couldn't get secret openshift-nmstate/plugin-serving-cert: secret "plugin-serving-cert" not found Nov 28 13:30:05 crc kubenswrapper[4631]: E1128 13:30:05.860015 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/61c15dca-2c0d-4c3f-b63c-6a6b7be6ad99-plugin-serving-cert podName:61c15dca-2c0d-4c3f-b63c-6a6b7be6ad99 nodeName:}" failed. No retries permitted until 2025-11-28 13:30:06.359981683 +0000 UTC m=+563.167285027 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "plugin-serving-cert" (UniqueName: "kubernetes.io/secret/61c15dca-2c0d-4c3f-b63c-6a6b7be6ad99-plugin-serving-cert") pod "nmstate-console-plugin-7fbb5f6569-k9dxt" (UID: "61c15dca-2c0d-4c3f-b63c-6a6b7be6ad99") : secret "plugin-serving-cert" not found Nov 28 13:30:05 crc kubenswrapper[4631]: I1128 13:30:05.860865 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/61c15dca-2c0d-4c3f-b63c-6a6b7be6ad99-nginx-conf\") pod \"nmstate-console-plugin-7fbb5f6569-k9dxt\" (UID: \"61c15dca-2c0d-4c3f-b63c-6a6b7be6ad99\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-k9dxt" Nov 28 13:30:05 crc kubenswrapper[4631]: I1128 13:30:05.869130 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-7998b944c5-s5ftv"] Nov 28 13:30:05 crc kubenswrapper[4631]: I1128 13:30:05.870136 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-7998b944c5-s5ftv" Nov 28 13:30:05 crc kubenswrapper[4631]: I1128 13:30:05.942486 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xxmdz\" (UniqueName: \"kubernetes.io/projected/61c15dca-2c0d-4c3f-b63c-6a6b7be6ad99-kube-api-access-xxmdz\") pod \"nmstate-console-plugin-7fbb5f6569-k9dxt\" (UID: \"61c15dca-2c0d-4c3f-b63c-6a6b7be6ad99\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-k9dxt" Nov 28 13:30:05 crc kubenswrapper[4631]: I1128 13:30:05.943683 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-7998b944c5-s5ftv"] Nov 28 13:30:05 crc kubenswrapper[4631]: I1128 13:30:05.969875 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/8c20e6bf-10ce-41a7-a27a-2854f55cf40f-console-oauth-config\") pod \"console-7998b944c5-s5ftv\" (UID: \"8c20e6bf-10ce-41a7-a27a-2854f55cf40f\") " pod="openshift-console/console-7998b944c5-s5ftv" Nov 28 13:30:05 crc kubenswrapper[4631]: I1128 13:30:05.969957 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9m5b8\" (UniqueName: \"kubernetes.io/projected/8c20e6bf-10ce-41a7-a27a-2854f55cf40f-kube-api-access-9m5b8\") pod \"console-7998b944c5-s5ftv\" (UID: \"8c20e6bf-10ce-41a7-a27a-2854f55cf40f\") " pod="openshift-console/console-7998b944c5-s5ftv" Nov 28 13:30:05 crc kubenswrapper[4631]: I1128 13:30:05.970013 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/8c20e6bf-10ce-41a7-a27a-2854f55cf40f-service-ca\") pod \"console-7998b944c5-s5ftv\" (UID: \"8c20e6bf-10ce-41a7-a27a-2854f55cf40f\") " pod="openshift-console/console-7998b944c5-s5ftv" Nov 28 13:30:05 crc kubenswrapper[4631]: I1128 13:30:05.970038 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/8c20e6bf-10ce-41a7-a27a-2854f55cf40f-oauth-serving-cert\") pod \"console-7998b944c5-s5ftv\" (UID: \"8c20e6bf-10ce-41a7-a27a-2854f55cf40f\") " pod="openshift-console/console-7998b944c5-s5ftv" Nov 28 13:30:05 crc kubenswrapper[4631]: I1128 13:30:05.970054 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/8c20e6bf-10ce-41a7-a27a-2854f55cf40f-console-config\") pod \"console-7998b944c5-s5ftv\" (UID: \"8c20e6bf-10ce-41a7-a27a-2854f55cf40f\") " pod="openshift-console/console-7998b944c5-s5ftv" Nov 28 13:30:05 crc kubenswrapper[4631]: I1128 13:30:05.970147 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8c20e6bf-10ce-41a7-a27a-2854f55cf40f-trusted-ca-bundle\") pod \"console-7998b944c5-s5ftv\" (UID: \"8c20e6bf-10ce-41a7-a27a-2854f55cf40f\") " pod="openshift-console/console-7998b944c5-s5ftv" Nov 28 13:30:05 crc kubenswrapper[4631]: I1128 13:30:05.970196 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/8c20e6bf-10ce-41a7-a27a-2854f55cf40f-console-serving-cert\") pod \"console-7998b944c5-s5ftv\" (UID: \"8c20e6bf-10ce-41a7-a27a-2854f55cf40f\") " pod="openshift-console/console-7998b944c5-s5ftv" Nov 28 13:30:06 crc kubenswrapper[4631]: I1128 13:30:06.023778 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-9t5sn" event={"ID":"6dd14cda-0401-4840-8e08-aebcb536ffb9","Type":"ContainerStarted","Data":"0968df6e49555cf352b0c1683b64d0f6e5147390a45e1441cffd182477fb0529"} Nov 28 13:30:06 crc kubenswrapper[4631]: I1128 13:30:06.071461 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8c20e6bf-10ce-41a7-a27a-2854f55cf40f-trusted-ca-bundle\") pod \"console-7998b944c5-s5ftv\" (UID: \"8c20e6bf-10ce-41a7-a27a-2854f55cf40f\") " pod="openshift-console/console-7998b944c5-s5ftv" Nov 28 13:30:06 crc kubenswrapper[4631]: I1128 13:30:06.071532 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/8c20e6bf-10ce-41a7-a27a-2854f55cf40f-console-serving-cert\") pod \"console-7998b944c5-s5ftv\" (UID: \"8c20e6bf-10ce-41a7-a27a-2854f55cf40f\") " pod="openshift-console/console-7998b944c5-s5ftv" Nov 28 13:30:06 crc kubenswrapper[4631]: I1128 13:30:06.071577 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/8c20e6bf-10ce-41a7-a27a-2854f55cf40f-console-oauth-config\") pod \"console-7998b944c5-s5ftv\" (UID: \"8c20e6bf-10ce-41a7-a27a-2854f55cf40f\") " pod="openshift-console/console-7998b944c5-s5ftv" Nov 28 13:30:06 crc kubenswrapper[4631]: I1128 13:30:06.071598 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9m5b8\" (UniqueName: \"kubernetes.io/projected/8c20e6bf-10ce-41a7-a27a-2854f55cf40f-kube-api-access-9m5b8\") pod \"console-7998b944c5-s5ftv\" (UID: \"8c20e6bf-10ce-41a7-a27a-2854f55cf40f\") " pod="openshift-console/console-7998b944c5-s5ftv" Nov 28 13:30:06 crc kubenswrapper[4631]: I1128 13:30:06.071627 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/8c20e6bf-10ce-41a7-a27a-2854f55cf40f-service-ca\") pod \"console-7998b944c5-s5ftv\" (UID: \"8c20e6bf-10ce-41a7-a27a-2854f55cf40f\") " pod="openshift-console/console-7998b944c5-s5ftv" Nov 28 13:30:06 crc kubenswrapper[4631]: I1128 13:30:06.071648 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/8c20e6bf-10ce-41a7-a27a-2854f55cf40f-oauth-serving-cert\") pod \"console-7998b944c5-s5ftv\" (UID: \"8c20e6bf-10ce-41a7-a27a-2854f55cf40f\") " pod="openshift-console/console-7998b944c5-s5ftv" Nov 28 13:30:06 crc kubenswrapper[4631]: I1128 13:30:06.071666 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/8c20e6bf-10ce-41a7-a27a-2854f55cf40f-console-config\") pod \"console-7998b944c5-s5ftv\" (UID: \"8c20e6bf-10ce-41a7-a27a-2854f55cf40f\") " pod="openshift-console/console-7998b944c5-s5ftv" Nov 28 13:30:06 crc kubenswrapper[4631]: I1128 13:30:06.072561 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8c20e6bf-10ce-41a7-a27a-2854f55cf40f-trusted-ca-bundle\") pod \"console-7998b944c5-s5ftv\" (UID: \"8c20e6bf-10ce-41a7-a27a-2854f55cf40f\") " pod="openshift-console/console-7998b944c5-s5ftv" Nov 28 13:30:06 crc kubenswrapper[4631]: I1128 13:30:06.072972 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/8c20e6bf-10ce-41a7-a27a-2854f55cf40f-service-ca\") pod \"console-7998b944c5-s5ftv\" (UID: \"8c20e6bf-10ce-41a7-a27a-2854f55cf40f\") " pod="openshift-console/console-7998b944c5-s5ftv" Nov 28 13:30:06 crc kubenswrapper[4631]: I1128 13:30:06.073114 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/8c20e6bf-10ce-41a7-a27a-2854f55cf40f-oauth-serving-cert\") pod \"console-7998b944c5-s5ftv\" (UID: \"8c20e6bf-10ce-41a7-a27a-2854f55cf40f\") " pod="openshift-console/console-7998b944c5-s5ftv" Nov 28 13:30:06 crc kubenswrapper[4631]: I1128 13:30:06.073674 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/8c20e6bf-10ce-41a7-a27a-2854f55cf40f-console-config\") pod \"console-7998b944c5-s5ftv\" (UID: \"8c20e6bf-10ce-41a7-a27a-2854f55cf40f\") " pod="openshift-console/console-7998b944c5-s5ftv" Nov 28 13:30:06 crc kubenswrapper[4631]: I1128 13:30:06.078858 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/8c20e6bf-10ce-41a7-a27a-2854f55cf40f-console-oauth-config\") pod \"console-7998b944c5-s5ftv\" (UID: \"8c20e6bf-10ce-41a7-a27a-2854f55cf40f\") " pod="openshift-console/console-7998b944c5-s5ftv" Nov 28 13:30:06 crc kubenswrapper[4631]: I1128 13:30:06.079343 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/8c20e6bf-10ce-41a7-a27a-2854f55cf40f-console-serving-cert\") pod \"console-7998b944c5-s5ftv\" (UID: \"8c20e6bf-10ce-41a7-a27a-2854f55cf40f\") " pod="openshift-console/console-7998b944c5-s5ftv" Nov 28 13:30:06 crc kubenswrapper[4631]: I1128 13:30:06.091130 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9m5b8\" (UniqueName: \"kubernetes.io/projected/8c20e6bf-10ce-41a7-a27a-2854f55cf40f-kube-api-access-9m5b8\") pod \"console-7998b944c5-s5ftv\" (UID: \"8c20e6bf-10ce-41a7-a27a-2854f55cf40f\") " pod="openshift-console/console-7998b944c5-s5ftv" Nov 28 13:30:06 crc kubenswrapper[4631]: I1128 13:30:06.095557 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-5f6d4c5ccb-b2ssz"] Nov 28 13:30:06 crc kubenswrapper[4631]: I1128 13:30:06.164123 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-7f946cbc9-z7fvn"] Nov 28 13:30:06 crc kubenswrapper[4631]: W1128 13:30:06.165982 4631 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod40f6a5ae_0fc9_4af6_a001_667ce456f18a.slice/crio-627f10124b53d65b4dd18cd4dbc7073698864d5b0ac5849305895553c644f339 WatchSource:0}: Error finding container 627f10124b53d65b4dd18cd4dbc7073698864d5b0ac5849305895553c644f339: Status 404 returned error can't find the container with id 627f10124b53d65b4dd18cd4dbc7073698864d5b0ac5849305895553c644f339 Nov 28 13:30:06 crc kubenswrapper[4631]: I1128 13:30:06.254570 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-7998b944c5-s5ftv" Nov 28 13:30:06 crc kubenswrapper[4631]: I1128 13:30:06.381302 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/61c15dca-2c0d-4c3f-b63c-6a6b7be6ad99-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-k9dxt\" (UID: \"61c15dca-2c0d-4c3f-b63c-6a6b7be6ad99\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-k9dxt" Nov 28 13:30:06 crc kubenswrapper[4631]: I1128 13:30:06.385300 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/61c15dca-2c0d-4c3f-b63c-6a6b7be6ad99-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-k9dxt\" (UID: \"61c15dca-2c0d-4c3f-b63c-6a6b7be6ad99\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-k9dxt" Nov 28 13:30:06 crc kubenswrapper[4631]: I1128 13:30:06.540120 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-k9dxt" Nov 28 13:30:06 crc kubenswrapper[4631]: I1128 13:30:06.687654 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-7998b944c5-s5ftv"] Nov 28 13:30:06 crc kubenswrapper[4631]: W1128 13:30:06.740325 4631 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8c20e6bf_10ce_41a7_a27a_2854f55cf40f.slice/crio-7f5bf411f4c9b36a863ae4d75a8a1b64a27aff384c61a1e5af74ae57f047e5de WatchSource:0}: Error finding container 7f5bf411f4c9b36a863ae4d75a8a1b64a27aff384c61a1e5af74ae57f047e5de: Status 404 returned error can't find the container with id 7f5bf411f4c9b36a863ae4d75a8a1b64a27aff384c61a1e5af74ae57f047e5de Nov 28 13:30:06 crc kubenswrapper[4631]: I1128 13:30:06.800975 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7fbb5f6569-k9dxt"] Nov 28 13:30:06 crc kubenswrapper[4631]: W1128 13:30:06.809460 4631 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod61c15dca_2c0d_4c3f_b63c_6a6b7be6ad99.slice/crio-abeb1ffe7e6f9a8cce4ecb85f386e3b04630171ced0712b8610a3607e450de22 WatchSource:0}: Error finding container abeb1ffe7e6f9a8cce4ecb85f386e3b04630171ced0712b8610a3607e450de22: Status 404 returned error can't find the container with id abeb1ffe7e6f9a8cce4ecb85f386e3b04630171ced0712b8610a3607e450de22 Nov 28 13:30:07 crc kubenswrapper[4631]: I1128 13:30:07.032471 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-z7fvn" event={"ID":"40f6a5ae-0fc9-4af6-a001-667ce456f18a","Type":"ContainerStarted","Data":"627f10124b53d65b4dd18cd4dbc7073698864d5b0ac5849305895553c644f339"} Nov 28 13:30:07 crc kubenswrapper[4631]: I1128 13:30:07.033860 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-k9dxt" event={"ID":"61c15dca-2c0d-4c3f-b63c-6a6b7be6ad99","Type":"ContainerStarted","Data":"abeb1ffe7e6f9a8cce4ecb85f386e3b04630171ced0712b8610a3607e450de22"} Nov 28 13:30:07 crc kubenswrapper[4631]: I1128 13:30:07.035123 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-b2ssz" event={"ID":"f4c0815f-e4ee-4be8-855a-9bf949bf91a0","Type":"ContainerStarted","Data":"72354b82cea682bc8c692df2ab246717a45563e5d79edc60adf24291672c52ad"} Nov 28 13:30:07 crc kubenswrapper[4631]: I1128 13:30:07.036914 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-7998b944c5-s5ftv" event={"ID":"8c20e6bf-10ce-41a7-a27a-2854f55cf40f","Type":"ContainerStarted","Data":"120fadd10a264d5ed2c39f150b8589cc2fb6926f804f96d64a4f724c5fcc2b1c"} Nov 28 13:30:07 crc kubenswrapper[4631]: I1128 13:30:07.036996 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-7998b944c5-s5ftv" event={"ID":"8c20e6bf-10ce-41a7-a27a-2854f55cf40f","Type":"ContainerStarted","Data":"7f5bf411f4c9b36a863ae4d75a8a1b64a27aff384c61a1e5af74ae57f047e5de"} Nov 28 13:30:07 crc kubenswrapper[4631]: I1128 13:30:07.081777 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-7998b944c5-s5ftv" podStartSLOduration=2.081739239 podStartE2EDuration="2.081739239s" podCreationTimestamp="2025-11-28 13:30:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 13:30:07.075731699 +0000 UTC m=+563.883035043" watchObservedRunningTime="2025-11-28 13:30:07.081739239 +0000 UTC m=+563.889042583" Nov 28 13:30:09 crc kubenswrapper[4631]: I1128 13:30:09.058393 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-z7fvn" event={"ID":"40f6a5ae-0fc9-4af6-a001-667ce456f18a","Type":"ContainerStarted","Data":"9173fe2365f149a39c2565e4822773d78fd5fbd20b9b3f5ae07fdf2f1f9f5f5d"} Nov 28 13:30:10 crc kubenswrapper[4631]: I1128 13:30:10.069816 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-b2ssz" event={"ID":"f4c0815f-e4ee-4be8-855a-9bf949bf91a0","Type":"ContainerStarted","Data":"eb7feb33c290325b6f0f3ac6505cb67bfae84a8a72d59ce10fbdf4c4dd72d03e"} Nov 28 13:30:10 crc kubenswrapper[4631]: I1128 13:30:10.070403 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-b2ssz" Nov 28 13:30:10 crc kubenswrapper[4631]: I1128 13:30:10.074930 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-9t5sn" event={"ID":"6dd14cda-0401-4840-8e08-aebcb536ffb9","Type":"ContainerStarted","Data":"47411c731b1db3ccb6445edfa2ae76be501a07adbb832a5dee0371e26552218d"} Nov 28 13:30:10 crc kubenswrapper[4631]: I1128 13:30:10.075114 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-9t5sn" Nov 28 13:30:10 crc kubenswrapper[4631]: I1128 13:30:10.108106 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-b2ssz" podStartSLOduration=2.37976761 podStartE2EDuration="5.108083074s" podCreationTimestamp="2025-11-28 13:30:05 +0000 UTC" firstStartedPulling="2025-11-28 13:30:06.112660163 +0000 UTC m=+562.919963497" lastFinishedPulling="2025-11-28 13:30:08.840975617 +0000 UTC m=+565.648278961" observedRunningTime="2025-11-28 13:30:10.091460619 +0000 UTC m=+566.898763963" watchObservedRunningTime="2025-11-28 13:30:10.108083074 +0000 UTC m=+566.915386418" Nov 28 13:30:10 crc kubenswrapper[4631]: I1128 13:30:10.116044 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-9t5sn" podStartSLOduration=2.109945543 podStartE2EDuration="5.116016942s" podCreationTimestamp="2025-11-28 13:30:05 +0000 UTC" firstStartedPulling="2025-11-28 13:30:05.851557253 +0000 UTC m=+562.658860597" lastFinishedPulling="2025-11-28 13:30:08.857628642 +0000 UTC m=+565.664931996" observedRunningTime="2025-11-28 13:30:10.107461348 +0000 UTC m=+566.914764692" watchObservedRunningTime="2025-11-28 13:30:10.116016942 +0000 UTC m=+566.923320336" Nov 28 13:30:11 crc kubenswrapper[4631]: I1128 13:30:11.086250 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-k9dxt" event={"ID":"61c15dca-2c0d-4c3f-b63c-6a6b7be6ad99","Type":"ContainerStarted","Data":"49c6b3c538c8e339b63833293ad8bb73086ebd7e07dab199ed520fca39a34995"} Nov 28 13:30:11 crc kubenswrapper[4631]: I1128 13:30:11.118045 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-k9dxt" podStartSLOduration=2.798136693 podStartE2EDuration="6.118013637s" podCreationTimestamp="2025-11-28 13:30:05 +0000 UTC" firstStartedPulling="2025-11-28 13:30:06.812227588 +0000 UTC m=+563.619530932" lastFinishedPulling="2025-11-28 13:30:10.132104532 +0000 UTC m=+566.939407876" observedRunningTime="2025-11-28 13:30:11.112373886 +0000 UTC m=+567.919677230" watchObservedRunningTime="2025-11-28 13:30:11.118013637 +0000 UTC m=+567.925316981" Nov 28 13:30:12 crc kubenswrapper[4631]: I1128 13:30:12.096916 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-z7fvn" event={"ID":"40f6a5ae-0fc9-4af6-a001-667ce456f18a","Type":"ContainerStarted","Data":"fe8e10233802bc4a0e4910eeb7e4030ab7de2d9f9850f11a0e41d37cce11f2b4"} Nov 28 13:30:12 crc kubenswrapper[4631]: I1128 13:30:12.118591 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-z7fvn" podStartSLOduration=1.6501533369999999 podStartE2EDuration="7.118565677s" podCreationTimestamp="2025-11-28 13:30:05 +0000 UTC" firstStartedPulling="2025-11-28 13:30:06.169614064 +0000 UTC m=+562.976917408" lastFinishedPulling="2025-11-28 13:30:11.638026404 +0000 UTC m=+568.445329748" observedRunningTime="2025-11-28 13:30:12.116733311 +0000 UTC m=+568.924036675" watchObservedRunningTime="2025-11-28 13:30:12.118565677 +0000 UTC m=+568.925869061" Nov 28 13:30:15 crc kubenswrapper[4631]: I1128 13:30:15.831881 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-9t5sn" Nov 28 13:30:16 crc kubenswrapper[4631]: I1128 13:30:16.255942 4631 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-7998b944c5-s5ftv" Nov 28 13:30:16 crc kubenswrapper[4631]: I1128 13:30:16.256040 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-7998b944c5-s5ftv" Nov 28 13:30:16 crc kubenswrapper[4631]: I1128 13:30:16.265739 4631 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-7998b944c5-s5ftv" Nov 28 13:30:17 crc kubenswrapper[4631]: I1128 13:30:17.133667 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-7998b944c5-s5ftv" Nov 28 13:30:17 crc kubenswrapper[4631]: I1128 13:30:17.207138 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-659jg"] Nov 28 13:30:19 crc kubenswrapper[4631]: I1128 13:30:19.634714 4631 patch_prober.go:28] interesting pod/machine-config-daemon-47sxw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 28 13:30:19 crc kubenswrapper[4631]: I1128 13:30:19.635140 4631 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" podUID="8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 28 13:30:25 crc kubenswrapper[4631]: I1128 13:30:25.740415 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-b2ssz" Nov 28 13:30:41 crc kubenswrapper[4631]: I1128 13:30:41.411543 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83rfhlr"] Nov 28 13:30:41 crc kubenswrapper[4631]: I1128 13:30:41.413656 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83rfhlr" Nov 28 13:30:41 crc kubenswrapper[4631]: I1128 13:30:41.415832 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Nov 28 13:30:41 crc kubenswrapper[4631]: I1128 13:30:41.422271 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83rfhlr"] Nov 28 13:30:41 crc kubenswrapper[4631]: I1128 13:30:41.444022 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ddb177f8-0125-45bc-bbc4-130d689e10c9-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83rfhlr\" (UID: \"ddb177f8-0125-45bc-bbc4-130d689e10c9\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83rfhlr" Nov 28 13:30:41 crc kubenswrapper[4631]: I1128 13:30:41.444070 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ddb177f8-0125-45bc-bbc4-130d689e10c9-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83rfhlr\" (UID: \"ddb177f8-0125-45bc-bbc4-130d689e10c9\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83rfhlr" Nov 28 13:30:41 crc kubenswrapper[4631]: I1128 13:30:41.444108 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hcps9\" (UniqueName: \"kubernetes.io/projected/ddb177f8-0125-45bc-bbc4-130d689e10c9-kube-api-access-hcps9\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83rfhlr\" (UID: \"ddb177f8-0125-45bc-bbc4-130d689e10c9\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83rfhlr" Nov 28 13:30:41 crc kubenswrapper[4631]: I1128 13:30:41.545193 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ddb177f8-0125-45bc-bbc4-130d689e10c9-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83rfhlr\" (UID: \"ddb177f8-0125-45bc-bbc4-130d689e10c9\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83rfhlr" Nov 28 13:30:41 crc kubenswrapper[4631]: I1128 13:30:41.545335 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hcps9\" (UniqueName: \"kubernetes.io/projected/ddb177f8-0125-45bc-bbc4-130d689e10c9-kube-api-access-hcps9\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83rfhlr\" (UID: \"ddb177f8-0125-45bc-bbc4-130d689e10c9\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83rfhlr" Nov 28 13:30:41 crc kubenswrapper[4631]: I1128 13:30:41.545483 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ddb177f8-0125-45bc-bbc4-130d689e10c9-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83rfhlr\" (UID: \"ddb177f8-0125-45bc-bbc4-130d689e10c9\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83rfhlr" Nov 28 13:30:41 crc kubenswrapper[4631]: I1128 13:30:41.545991 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ddb177f8-0125-45bc-bbc4-130d689e10c9-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83rfhlr\" (UID: \"ddb177f8-0125-45bc-bbc4-130d689e10c9\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83rfhlr" Nov 28 13:30:41 crc kubenswrapper[4631]: I1128 13:30:41.546079 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ddb177f8-0125-45bc-bbc4-130d689e10c9-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83rfhlr\" (UID: \"ddb177f8-0125-45bc-bbc4-130d689e10c9\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83rfhlr" Nov 28 13:30:41 crc kubenswrapper[4631]: I1128 13:30:41.567082 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hcps9\" (UniqueName: \"kubernetes.io/projected/ddb177f8-0125-45bc-bbc4-130d689e10c9-kube-api-access-hcps9\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83rfhlr\" (UID: \"ddb177f8-0125-45bc-bbc4-130d689e10c9\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83rfhlr" Nov 28 13:30:41 crc kubenswrapper[4631]: I1128 13:30:41.731961 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83rfhlr" Nov 28 13:30:42 crc kubenswrapper[4631]: I1128 13:30:42.018022 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83rfhlr"] Nov 28 13:30:42 crc kubenswrapper[4631]: I1128 13:30:42.255486 4631 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-659jg" podUID="8e852fdc-5b82-4b51-b0e5-804a1b3be0ec" containerName="console" containerID="cri-o://6d1ef439fe58f9af39a531e0eaf632db22a7ed41276b202ee16433099c70ecc3" gracePeriod=15 Nov 28 13:30:42 crc kubenswrapper[4631]: I1128 13:30:42.316781 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83rfhlr" event={"ID":"ddb177f8-0125-45bc-bbc4-130d689e10c9","Type":"ContainerStarted","Data":"5854322c5d50e05ee933646263681ba756d95a486e97fe1226b7c687d1fcdf69"} Nov 28 13:30:42 crc kubenswrapper[4631]: I1128 13:30:42.317335 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83rfhlr" event={"ID":"ddb177f8-0125-45bc-bbc4-130d689e10c9","Type":"ContainerStarted","Data":"5d758f5bb81955305c08292ba8e424494aaf719ae58a50ff17d885ed7616d47f"} Nov 28 13:30:42 crc kubenswrapper[4631]: I1128 13:30:42.561236 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-659jg_8e852fdc-5b82-4b51-b0e5-804a1b3be0ec/console/0.log" Nov 28 13:30:42 crc kubenswrapper[4631]: I1128 13:30:42.561361 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-659jg" Nov 28 13:30:42 crc kubenswrapper[4631]: I1128 13:30:42.658796 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/8e852fdc-5b82-4b51-b0e5-804a1b3be0ec-console-oauth-config\") pod \"8e852fdc-5b82-4b51-b0e5-804a1b3be0ec\" (UID: \"8e852fdc-5b82-4b51-b0e5-804a1b3be0ec\") " Nov 28 13:30:42 crc kubenswrapper[4631]: I1128 13:30:42.658915 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/8e852fdc-5b82-4b51-b0e5-804a1b3be0ec-console-config\") pod \"8e852fdc-5b82-4b51-b0e5-804a1b3be0ec\" (UID: \"8e852fdc-5b82-4b51-b0e5-804a1b3be0ec\") " Nov 28 13:30:42 crc kubenswrapper[4631]: I1128 13:30:42.658961 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wrx26\" (UniqueName: \"kubernetes.io/projected/8e852fdc-5b82-4b51-b0e5-804a1b3be0ec-kube-api-access-wrx26\") pod \"8e852fdc-5b82-4b51-b0e5-804a1b3be0ec\" (UID: \"8e852fdc-5b82-4b51-b0e5-804a1b3be0ec\") " Nov 28 13:30:42 crc kubenswrapper[4631]: I1128 13:30:42.658980 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/8e852fdc-5b82-4b51-b0e5-804a1b3be0ec-console-serving-cert\") pod \"8e852fdc-5b82-4b51-b0e5-804a1b3be0ec\" (UID: \"8e852fdc-5b82-4b51-b0e5-804a1b3be0ec\") " Nov 28 13:30:42 crc kubenswrapper[4631]: I1128 13:30:42.659010 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8e852fdc-5b82-4b51-b0e5-804a1b3be0ec-trusted-ca-bundle\") pod \"8e852fdc-5b82-4b51-b0e5-804a1b3be0ec\" (UID: \"8e852fdc-5b82-4b51-b0e5-804a1b3be0ec\") " Nov 28 13:30:42 crc kubenswrapper[4631]: I1128 13:30:42.659040 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/8e852fdc-5b82-4b51-b0e5-804a1b3be0ec-service-ca\") pod \"8e852fdc-5b82-4b51-b0e5-804a1b3be0ec\" (UID: \"8e852fdc-5b82-4b51-b0e5-804a1b3be0ec\") " Nov 28 13:30:42 crc kubenswrapper[4631]: I1128 13:30:42.659057 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/8e852fdc-5b82-4b51-b0e5-804a1b3be0ec-oauth-serving-cert\") pod \"8e852fdc-5b82-4b51-b0e5-804a1b3be0ec\" (UID: \"8e852fdc-5b82-4b51-b0e5-804a1b3be0ec\") " Nov 28 13:30:42 crc kubenswrapper[4631]: I1128 13:30:42.660066 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8e852fdc-5b82-4b51-b0e5-804a1b3be0ec-console-config" (OuterVolumeSpecName: "console-config") pod "8e852fdc-5b82-4b51-b0e5-804a1b3be0ec" (UID: "8e852fdc-5b82-4b51-b0e5-804a1b3be0ec"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 13:30:42 crc kubenswrapper[4631]: I1128 13:30:42.660204 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8e852fdc-5b82-4b51-b0e5-804a1b3be0ec-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "8e852fdc-5b82-4b51-b0e5-804a1b3be0ec" (UID: "8e852fdc-5b82-4b51-b0e5-804a1b3be0ec"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 13:30:42 crc kubenswrapper[4631]: I1128 13:30:42.660231 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8e852fdc-5b82-4b51-b0e5-804a1b3be0ec-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "8e852fdc-5b82-4b51-b0e5-804a1b3be0ec" (UID: "8e852fdc-5b82-4b51-b0e5-804a1b3be0ec"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 13:30:42 crc kubenswrapper[4631]: I1128 13:30:42.660246 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8e852fdc-5b82-4b51-b0e5-804a1b3be0ec-service-ca" (OuterVolumeSpecName: "service-ca") pod "8e852fdc-5b82-4b51-b0e5-804a1b3be0ec" (UID: "8e852fdc-5b82-4b51-b0e5-804a1b3be0ec"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 13:30:42 crc kubenswrapper[4631]: I1128 13:30:42.665084 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8e852fdc-5b82-4b51-b0e5-804a1b3be0ec-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "8e852fdc-5b82-4b51-b0e5-804a1b3be0ec" (UID: "8e852fdc-5b82-4b51-b0e5-804a1b3be0ec"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:30:42 crc kubenswrapper[4631]: I1128 13:30:42.665143 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8e852fdc-5b82-4b51-b0e5-804a1b3be0ec-kube-api-access-wrx26" (OuterVolumeSpecName: "kube-api-access-wrx26") pod "8e852fdc-5b82-4b51-b0e5-804a1b3be0ec" (UID: "8e852fdc-5b82-4b51-b0e5-804a1b3be0ec"). InnerVolumeSpecName "kube-api-access-wrx26". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:30:42 crc kubenswrapper[4631]: I1128 13:30:42.666442 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8e852fdc-5b82-4b51-b0e5-804a1b3be0ec-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "8e852fdc-5b82-4b51-b0e5-804a1b3be0ec" (UID: "8e852fdc-5b82-4b51-b0e5-804a1b3be0ec"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:30:42 crc kubenswrapper[4631]: I1128 13:30:42.759750 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wrx26\" (UniqueName: \"kubernetes.io/projected/8e852fdc-5b82-4b51-b0e5-804a1b3be0ec-kube-api-access-wrx26\") on node \"crc\" DevicePath \"\"" Nov 28 13:30:42 crc kubenswrapper[4631]: I1128 13:30:42.759787 4631 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/8e852fdc-5b82-4b51-b0e5-804a1b3be0ec-console-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 28 13:30:42 crc kubenswrapper[4631]: I1128 13:30:42.759800 4631 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8e852fdc-5b82-4b51-b0e5-804a1b3be0ec-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 13:30:42 crc kubenswrapper[4631]: I1128 13:30:42.759810 4631 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/8e852fdc-5b82-4b51-b0e5-804a1b3be0ec-service-ca\") on node \"crc\" DevicePath \"\"" Nov 28 13:30:42 crc kubenswrapper[4631]: I1128 13:30:42.759818 4631 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/8e852fdc-5b82-4b51-b0e5-804a1b3be0ec-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 28 13:30:42 crc kubenswrapper[4631]: I1128 13:30:42.759828 4631 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/8e852fdc-5b82-4b51-b0e5-804a1b3be0ec-console-oauth-config\") on node \"crc\" DevicePath \"\"" Nov 28 13:30:42 crc kubenswrapper[4631]: I1128 13:30:42.759839 4631 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/8e852fdc-5b82-4b51-b0e5-804a1b3be0ec-console-config\") on node \"crc\" DevicePath \"\"" Nov 28 13:30:43 crc kubenswrapper[4631]: I1128 13:30:43.325828 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-659jg_8e852fdc-5b82-4b51-b0e5-804a1b3be0ec/console/0.log" Nov 28 13:30:43 crc kubenswrapper[4631]: I1128 13:30:43.325918 4631 generic.go:334] "Generic (PLEG): container finished" podID="8e852fdc-5b82-4b51-b0e5-804a1b3be0ec" containerID="6d1ef439fe58f9af39a531e0eaf632db22a7ed41276b202ee16433099c70ecc3" exitCode=2 Nov 28 13:30:43 crc kubenswrapper[4631]: I1128 13:30:43.326205 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-659jg" Nov 28 13:30:43 crc kubenswrapper[4631]: I1128 13:30:43.326277 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-659jg" event={"ID":"8e852fdc-5b82-4b51-b0e5-804a1b3be0ec","Type":"ContainerDied","Data":"6d1ef439fe58f9af39a531e0eaf632db22a7ed41276b202ee16433099c70ecc3"} Nov 28 13:30:43 crc kubenswrapper[4631]: I1128 13:30:43.326444 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-659jg" event={"ID":"8e852fdc-5b82-4b51-b0e5-804a1b3be0ec","Type":"ContainerDied","Data":"df150edf9aded98b366ef2ec156a46838176f0602de4074397fc7790e5401dcc"} Nov 28 13:30:43 crc kubenswrapper[4631]: I1128 13:30:43.326494 4631 scope.go:117] "RemoveContainer" containerID="6d1ef439fe58f9af39a531e0eaf632db22a7ed41276b202ee16433099c70ecc3" Nov 28 13:30:43 crc kubenswrapper[4631]: I1128 13:30:43.328494 4631 generic.go:334] "Generic (PLEG): container finished" podID="ddb177f8-0125-45bc-bbc4-130d689e10c9" containerID="5854322c5d50e05ee933646263681ba756d95a486e97fe1226b7c687d1fcdf69" exitCode=0 Nov 28 13:30:43 crc kubenswrapper[4631]: I1128 13:30:43.328552 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83rfhlr" event={"ID":"ddb177f8-0125-45bc-bbc4-130d689e10c9","Type":"ContainerDied","Data":"5854322c5d50e05ee933646263681ba756d95a486e97fe1226b7c687d1fcdf69"} Nov 28 13:30:43 crc kubenswrapper[4631]: I1128 13:30:43.361175 4631 scope.go:117] "RemoveContainer" containerID="6d1ef439fe58f9af39a531e0eaf632db22a7ed41276b202ee16433099c70ecc3" Nov 28 13:30:43 crc kubenswrapper[4631]: E1128 13:30:43.361853 4631 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6d1ef439fe58f9af39a531e0eaf632db22a7ed41276b202ee16433099c70ecc3\": container with ID starting with 6d1ef439fe58f9af39a531e0eaf632db22a7ed41276b202ee16433099c70ecc3 not found: ID does not exist" containerID="6d1ef439fe58f9af39a531e0eaf632db22a7ed41276b202ee16433099c70ecc3" Nov 28 13:30:43 crc kubenswrapper[4631]: I1128 13:30:43.361902 4631 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6d1ef439fe58f9af39a531e0eaf632db22a7ed41276b202ee16433099c70ecc3"} err="failed to get container status \"6d1ef439fe58f9af39a531e0eaf632db22a7ed41276b202ee16433099c70ecc3\": rpc error: code = NotFound desc = could not find container \"6d1ef439fe58f9af39a531e0eaf632db22a7ed41276b202ee16433099c70ecc3\": container with ID starting with 6d1ef439fe58f9af39a531e0eaf632db22a7ed41276b202ee16433099c70ecc3 not found: ID does not exist" Nov 28 13:30:43 crc kubenswrapper[4631]: I1128 13:30:43.384502 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-659jg"] Nov 28 13:30:43 crc kubenswrapper[4631]: I1128 13:30:43.388201 4631 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-659jg"] Nov 28 13:30:43 crc kubenswrapper[4631]: I1128 13:30:43.523525 4631 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8e852fdc-5b82-4b51-b0e5-804a1b3be0ec" path="/var/lib/kubelet/pods/8e852fdc-5b82-4b51-b0e5-804a1b3be0ec/volumes" Nov 28 13:30:46 crc kubenswrapper[4631]: I1128 13:30:46.356360 4631 generic.go:334] "Generic (PLEG): container finished" podID="ddb177f8-0125-45bc-bbc4-130d689e10c9" containerID="e2c5be10a6fdeaa25caffb452a5cbed66fa4aef660eb6522c8b55ac0fb33cb71" exitCode=0 Nov 28 13:30:46 crc kubenswrapper[4631]: I1128 13:30:46.356438 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83rfhlr" event={"ID":"ddb177f8-0125-45bc-bbc4-130d689e10c9","Type":"ContainerDied","Data":"e2c5be10a6fdeaa25caffb452a5cbed66fa4aef660eb6522c8b55ac0fb33cb71"} Nov 28 13:30:47 crc kubenswrapper[4631]: I1128 13:30:47.370458 4631 generic.go:334] "Generic (PLEG): container finished" podID="ddb177f8-0125-45bc-bbc4-130d689e10c9" containerID="ab59f1ff2a2c795b62566ceca1b46c239e4f763ec049c7f1ca354b04b661dcc6" exitCode=0 Nov 28 13:30:47 crc kubenswrapper[4631]: I1128 13:30:47.370523 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83rfhlr" event={"ID":"ddb177f8-0125-45bc-bbc4-130d689e10c9","Type":"ContainerDied","Data":"ab59f1ff2a2c795b62566ceca1b46c239e4f763ec049c7f1ca354b04b661dcc6"} Nov 28 13:30:48 crc kubenswrapper[4631]: I1128 13:30:48.630232 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83rfhlr" Nov 28 13:30:48 crc kubenswrapper[4631]: I1128 13:30:48.752632 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hcps9\" (UniqueName: \"kubernetes.io/projected/ddb177f8-0125-45bc-bbc4-130d689e10c9-kube-api-access-hcps9\") pod \"ddb177f8-0125-45bc-bbc4-130d689e10c9\" (UID: \"ddb177f8-0125-45bc-bbc4-130d689e10c9\") " Nov 28 13:30:48 crc kubenswrapper[4631]: I1128 13:30:48.752759 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ddb177f8-0125-45bc-bbc4-130d689e10c9-bundle\") pod \"ddb177f8-0125-45bc-bbc4-130d689e10c9\" (UID: \"ddb177f8-0125-45bc-bbc4-130d689e10c9\") " Nov 28 13:30:48 crc kubenswrapper[4631]: I1128 13:30:48.752808 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ddb177f8-0125-45bc-bbc4-130d689e10c9-util\") pod \"ddb177f8-0125-45bc-bbc4-130d689e10c9\" (UID: \"ddb177f8-0125-45bc-bbc4-130d689e10c9\") " Nov 28 13:30:48 crc kubenswrapper[4631]: I1128 13:30:48.754430 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ddb177f8-0125-45bc-bbc4-130d689e10c9-bundle" (OuterVolumeSpecName: "bundle") pod "ddb177f8-0125-45bc-bbc4-130d689e10c9" (UID: "ddb177f8-0125-45bc-bbc4-130d689e10c9"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 13:30:48 crc kubenswrapper[4631]: I1128 13:30:48.761541 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ddb177f8-0125-45bc-bbc4-130d689e10c9-kube-api-access-hcps9" (OuterVolumeSpecName: "kube-api-access-hcps9") pod "ddb177f8-0125-45bc-bbc4-130d689e10c9" (UID: "ddb177f8-0125-45bc-bbc4-130d689e10c9"). InnerVolumeSpecName "kube-api-access-hcps9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:30:48 crc kubenswrapper[4631]: I1128 13:30:48.762842 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ddb177f8-0125-45bc-bbc4-130d689e10c9-util" (OuterVolumeSpecName: "util") pod "ddb177f8-0125-45bc-bbc4-130d689e10c9" (UID: "ddb177f8-0125-45bc-bbc4-130d689e10c9"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 13:30:48 crc kubenswrapper[4631]: I1128 13:30:48.854526 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hcps9\" (UniqueName: \"kubernetes.io/projected/ddb177f8-0125-45bc-bbc4-130d689e10c9-kube-api-access-hcps9\") on node \"crc\" DevicePath \"\"" Nov 28 13:30:48 crc kubenswrapper[4631]: I1128 13:30:48.854566 4631 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ddb177f8-0125-45bc-bbc4-130d689e10c9-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 13:30:48 crc kubenswrapper[4631]: I1128 13:30:48.854596 4631 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ddb177f8-0125-45bc-bbc4-130d689e10c9-util\") on node \"crc\" DevicePath \"\"" Nov 28 13:30:49 crc kubenswrapper[4631]: I1128 13:30:49.388565 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83rfhlr" event={"ID":"ddb177f8-0125-45bc-bbc4-130d689e10c9","Type":"ContainerDied","Data":"5d758f5bb81955305c08292ba8e424494aaf719ae58a50ff17d885ed7616d47f"} Nov 28 13:30:49 crc kubenswrapper[4631]: I1128 13:30:49.388639 4631 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5d758f5bb81955305c08292ba8e424494aaf719ae58a50ff17d885ed7616d47f" Nov 28 13:30:49 crc kubenswrapper[4631]: I1128 13:30:49.388706 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83rfhlr" Nov 28 13:30:49 crc kubenswrapper[4631]: I1128 13:30:49.634401 4631 patch_prober.go:28] interesting pod/machine-config-daemon-47sxw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 28 13:30:49 crc kubenswrapper[4631]: I1128 13:30:49.634470 4631 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" podUID="8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 28 13:30:49 crc kubenswrapper[4631]: I1128 13:30:49.634525 4631 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" Nov 28 13:30:49 crc kubenswrapper[4631]: I1128 13:30:49.635474 4631 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"42d567fa8bd6c073cd89c5129f91adc5f5feefc7933a184e745232605d6012b2"} pod="openshift-machine-config-operator/machine-config-daemon-47sxw" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 28 13:30:49 crc kubenswrapper[4631]: I1128 13:30:49.635540 4631 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" podUID="8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1" containerName="machine-config-daemon" containerID="cri-o://42d567fa8bd6c073cd89c5129f91adc5f5feefc7933a184e745232605d6012b2" gracePeriod=600 Nov 28 13:30:50 crc kubenswrapper[4631]: I1128 13:30:50.402530 4631 generic.go:334] "Generic (PLEG): container finished" podID="8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1" containerID="42d567fa8bd6c073cd89c5129f91adc5f5feefc7933a184e745232605d6012b2" exitCode=0 Nov 28 13:30:50 crc kubenswrapper[4631]: I1128 13:30:50.402583 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" event={"ID":"8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1","Type":"ContainerDied","Data":"42d567fa8bd6c073cd89c5129f91adc5f5feefc7933a184e745232605d6012b2"} Nov 28 13:30:50 crc kubenswrapper[4631]: I1128 13:30:50.402672 4631 scope.go:117] "RemoveContainer" containerID="b04ff78ff7998dc8957c0acf1246dd6adabf1e8af15924c3953943c9bdb3150a" Nov 28 13:30:51 crc kubenswrapper[4631]: I1128 13:30:51.411127 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" event={"ID":"8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1","Type":"ContainerStarted","Data":"914e9854c9ef96234991a24021b15fad0f64f1894afc1a19ce66cd4bc2c2af7d"} Nov 28 13:31:03 crc kubenswrapper[4631]: I1128 13:31:03.761945 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-7c98947894-pw5mq"] Nov 28 13:31:03 crc kubenswrapper[4631]: E1128 13:31:03.762978 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8e852fdc-5b82-4b51-b0e5-804a1b3be0ec" containerName="console" Nov 28 13:31:03 crc kubenswrapper[4631]: I1128 13:31:03.762993 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e852fdc-5b82-4b51-b0e5-804a1b3be0ec" containerName="console" Nov 28 13:31:03 crc kubenswrapper[4631]: E1128 13:31:03.763004 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ddb177f8-0125-45bc-bbc4-130d689e10c9" containerName="util" Nov 28 13:31:03 crc kubenswrapper[4631]: I1128 13:31:03.763012 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="ddb177f8-0125-45bc-bbc4-130d689e10c9" containerName="util" Nov 28 13:31:03 crc kubenswrapper[4631]: E1128 13:31:03.763032 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ddb177f8-0125-45bc-bbc4-130d689e10c9" containerName="extract" Nov 28 13:31:03 crc kubenswrapper[4631]: I1128 13:31:03.763040 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="ddb177f8-0125-45bc-bbc4-130d689e10c9" containerName="extract" Nov 28 13:31:03 crc kubenswrapper[4631]: E1128 13:31:03.763063 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ddb177f8-0125-45bc-bbc4-130d689e10c9" containerName="pull" Nov 28 13:31:03 crc kubenswrapper[4631]: I1128 13:31:03.763070 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="ddb177f8-0125-45bc-bbc4-130d689e10c9" containerName="pull" Nov 28 13:31:03 crc kubenswrapper[4631]: I1128 13:31:03.763350 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="8e852fdc-5b82-4b51-b0e5-804a1b3be0ec" containerName="console" Nov 28 13:31:03 crc kubenswrapper[4631]: I1128 13:31:03.763369 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="ddb177f8-0125-45bc-bbc4-130d689e10c9" containerName="extract" Nov 28 13:31:03 crc kubenswrapper[4631]: I1128 13:31:03.764034 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-7c98947894-pw5mq" Nov 28 13:31:03 crc kubenswrapper[4631]: I1128 13:31:03.769195 4631 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Nov 28 13:31:03 crc kubenswrapper[4631]: I1128 13:31:03.770498 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Nov 28 13:31:03 crc kubenswrapper[4631]: I1128 13:31:03.770701 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Nov 28 13:31:03 crc kubenswrapper[4631]: I1128 13:31:03.770968 4631 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Nov 28 13:31:03 crc kubenswrapper[4631]: I1128 13:31:03.771069 4631 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-k56pk" Nov 28 13:31:03 crc kubenswrapper[4631]: I1128 13:31:03.843115 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-7c98947894-pw5mq"] Nov 28 13:31:03 crc kubenswrapper[4631]: I1128 13:31:03.930049 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vkscp\" (UniqueName: \"kubernetes.io/projected/299f5184-243e-4fe0-b60e-2d40eee5a3f6-kube-api-access-vkscp\") pod \"metallb-operator-controller-manager-7c98947894-pw5mq\" (UID: \"299f5184-243e-4fe0-b60e-2d40eee5a3f6\") " pod="metallb-system/metallb-operator-controller-manager-7c98947894-pw5mq" Nov 28 13:31:03 crc kubenswrapper[4631]: I1128 13:31:03.930321 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/299f5184-243e-4fe0-b60e-2d40eee5a3f6-apiservice-cert\") pod \"metallb-operator-controller-manager-7c98947894-pw5mq\" (UID: \"299f5184-243e-4fe0-b60e-2d40eee5a3f6\") " pod="metallb-system/metallb-operator-controller-manager-7c98947894-pw5mq" Nov 28 13:31:03 crc kubenswrapper[4631]: I1128 13:31:03.930453 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/299f5184-243e-4fe0-b60e-2d40eee5a3f6-webhook-cert\") pod \"metallb-operator-controller-manager-7c98947894-pw5mq\" (UID: \"299f5184-243e-4fe0-b60e-2d40eee5a3f6\") " pod="metallb-system/metallb-operator-controller-manager-7c98947894-pw5mq" Nov 28 13:31:04 crc kubenswrapper[4631]: I1128 13:31:04.022601 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-fc587766c-d92lf"] Nov 28 13:31:04 crc kubenswrapper[4631]: I1128 13:31:04.023562 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-fc587766c-d92lf" Nov 28 13:31:04 crc kubenswrapper[4631]: I1128 13:31:04.029180 4631 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-47zj6" Nov 28 13:31:04 crc kubenswrapper[4631]: I1128 13:31:04.029368 4631 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Nov 28 13:31:04 crc kubenswrapper[4631]: I1128 13:31:04.029208 4631 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Nov 28 13:31:04 crc kubenswrapper[4631]: I1128 13:31:04.031545 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vkscp\" (UniqueName: \"kubernetes.io/projected/299f5184-243e-4fe0-b60e-2d40eee5a3f6-kube-api-access-vkscp\") pod \"metallb-operator-controller-manager-7c98947894-pw5mq\" (UID: \"299f5184-243e-4fe0-b60e-2d40eee5a3f6\") " pod="metallb-system/metallb-operator-controller-manager-7c98947894-pw5mq" Nov 28 13:31:04 crc kubenswrapper[4631]: I1128 13:31:04.031665 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/299f5184-243e-4fe0-b60e-2d40eee5a3f6-apiservice-cert\") pod \"metallb-operator-controller-manager-7c98947894-pw5mq\" (UID: \"299f5184-243e-4fe0-b60e-2d40eee5a3f6\") " pod="metallb-system/metallb-operator-controller-manager-7c98947894-pw5mq" Nov 28 13:31:04 crc kubenswrapper[4631]: I1128 13:31:04.031754 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/299f5184-243e-4fe0-b60e-2d40eee5a3f6-webhook-cert\") pod \"metallb-operator-controller-manager-7c98947894-pw5mq\" (UID: \"299f5184-243e-4fe0-b60e-2d40eee5a3f6\") " pod="metallb-system/metallb-operator-controller-manager-7c98947894-pw5mq" Nov 28 13:31:04 crc kubenswrapper[4631]: I1128 13:31:04.040463 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/299f5184-243e-4fe0-b60e-2d40eee5a3f6-webhook-cert\") pod \"metallb-operator-controller-manager-7c98947894-pw5mq\" (UID: \"299f5184-243e-4fe0-b60e-2d40eee5a3f6\") " pod="metallb-system/metallb-operator-controller-manager-7c98947894-pw5mq" Nov 28 13:31:04 crc kubenswrapper[4631]: I1128 13:31:04.049831 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-fc587766c-d92lf"] Nov 28 13:31:04 crc kubenswrapper[4631]: I1128 13:31:04.053164 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/299f5184-243e-4fe0-b60e-2d40eee5a3f6-apiservice-cert\") pod \"metallb-operator-controller-manager-7c98947894-pw5mq\" (UID: \"299f5184-243e-4fe0-b60e-2d40eee5a3f6\") " pod="metallb-system/metallb-operator-controller-manager-7c98947894-pw5mq" Nov 28 13:31:04 crc kubenswrapper[4631]: I1128 13:31:04.062451 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vkscp\" (UniqueName: \"kubernetes.io/projected/299f5184-243e-4fe0-b60e-2d40eee5a3f6-kube-api-access-vkscp\") pod \"metallb-operator-controller-manager-7c98947894-pw5mq\" (UID: \"299f5184-243e-4fe0-b60e-2d40eee5a3f6\") " pod="metallb-system/metallb-operator-controller-manager-7c98947894-pw5mq" Nov 28 13:31:04 crc kubenswrapper[4631]: I1128 13:31:04.092349 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-7c98947894-pw5mq" Nov 28 13:31:04 crc kubenswrapper[4631]: I1128 13:31:04.135202 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ktmkd\" (UniqueName: \"kubernetes.io/projected/dc71ed9c-9a56-449a-afab-1a28a746fb2e-kube-api-access-ktmkd\") pod \"metallb-operator-webhook-server-fc587766c-d92lf\" (UID: \"dc71ed9c-9a56-449a-afab-1a28a746fb2e\") " pod="metallb-system/metallb-operator-webhook-server-fc587766c-d92lf" Nov 28 13:31:04 crc kubenswrapper[4631]: I1128 13:31:04.135312 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/dc71ed9c-9a56-449a-afab-1a28a746fb2e-apiservice-cert\") pod \"metallb-operator-webhook-server-fc587766c-d92lf\" (UID: \"dc71ed9c-9a56-449a-afab-1a28a746fb2e\") " pod="metallb-system/metallb-operator-webhook-server-fc587766c-d92lf" Nov 28 13:31:04 crc kubenswrapper[4631]: I1128 13:31:04.135350 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/dc71ed9c-9a56-449a-afab-1a28a746fb2e-webhook-cert\") pod \"metallb-operator-webhook-server-fc587766c-d92lf\" (UID: \"dc71ed9c-9a56-449a-afab-1a28a746fb2e\") " pod="metallb-system/metallb-operator-webhook-server-fc587766c-d92lf" Nov 28 13:31:04 crc kubenswrapper[4631]: I1128 13:31:04.242167 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/dc71ed9c-9a56-449a-afab-1a28a746fb2e-webhook-cert\") pod \"metallb-operator-webhook-server-fc587766c-d92lf\" (UID: \"dc71ed9c-9a56-449a-afab-1a28a746fb2e\") " pod="metallb-system/metallb-operator-webhook-server-fc587766c-d92lf" Nov 28 13:31:04 crc kubenswrapper[4631]: I1128 13:31:04.242554 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ktmkd\" (UniqueName: \"kubernetes.io/projected/dc71ed9c-9a56-449a-afab-1a28a746fb2e-kube-api-access-ktmkd\") pod \"metallb-operator-webhook-server-fc587766c-d92lf\" (UID: \"dc71ed9c-9a56-449a-afab-1a28a746fb2e\") " pod="metallb-system/metallb-operator-webhook-server-fc587766c-d92lf" Nov 28 13:31:04 crc kubenswrapper[4631]: I1128 13:31:04.242614 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/dc71ed9c-9a56-449a-afab-1a28a746fb2e-apiservice-cert\") pod \"metallb-operator-webhook-server-fc587766c-d92lf\" (UID: \"dc71ed9c-9a56-449a-afab-1a28a746fb2e\") " pod="metallb-system/metallb-operator-webhook-server-fc587766c-d92lf" Nov 28 13:31:04 crc kubenswrapper[4631]: I1128 13:31:04.249098 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/dc71ed9c-9a56-449a-afab-1a28a746fb2e-webhook-cert\") pod \"metallb-operator-webhook-server-fc587766c-d92lf\" (UID: \"dc71ed9c-9a56-449a-afab-1a28a746fb2e\") " pod="metallb-system/metallb-operator-webhook-server-fc587766c-d92lf" Nov 28 13:31:04 crc kubenswrapper[4631]: I1128 13:31:04.250486 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/dc71ed9c-9a56-449a-afab-1a28a746fb2e-apiservice-cert\") pod \"metallb-operator-webhook-server-fc587766c-d92lf\" (UID: \"dc71ed9c-9a56-449a-afab-1a28a746fb2e\") " pod="metallb-system/metallb-operator-webhook-server-fc587766c-d92lf" Nov 28 13:31:04 crc kubenswrapper[4631]: I1128 13:31:04.265943 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ktmkd\" (UniqueName: \"kubernetes.io/projected/dc71ed9c-9a56-449a-afab-1a28a746fb2e-kube-api-access-ktmkd\") pod \"metallb-operator-webhook-server-fc587766c-d92lf\" (UID: \"dc71ed9c-9a56-449a-afab-1a28a746fb2e\") " pod="metallb-system/metallb-operator-webhook-server-fc587766c-d92lf" Nov 28 13:31:04 crc kubenswrapper[4631]: I1128 13:31:04.340571 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-fc587766c-d92lf" Nov 28 13:31:04 crc kubenswrapper[4631]: I1128 13:31:04.619409 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-7c98947894-pw5mq"] Nov 28 13:31:04 crc kubenswrapper[4631]: W1128 13:31:04.621855 4631 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod299f5184_243e_4fe0_b60e_2d40eee5a3f6.slice/crio-2c0b743de0d68f63fba5355115876d3737c293178079c51112b2a93c6f9802ed WatchSource:0}: Error finding container 2c0b743de0d68f63fba5355115876d3737c293178079c51112b2a93c6f9802ed: Status 404 returned error can't find the container with id 2c0b743de0d68f63fba5355115876d3737c293178079c51112b2a93c6f9802ed Nov 28 13:31:04 crc kubenswrapper[4631]: I1128 13:31:04.859006 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-fc587766c-d92lf"] Nov 28 13:31:04 crc kubenswrapper[4631]: W1128 13:31:04.863337 4631 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddc71ed9c_9a56_449a_afab_1a28a746fb2e.slice/crio-e47fee1b00b686cfa36cba13a53a8ad4768d5ab6381d8b1ed9c53eb22dd98520 WatchSource:0}: Error finding container e47fee1b00b686cfa36cba13a53a8ad4768d5ab6381d8b1ed9c53eb22dd98520: Status 404 returned error can't find the container with id e47fee1b00b686cfa36cba13a53a8ad4768d5ab6381d8b1ed9c53eb22dd98520 Nov 28 13:31:05 crc kubenswrapper[4631]: I1128 13:31:05.502937 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-7c98947894-pw5mq" event={"ID":"299f5184-243e-4fe0-b60e-2d40eee5a3f6","Type":"ContainerStarted","Data":"2c0b743de0d68f63fba5355115876d3737c293178079c51112b2a93c6f9802ed"} Nov 28 13:31:05 crc kubenswrapper[4631]: I1128 13:31:05.504610 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-fc587766c-d92lf" event={"ID":"dc71ed9c-9a56-449a-afab-1a28a746fb2e","Type":"ContainerStarted","Data":"e47fee1b00b686cfa36cba13a53a8ad4768d5ab6381d8b1ed9c53eb22dd98520"} Nov 28 13:31:13 crc kubenswrapper[4631]: I1128 13:31:13.577964 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-fc587766c-d92lf" event={"ID":"dc71ed9c-9a56-449a-afab-1a28a746fb2e","Type":"ContainerStarted","Data":"fcb1672151f41026a06ea2c6210c3439fea2b90f30dc4b4ef74319702f867fff"} Nov 28 13:31:13 crc kubenswrapper[4631]: I1128 13:31:13.578809 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-fc587766c-d92lf" Nov 28 13:31:13 crc kubenswrapper[4631]: I1128 13:31:13.584680 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-7c98947894-pw5mq" event={"ID":"299f5184-243e-4fe0-b60e-2d40eee5a3f6","Type":"ContainerStarted","Data":"3c0c7be5036e7d4ccc57ff46a787c48480a715d0e3cdc5f27acc18fe204ef3eb"} Nov 28 13:31:13 crc kubenswrapper[4631]: I1128 13:31:13.585199 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-7c98947894-pw5mq" Nov 28 13:31:13 crc kubenswrapper[4631]: I1128 13:31:13.613711 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-fc587766c-d92lf" podStartSLOduration=1.647298079 podStartE2EDuration="9.613688699s" podCreationTimestamp="2025-11-28 13:31:04 +0000 UTC" firstStartedPulling="2025-11-28 13:31:04.867401666 +0000 UTC m=+621.674705010" lastFinishedPulling="2025-11-28 13:31:12.833792276 +0000 UTC m=+629.641095630" observedRunningTime="2025-11-28 13:31:13.609226128 +0000 UTC m=+630.416529472" watchObservedRunningTime="2025-11-28 13:31:13.613688699 +0000 UTC m=+630.420992043" Nov 28 13:31:13 crc kubenswrapper[4631]: I1128 13:31:13.645594 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-7c98947894-pw5mq" podStartSLOduration=2.456584111 podStartE2EDuration="10.645570855s" podCreationTimestamp="2025-11-28 13:31:03 +0000 UTC" firstStartedPulling="2025-11-28 13:31:04.624727235 +0000 UTC m=+621.432030579" lastFinishedPulling="2025-11-28 13:31:12.813713979 +0000 UTC m=+629.621017323" observedRunningTime="2025-11-28 13:31:13.645158496 +0000 UTC m=+630.452461840" watchObservedRunningTime="2025-11-28 13:31:13.645570855 +0000 UTC m=+630.452874199" Nov 28 13:31:24 crc kubenswrapper[4631]: I1128 13:31:24.354624 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-fc587766c-d92lf" Nov 28 13:31:39 crc kubenswrapper[4631]: I1128 13:31:39.962686 4631 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Nov 28 13:31:44 crc kubenswrapper[4631]: I1128 13:31:44.096325 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-7c98947894-pw5mq" Nov 28 13:31:44 crc kubenswrapper[4631]: I1128 13:31:44.963584 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-9vxk7"] Nov 28 13:31:44 crc kubenswrapper[4631]: I1128 13:31:44.964423 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-9vxk7" Nov 28 13:31:44 crc kubenswrapper[4631]: I1128 13:31:44.967416 4631 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Nov 28 13:31:44 crc kubenswrapper[4631]: I1128 13:31:44.967689 4631 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-2ls9j" Nov 28 13:31:44 crc kubenswrapper[4631]: I1128 13:31:44.972280 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-5cpds"] Nov 28 13:31:44 crc kubenswrapper[4631]: I1128 13:31:44.975727 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-5cpds" Nov 28 13:31:44 crc kubenswrapper[4631]: I1128 13:31:44.981567 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Nov 28 13:31:44 crc kubenswrapper[4631]: I1128 13:31:44.982065 4631 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Nov 28 13:31:44 crc kubenswrapper[4631]: I1128 13:31:44.983000 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/15c5a33d-b3c0-45aa-963f-b6d9aa639edd-frr-sockets\") pod \"frr-k8s-5cpds\" (UID: \"15c5a33d-b3c0-45aa-963f-b6d9aa639edd\") " pod="metallb-system/frr-k8s-5cpds" Nov 28 13:31:44 crc kubenswrapper[4631]: I1128 13:31:44.983027 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/020cecb2-380e-4c2a-8741-eee44e230abf-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-9vxk7\" (UID: \"020cecb2-380e-4c2a-8741-eee44e230abf\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-9vxk7" Nov 28 13:31:44 crc kubenswrapper[4631]: I1128 13:31:44.983055 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/15c5a33d-b3c0-45aa-963f-b6d9aa639edd-frr-conf\") pod \"frr-k8s-5cpds\" (UID: \"15c5a33d-b3c0-45aa-963f-b6d9aa639edd\") " pod="metallb-system/frr-k8s-5cpds" Nov 28 13:31:44 crc kubenswrapper[4631]: I1128 13:31:44.983081 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rmvlg\" (UniqueName: \"kubernetes.io/projected/15c5a33d-b3c0-45aa-963f-b6d9aa639edd-kube-api-access-rmvlg\") pod \"frr-k8s-5cpds\" (UID: \"15c5a33d-b3c0-45aa-963f-b6d9aa639edd\") " pod="metallb-system/frr-k8s-5cpds" Nov 28 13:31:44 crc kubenswrapper[4631]: I1128 13:31:44.983109 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/15c5a33d-b3c0-45aa-963f-b6d9aa639edd-reloader\") pod \"frr-k8s-5cpds\" (UID: \"15c5a33d-b3c0-45aa-963f-b6d9aa639edd\") " pod="metallb-system/frr-k8s-5cpds" Nov 28 13:31:44 crc kubenswrapper[4631]: I1128 13:31:44.983128 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/15c5a33d-b3c0-45aa-963f-b6d9aa639edd-metrics\") pod \"frr-k8s-5cpds\" (UID: \"15c5a33d-b3c0-45aa-963f-b6d9aa639edd\") " pod="metallb-system/frr-k8s-5cpds" Nov 28 13:31:44 crc kubenswrapper[4631]: I1128 13:31:44.983151 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/15c5a33d-b3c0-45aa-963f-b6d9aa639edd-frr-startup\") pod \"frr-k8s-5cpds\" (UID: \"15c5a33d-b3c0-45aa-963f-b6d9aa639edd\") " pod="metallb-system/frr-k8s-5cpds" Nov 28 13:31:44 crc kubenswrapper[4631]: I1128 13:31:44.983168 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/15c5a33d-b3c0-45aa-963f-b6d9aa639edd-metrics-certs\") pod \"frr-k8s-5cpds\" (UID: \"15c5a33d-b3c0-45aa-963f-b6d9aa639edd\") " pod="metallb-system/frr-k8s-5cpds" Nov 28 13:31:44 crc kubenswrapper[4631]: I1128 13:31:44.983188 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rxkz8\" (UniqueName: \"kubernetes.io/projected/020cecb2-380e-4c2a-8741-eee44e230abf-kube-api-access-rxkz8\") pod \"frr-k8s-webhook-server-7fcb986d4-9vxk7\" (UID: \"020cecb2-380e-4c2a-8741-eee44e230abf\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-9vxk7" Nov 28 13:31:44 crc kubenswrapper[4631]: I1128 13:31:44.987519 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-9vxk7"] Nov 28 13:31:45 crc kubenswrapper[4631]: I1128 13:31:45.083603 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/15c5a33d-b3c0-45aa-963f-b6d9aa639edd-frr-startup\") pod \"frr-k8s-5cpds\" (UID: \"15c5a33d-b3c0-45aa-963f-b6d9aa639edd\") " pod="metallb-system/frr-k8s-5cpds" Nov 28 13:31:45 crc kubenswrapper[4631]: I1128 13:31:45.083647 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/15c5a33d-b3c0-45aa-963f-b6d9aa639edd-metrics-certs\") pod \"frr-k8s-5cpds\" (UID: \"15c5a33d-b3c0-45aa-963f-b6d9aa639edd\") " pod="metallb-system/frr-k8s-5cpds" Nov 28 13:31:45 crc kubenswrapper[4631]: I1128 13:31:45.083675 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rxkz8\" (UniqueName: \"kubernetes.io/projected/020cecb2-380e-4c2a-8741-eee44e230abf-kube-api-access-rxkz8\") pod \"frr-k8s-webhook-server-7fcb986d4-9vxk7\" (UID: \"020cecb2-380e-4c2a-8741-eee44e230abf\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-9vxk7" Nov 28 13:31:45 crc kubenswrapper[4631]: I1128 13:31:45.083724 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/15c5a33d-b3c0-45aa-963f-b6d9aa639edd-frr-sockets\") pod \"frr-k8s-5cpds\" (UID: \"15c5a33d-b3c0-45aa-963f-b6d9aa639edd\") " pod="metallb-system/frr-k8s-5cpds" Nov 28 13:31:45 crc kubenswrapper[4631]: I1128 13:31:45.083744 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/020cecb2-380e-4c2a-8741-eee44e230abf-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-9vxk7\" (UID: \"020cecb2-380e-4c2a-8741-eee44e230abf\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-9vxk7" Nov 28 13:31:45 crc kubenswrapper[4631]: I1128 13:31:45.083778 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/15c5a33d-b3c0-45aa-963f-b6d9aa639edd-frr-conf\") pod \"frr-k8s-5cpds\" (UID: \"15c5a33d-b3c0-45aa-963f-b6d9aa639edd\") " pod="metallb-system/frr-k8s-5cpds" Nov 28 13:31:45 crc kubenswrapper[4631]: I1128 13:31:45.083811 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rmvlg\" (UniqueName: \"kubernetes.io/projected/15c5a33d-b3c0-45aa-963f-b6d9aa639edd-kube-api-access-rmvlg\") pod \"frr-k8s-5cpds\" (UID: \"15c5a33d-b3c0-45aa-963f-b6d9aa639edd\") " pod="metallb-system/frr-k8s-5cpds" Nov 28 13:31:45 crc kubenswrapper[4631]: I1128 13:31:45.083848 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/15c5a33d-b3c0-45aa-963f-b6d9aa639edd-reloader\") pod \"frr-k8s-5cpds\" (UID: \"15c5a33d-b3c0-45aa-963f-b6d9aa639edd\") " pod="metallb-system/frr-k8s-5cpds" Nov 28 13:31:45 crc kubenswrapper[4631]: I1128 13:31:45.083876 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/15c5a33d-b3c0-45aa-963f-b6d9aa639edd-metrics\") pod \"frr-k8s-5cpds\" (UID: \"15c5a33d-b3c0-45aa-963f-b6d9aa639edd\") " pod="metallb-system/frr-k8s-5cpds" Nov 28 13:31:45 crc kubenswrapper[4631]: E1128 13:31:45.084026 4631 secret.go:188] Couldn't get secret metallb-system/frr-k8s-webhook-server-cert: secret "frr-k8s-webhook-server-cert" not found Nov 28 13:31:45 crc kubenswrapper[4631]: E1128 13:31:45.084096 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/020cecb2-380e-4c2a-8741-eee44e230abf-cert podName:020cecb2-380e-4c2a-8741-eee44e230abf nodeName:}" failed. No retries permitted until 2025-11-28 13:31:45.584074572 +0000 UTC m=+662.391377906 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/020cecb2-380e-4c2a-8741-eee44e230abf-cert") pod "frr-k8s-webhook-server-7fcb986d4-9vxk7" (UID: "020cecb2-380e-4c2a-8741-eee44e230abf") : secret "frr-k8s-webhook-server-cert" not found Nov 28 13:31:45 crc kubenswrapper[4631]: E1128 13:31:45.084413 4631 secret.go:188] Couldn't get secret metallb-system/frr-k8s-certs-secret: secret "frr-k8s-certs-secret" not found Nov 28 13:31:45 crc kubenswrapper[4631]: E1128 13:31:45.084467 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/15c5a33d-b3c0-45aa-963f-b6d9aa639edd-metrics-certs podName:15c5a33d-b3c0-45aa-963f-b6d9aa639edd nodeName:}" failed. No retries permitted until 2025-11-28 13:31:45.584455761 +0000 UTC m=+662.391759115 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/15c5a33d-b3c0-45aa-963f-b6d9aa639edd-metrics-certs") pod "frr-k8s-5cpds" (UID: "15c5a33d-b3c0-45aa-963f-b6d9aa639edd") : secret "frr-k8s-certs-secret" not found Nov 28 13:31:45 crc kubenswrapper[4631]: I1128 13:31:45.085160 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/15c5a33d-b3c0-45aa-963f-b6d9aa639edd-frr-conf\") pod \"frr-k8s-5cpds\" (UID: \"15c5a33d-b3c0-45aa-963f-b6d9aa639edd\") " pod="metallb-system/frr-k8s-5cpds" Nov 28 13:31:45 crc kubenswrapper[4631]: I1128 13:31:45.085192 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/15c5a33d-b3c0-45aa-963f-b6d9aa639edd-metrics\") pod \"frr-k8s-5cpds\" (UID: \"15c5a33d-b3c0-45aa-963f-b6d9aa639edd\") " pod="metallb-system/frr-k8s-5cpds" Nov 28 13:31:45 crc kubenswrapper[4631]: I1128 13:31:45.085270 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/15c5a33d-b3c0-45aa-963f-b6d9aa639edd-frr-sockets\") pod \"frr-k8s-5cpds\" (UID: \"15c5a33d-b3c0-45aa-963f-b6d9aa639edd\") " pod="metallb-system/frr-k8s-5cpds" Nov 28 13:31:45 crc kubenswrapper[4631]: I1128 13:31:45.085306 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/15c5a33d-b3c0-45aa-963f-b6d9aa639edd-reloader\") pod \"frr-k8s-5cpds\" (UID: \"15c5a33d-b3c0-45aa-963f-b6d9aa639edd\") " pod="metallb-system/frr-k8s-5cpds" Nov 28 13:31:45 crc kubenswrapper[4631]: I1128 13:31:45.085408 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/15c5a33d-b3c0-45aa-963f-b6d9aa639edd-frr-startup\") pod \"frr-k8s-5cpds\" (UID: \"15c5a33d-b3c0-45aa-963f-b6d9aa639edd\") " pod="metallb-system/frr-k8s-5cpds" Nov 28 13:31:45 crc kubenswrapper[4631]: I1128 13:31:45.119811 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rxkz8\" (UniqueName: \"kubernetes.io/projected/020cecb2-380e-4c2a-8741-eee44e230abf-kube-api-access-rxkz8\") pod \"frr-k8s-webhook-server-7fcb986d4-9vxk7\" (UID: \"020cecb2-380e-4c2a-8741-eee44e230abf\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-9vxk7" Nov 28 13:31:45 crc kubenswrapper[4631]: I1128 13:31:45.122108 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rmvlg\" (UniqueName: \"kubernetes.io/projected/15c5a33d-b3c0-45aa-963f-b6d9aa639edd-kube-api-access-rmvlg\") pod \"frr-k8s-5cpds\" (UID: \"15c5a33d-b3c0-45aa-963f-b6d9aa639edd\") " pod="metallb-system/frr-k8s-5cpds" Nov 28 13:31:45 crc kubenswrapper[4631]: I1128 13:31:45.124128 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-d24rl"] Nov 28 13:31:45 crc kubenswrapper[4631]: I1128 13:31:45.135239 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-d24rl" Nov 28 13:31:45 crc kubenswrapper[4631]: I1128 13:31:45.139025 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-f8648f98b-fpkw8"] Nov 28 13:31:45 crc kubenswrapper[4631]: I1128 13:31:45.140115 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-f8648f98b-fpkw8" Nov 28 13:31:45 crc kubenswrapper[4631]: I1128 13:31:45.141213 4631 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Nov 28 13:31:45 crc kubenswrapper[4631]: I1128 13:31:45.141419 4631 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Nov 28 13:31:45 crc kubenswrapper[4631]: I1128 13:31:45.141606 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Nov 28 13:31:45 crc kubenswrapper[4631]: I1128 13:31:45.141899 4631 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Nov 28 13:31:45 crc kubenswrapper[4631]: I1128 13:31:45.142072 4631 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-qtbxc" Nov 28 13:31:45 crc kubenswrapper[4631]: I1128 13:31:45.167187 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-f8648f98b-fpkw8"] Nov 28 13:31:45 crc kubenswrapper[4631]: I1128 13:31:45.184811 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/255213e3-20a3-4b0b-a165-c598ecce4766-metallb-excludel2\") pod \"speaker-d24rl\" (UID: \"255213e3-20a3-4b0b-a165-c598ecce4766\") " pod="metallb-system/speaker-d24rl" Nov 28 13:31:45 crc kubenswrapper[4631]: I1128 13:31:45.184860 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f9e35a8e-a1f5-40f9-b673-533ee2f1425f-cert\") pod \"controller-f8648f98b-fpkw8\" (UID: \"f9e35a8e-a1f5-40f9-b673-533ee2f1425f\") " pod="metallb-system/controller-f8648f98b-fpkw8" Nov 28 13:31:45 crc kubenswrapper[4631]: I1128 13:31:45.184913 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b22dw\" (UniqueName: \"kubernetes.io/projected/255213e3-20a3-4b0b-a165-c598ecce4766-kube-api-access-b22dw\") pod \"speaker-d24rl\" (UID: \"255213e3-20a3-4b0b-a165-c598ecce4766\") " pod="metallb-system/speaker-d24rl" Nov 28 13:31:45 crc kubenswrapper[4631]: I1128 13:31:45.185003 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/255213e3-20a3-4b0b-a165-c598ecce4766-memberlist\") pod \"speaker-d24rl\" (UID: \"255213e3-20a3-4b0b-a165-c598ecce4766\") " pod="metallb-system/speaker-d24rl" Nov 28 13:31:45 crc kubenswrapper[4631]: I1128 13:31:45.185186 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/255213e3-20a3-4b0b-a165-c598ecce4766-metrics-certs\") pod \"speaker-d24rl\" (UID: \"255213e3-20a3-4b0b-a165-c598ecce4766\") " pod="metallb-system/speaker-d24rl" Nov 28 13:31:45 crc kubenswrapper[4631]: I1128 13:31:45.185218 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f9e35a8e-a1f5-40f9-b673-533ee2f1425f-metrics-certs\") pod \"controller-f8648f98b-fpkw8\" (UID: \"f9e35a8e-a1f5-40f9-b673-533ee2f1425f\") " pod="metallb-system/controller-f8648f98b-fpkw8" Nov 28 13:31:45 crc kubenswrapper[4631]: I1128 13:31:45.185242 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bdgn6\" (UniqueName: \"kubernetes.io/projected/f9e35a8e-a1f5-40f9-b673-533ee2f1425f-kube-api-access-bdgn6\") pod \"controller-f8648f98b-fpkw8\" (UID: \"f9e35a8e-a1f5-40f9-b673-533ee2f1425f\") " pod="metallb-system/controller-f8648f98b-fpkw8" Nov 28 13:31:45 crc kubenswrapper[4631]: I1128 13:31:45.286003 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/255213e3-20a3-4b0b-a165-c598ecce4766-metrics-certs\") pod \"speaker-d24rl\" (UID: \"255213e3-20a3-4b0b-a165-c598ecce4766\") " pod="metallb-system/speaker-d24rl" Nov 28 13:31:45 crc kubenswrapper[4631]: I1128 13:31:45.286052 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f9e35a8e-a1f5-40f9-b673-533ee2f1425f-metrics-certs\") pod \"controller-f8648f98b-fpkw8\" (UID: \"f9e35a8e-a1f5-40f9-b673-533ee2f1425f\") " pod="metallb-system/controller-f8648f98b-fpkw8" Nov 28 13:31:45 crc kubenswrapper[4631]: I1128 13:31:45.286072 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bdgn6\" (UniqueName: \"kubernetes.io/projected/f9e35a8e-a1f5-40f9-b673-533ee2f1425f-kube-api-access-bdgn6\") pod \"controller-f8648f98b-fpkw8\" (UID: \"f9e35a8e-a1f5-40f9-b673-533ee2f1425f\") " pod="metallb-system/controller-f8648f98b-fpkw8" Nov 28 13:31:45 crc kubenswrapper[4631]: I1128 13:31:45.286108 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/255213e3-20a3-4b0b-a165-c598ecce4766-metallb-excludel2\") pod \"speaker-d24rl\" (UID: \"255213e3-20a3-4b0b-a165-c598ecce4766\") " pod="metallb-system/speaker-d24rl" Nov 28 13:31:45 crc kubenswrapper[4631]: I1128 13:31:45.286125 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f9e35a8e-a1f5-40f9-b673-533ee2f1425f-cert\") pod \"controller-f8648f98b-fpkw8\" (UID: \"f9e35a8e-a1f5-40f9-b673-533ee2f1425f\") " pod="metallb-system/controller-f8648f98b-fpkw8" Nov 28 13:31:45 crc kubenswrapper[4631]: I1128 13:31:45.286140 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b22dw\" (UniqueName: \"kubernetes.io/projected/255213e3-20a3-4b0b-a165-c598ecce4766-kube-api-access-b22dw\") pod \"speaker-d24rl\" (UID: \"255213e3-20a3-4b0b-a165-c598ecce4766\") " pod="metallb-system/speaker-d24rl" Nov 28 13:31:45 crc kubenswrapper[4631]: I1128 13:31:45.286159 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/255213e3-20a3-4b0b-a165-c598ecce4766-memberlist\") pod \"speaker-d24rl\" (UID: \"255213e3-20a3-4b0b-a165-c598ecce4766\") " pod="metallb-system/speaker-d24rl" Nov 28 13:31:45 crc kubenswrapper[4631]: E1128 13:31:45.286265 4631 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Nov 28 13:31:45 crc kubenswrapper[4631]: E1128 13:31:45.286330 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/255213e3-20a3-4b0b-a165-c598ecce4766-memberlist podName:255213e3-20a3-4b0b-a165-c598ecce4766 nodeName:}" failed. No retries permitted until 2025-11-28 13:31:45.786314721 +0000 UTC m=+662.593618065 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/255213e3-20a3-4b0b-a165-c598ecce4766-memberlist") pod "speaker-d24rl" (UID: "255213e3-20a3-4b0b-a165-c598ecce4766") : secret "metallb-memberlist" not found Nov 28 13:31:45 crc kubenswrapper[4631]: I1128 13:31:45.286991 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/255213e3-20a3-4b0b-a165-c598ecce4766-metallb-excludel2\") pod \"speaker-d24rl\" (UID: \"255213e3-20a3-4b0b-a165-c598ecce4766\") " pod="metallb-system/speaker-d24rl" Nov 28 13:31:45 crc kubenswrapper[4631]: E1128 13:31:45.287080 4631 secret.go:188] Couldn't get secret metallb-system/controller-certs-secret: secret "controller-certs-secret" not found Nov 28 13:31:45 crc kubenswrapper[4631]: E1128 13:31:45.287125 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f9e35a8e-a1f5-40f9-b673-533ee2f1425f-metrics-certs podName:f9e35a8e-a1f5-40f9-b673-533ee2f1425f nodeName:}" failed. No retries permitted until 2025-11-28 13:31:45.78711433 +0000 UTC m=+662.594417674 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/f9e35a8e-a1f5-40f9-b673-533ee2f1425f-metrics-certs") pod "controller-f8648f98b-fpkw8" (UID: "f9e35a8e-a1f5-40f9-b673-533ee2f1425f") : secret "controller-certs-secret" not found Nov 28 13:31:45 crc kubenswrapper[4631]: I1128 13:31:45.291553 4631 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Nov 28 13:31:45 crc kubenswrapper[4631]: I1128 13:31:45.297142 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/255213e3-20a3-4b0b-a165-c598ecce4766-metrics-certs\") pod \"speaker-d24rl\" (UID: \"255213e3-20a3-4b0b-a165-c598ecce4766\") " pod="metallb-system/speaker-d24rl" Nov 28 13:31:45 crc kubenswrapper[4631]: I1128 13:31:45.300449 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f9e35a8e-a1f5-40f9-b673-533ee2f1425f-cert\") pod \"controller-f8648f98b-fpkw8\" (UID: \"f9e35a8e-a1f5-40f9-b673-533ee2f1425f\") " pod="metallb-system/controller-f8648f98b-fpkw8" Nov 28 13:31:45 crc kubenswrapper[4631]: I1128 13:31:45.321596 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bdgn6\" (UniqueName: \"kubernetes.io/projected/f9e35a8e-a1f5-40f9-b673-533ee2f1425f-kube-api-access-bdgn6\") pod \"controller-f8648f98b-fpkw8\" (UID: \"f9e35a8e-a1f5-40f9-b673-533ee2f1425f\") " pod="metallb-system/controller-f8648f98b-fpkw8" Nov 28 13:31:45 crc kubenswrapper[4631]: I1128 13:31:45.325679 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b22dw\" (UniqueName: \"kubernetes.io/projected/255213e3-20a3-4b0b-a165-c598ecce4766-kube-api-access-b22dw\") pod \"speaker-d24rl\" (UID: \"255213e3-20a3-4b0b-a165-c598ecce4766\") " pod="metallb-system/speaker-d24rl" Nov 28 13:31:45 crc kubenswrapper[4631]: I1128 13:31:45.590961 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/020cecb2-380e-4c2a-8741-eee44e230abf-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-9vxk7\" (UID: \"020cecb2-380e-4c2a-8741-eee44e230abf\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-9vxk7" Nov 28 13:31:45 crc kubenswrapper[4631]: I1128 13:31:45.591095 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/15c5a33d-b3c0-45aa-963f-b6d9aa639edd-metrics-certs\") pod \"frr-k8s-5cpds\" (UID: \"15c5a33d-b3c0-45aa-963f-b6d9aa639edd\") " pod="metallb-system/frr-k8s-5cpds" Nov 28 13:31:45 crc kubenswrapper[4631]: I1128 13:31:45.594469 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/15c5a33d-b3c0-45aa-963f-b6d9aa639edd-metrics-certs\") pod \"frr-k8s-5cpds\" (UID: \"15c5a33d-b3c0-45aa-963f-b6d9aa639edd\") " pod="metallb-system/frr-k8s-5cpds" Nov 28 13:31:45 crc kubenswrapper[4631]: I1128 13:31:45.595344 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/020cecb2-380e-4c2a-8741-eee44e230abf-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-9vxk7\" (UID: \"020cecb2-380e-4c2a-8741-eee44e230abf\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-9vxk7" Nov 28 13:31:45 crc kubenswrapper[4631]: I1128 13:31:45.794513 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f9e35a8e-a1f5-40f9-b673-533ee2f1425f-metrics-certs\") pod \"controller-f8648f98b-fpkw8\" (UID: \"f9e35a8e-a1f5-40f9-b673-533ee2f1425f\") " pod="metallb-system/controller-f8648f98b-fpkw8" Nov 28 13:31:45 crc kubenswrapper[4631]: I1128 13:31:45.794712 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/255213e3-20a3-4b0b-a165-c598ecce4766-memberlist\") pod \"speaker-d24rl\" (UID: \"255213e3-20a3-4b0b-a165-c598ecce4766\") " pod="metallb-system/speaker-d24rl" Nov 28 13:31:45 crc kubenswrapper[4631]: E1128 13:31:45.794927 4631 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Nov 28 13:31:45 crc kubenswrapper[4631]: E1128 13:31:45.795035 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/255213e3-20a3-4b0b-a165-c598ecce4766-memberlist podName:255213e3-20a3-4b0b-a165-c598ecce4766 nodeName:}" failed. No retries permitted until 2025-11-28 13:31:46.794995597 +0000 UTC m=+663.602298941 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/255213e3-20a3-4b0b-a165-c598ecce4766-memberlist") pod "speaker-d24rl" (UID: "255213e3-20a3-4b0b-a165-c598ecce4766") : secret "metallb-memberlist" not found Nov 28 13:31:45 crc kubenswrapper[4631]: I1128 13:31:45.799090 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f9e35a8e-a1f5-40f9-b673-533ee2f1425f-metrics-certs\") pod \"controller-f8648f98b-fpkw8\" (UID: \"f9e35a8e-a1f5-40f9-b673-533ee2f1425f\") " pod="metallb-system/controller-f8648f98b-fpkw8" Nov 28 13:31:45 crc kubenswrapper[4631]: I1128 13:31:45.807177 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-f8648f98b-fpkw8" Nov 28 13:31:45 crc kubenswrapper[4631]: I1128 13:31:45.882016 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-9vxk7" Nov 28 13:31:45 crc kubenswrapper[4631]: I1128 13:31:45.893071 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-5cpds" Nov 28 13:31:46 crc kubenswrapper[4631]: I1128 13:31:46.177768 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-f8648f98b-fpkw8"] Nov 28 13:31:46 crc kubenswrapper[4631]: I1128 13:31:46.259886 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-9vxk7"] Nov 28 13:31:46 crc kubenswrapper[4631]: W1128 13:31:46.269068 4631 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod020cecb2_380e_4c2a_8741_eee44e230abf.slice/crio-fa31e1b46a201d725a252a4b86f037bd3b1d7209211c42e3e08cf041f0fbf5e4 WatchSource:0}: Error finding container fa31e1b46a201d725a252a4b86f037bd3b1d7209211c42e3e08cf041f0fbf5e4: Status 404 returned error can't find the container with id fa31e1b46a201d725a252a4b86f037bd3b1d7209211c42e3e08cf041f0fbf5e4 Nov 28 13:31:46 crc kubenswrapper[4631]: I1128 13:31:46.812925 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/255213e3-20a3-4b0b-a165-c598ecce4766-memberlist\") pod \"speaker-d24rl\" (UID: \"255213e3-20a3-4b0b-a165-c598ecce4766\") " pod="metallb-system/speaker-d24rl" Nov 28 13:31:46 crc kubenswrapper[4631]: E1128 13:31:46.813517 4631 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Nov 28 13:31:46 crc kubenswrapper[4631]: E1128 13:31:46.813597 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/255213e3-20a3-4b0b-a165-c598ecce4766-memberlist podName:255213e3-20a3-4b0b-a165-c598ecce4766 nodeName:}" failed. No retries permitted until 2025-11-28 13:31:48.813577689 +0000 UTC m=+665.620881043 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/255213e3-20a3-4b0b-a165-c598ecce4766-memberlist") pod "speaker-d24rl" (UID: "255213e3-20a3-4b0b-a165-c598ecce4766") : secret "metallb-memberlist" not found Nov 28 13:31:46 crc kubenswrapper[4631]: I1128 13:31:46.817499 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-9vxk7" event={"ID":"020cecb2-380e-4c2a-8741-eee44e230abf","Type":"ContainerStarted","Data":"fa31e1b46a201d725a252a4b86f037bd3b1d7209211c42e3e08cf041f0fbf5e4"} Nov 28 13:31:46 crc kubenswrapper[4631]: I1128 13:31:46.822367 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-fpkw8" event={"ID":"f9e35a8e-a1f5-40f9-b673-533ee2f1425f","Type":"ContainerStarted","Data":"feb78c0d23bfc606f91acc2f935f2170dbdbc1b0dc3a41ab3c279f111afe8ee8"} Nov 28 13:31:46 crc kubenswrapper[4631]: I1128 13:31:46.822434 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-fpkw8" event={"ID":"f9e35a8e-a1f5-40f9-b673-533ee2f1425f","Type":"ContainerStarted","Data":"85b8e79c24a85a37521bba6b99a034fdc9573d15e4cd9bc7ed8d1723d61b8c96"} Nov 28 13:31:46 crc kubenswrapper[4631]: I1128 13:31:46.823905 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-5cpds" event={"ID":"15c5a33d-b3c0-45aa-963f-b6d9aa639edd","Type":"ContainerStarted","Data":"58958ec8dc87bf9a86f1e171b44f13ae925425a49e59fa0a4376a73fd0a2d316"} Nov 28 13:31:47 crc kubenswrapper[4631]: I1128 13:31:47.837136 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-fpkw8" event={"ID":"f9e35a8e-a1f5-40f9-b673-533ee2f1425f","Type":"ContainerStarted","Data":"f11fab73fee547a6333edd086fb0462d36d460e21e13f9dc6cd54144d6dd860d"} Nov 28 13:31:47 crc kubenswrapper[4631]: I1128 13:31:47.838729 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-f8648f98b-fpkw8" Nov 28 13:31:47 crc kubenswrapper[4631]: I1128 13:31:47.916631 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-f8648f98b-fpkw8" podStartSLOduration=2.916610713 podStartE2EDuration="2.916610713s" podCreationTimestamp="2025-11-28 13:31:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 13:31:47.896326801 +0000 UTC m=+664.703630155" watchObservedRunningTime="2025-11-28 13:31:47.916610713 +0000 UTC m=+664.723914057" Nov 28 13:31:48 crc kubenswrapper[4631]: I1128 13:31:48.848347 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/255213e3-20a3-4b0b-a165-c598ecce4766-memberlist\") pod \"speaker-d24rl\" (UID: \"255213e3-20a3-4b0b-a165-c598ecce4766\") " pod="metallb-system/speaker-d24rl" Nov 28 13:31:48 crc kubenswrapper[4631]: I1128 13:31:48.860625 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/255213e3-20a3-4b0b-a165-c598ecce4766-memberlist\") pod \"speaker-d24rl\" (UID: \"255213e3-20a3-4b0b-a165-c598ecce4766\") " pod="metallb-system/speaker-d24rl" Nov 28 13:31:49 crc kubenswrapper[4631]: I1128 13:31:49.078672 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-d24rl" Nov 28 13:31:49 crc kubenswrapper[4631]: I1128 13:31:49.863761 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-d24rl" event={"ID":"255213e3-20a3-4b0b-a165-c598ecce4766","Type":"ContainerStarted","Data":"5c23271e2534e19ef5ab9f361c2944bf6e5181c6285b5cd6dd6e4a7d289ca733"} Nov 28 13:31:49 crc kubenswrapper[4631]: I1128 13:31:49.864065 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-d24rl" event={"ID":"255213e3-20a3-4b0b-a165-c598ecce4766","Type":"ContainerStarted","Data":"3966576e485baebd372367db073624e29055b529255327120cb5f39221d448d5"} Nov 28 13:31:50 crc kubenswrapper[4631]: I1128 13:31:50.871426 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-d24rl" event={"ID":"255213e3-20a3-4b0b-a165-c598ecce4766","Type":"ContainerStarted","Data":"3172e9045df5d93059dfb8b68ec21fa3f95a1e3497fb49891395ddaeb96b677a"} Nov 28 13:31:50 crc kubenswrapper[4631]: I1128 13:31:50.872055 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-d24rl" Nov 28 13:31:50 crc kubenswrapper[4631]: I1128 13:31:50.892805 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-d24rl" podStartSLOduration=5.89278212 podStartE2EDuration="5.89278212s" podCreationTimestamp="2025-11-28 13:31:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 13:31:50.888644258 +0000 UTC m=+667.695947592" watchObservedRunningTime="2025-11-28 13:31:50.89278212 +0000 UTC m=+667.700085464" Nov 28 13:31:56 crc kubenswrapper[4631]: I1128 13:31:56.910315 4631 generic.go:334] "Generic (PLEG): container finished" podID="15c5a33d-b3c0-45aa-963f-b6d9aa639edd" containerID="4a4124253b1f6d18594359672204f9f3632882525a8d8b70f5d5a091e0af1529" exitCode=0 Nov 28 13:31:56 crc kubenswrapper[4631]: I1128 13:31:56.910440 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-5cpds" event={"ID":"15c5a33d-b3c0-45aa-963f-b6d9aa639edd","Type":"ContainerDied","Data":"4a4124253b1f6d18594359672204f9f3632882525a8d8b70f5d5a091e0af1529"} Nov 28 13:31:56 crc kubenswrapper[4631]: I1128 13:31:56.912375 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-9vxk7" event={"ID":"020cecb2-380e-4c2a-8741-eee44e230abf","Type":"ContainerStarted","Data":"134b9f75c4c88086db0710f648fc5df257865ff0655c2ac959b349282edf607e"} Nov 28 13:31:56 crc kubenswrapper[4631]: I1128 13:31:56.912555 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-9vxk7" Nov 28 13:31:56 crc kubenswrapper[4631]: I1128 13:31:56.971926 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-9vxk7" podStartSLOduration=4.211502956 podStartE2EDuration="12.971904664s" podCreationTimestamp="2025-11-28 13:31:44 +0000 UTC" firstStartedPulling="2025-11-28 13:31:46.271202349 +0000 UTC m=+663.078505703" lastFinishedPulling="2025-11-28 13:31:55.031604047 +0000 UTC m=+671.838907411" observedRunningTime="2025-11-28 13:31:56.968958911 +0000 UTC m=+673.776262265" watchObservedRunningTime="2025-11-28 13:31:56.971904664 +0000 UTC m=+673.779208018" Nov 28 13:31:57 crc kubenswrapper[4631]: I1128 13:31:57.924665 4631 generic.go:334] "Generic (PLEG): container finished" podID="15c5a33d-b3c0-45aa-963f-b6d9aa639edd" containerID="2cac8faa80efd94c1bf5f2929f523e56e0f61f59c93e3cdee69d134545de47ba" exitCode=0 Nov 28 13:31:57 crc kubenswrapper[4631]: I1128 13:31:57.924729 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-5cpds" event={"ID":"15c5a33d-b3c0-45aa-963f-b6d9aa639edd","Type":"ContainerDied","Data":"2cac8faa80efd94c1bf5f2929f523e56e0f61f59c93e3cdee69d134545de47ba"} Nov 28 13:31:58 crc kubenswrapper[4631]: I1128 13:31:58.935469 4631 generic.go:334] "Generic (PLEG): container finished" podID="15c5a33d-b3c0-45aa-963f-b6d9aa639edd" containerID="18d580a03c53ff96602fe8f02aad0ec3d7b0b113e15b08ff9dc8778ab39626b7" exitCode=0 Nov 28 13:31:58 crc kubenswrapper[4631]: I1128 13:31:58.935850 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-5cpds" event={"ID":"15c5a33d-b3c0-45aa-963f-b6d9aa639edd","Type":"ContainerDied","Data":"18d580a03c53ff96602fe8f02aad0ec3d7b0b113e15b08ff9dc8778ab39626b7"} Nov 28 13:31:59 crc kubenswrapper[4631]: I1128 13:31:59.085092 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-d24rl" Nov 28 13:31:59 crc kubenswrapper[4631]: I1128 13:31:59.960283 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-5cpds" event={"ID":"15c5a33d-b3c0-45aa-963f-b6d9aa639edd","Type":"ContainerStarted","Data":"eba1b21ab9fe9ca4ac1334c3b701c2a967961b431b73aaf85c457c69fd50f187"} Nov 28 13:31:59 crc kubenswrapper[4631]: I1128 13:31:59.960716 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-5cpds" event={"ID":"15c5a33d-b3c0-45aa-963f-b6d9aa639edd","Type":"ContainerStarted","Data":"7ef6d8c3fd493fd227c3f3614b26cb600c734d2ae10f99c814e0ad6aa75b2d2c"} Nov 28 13:31:59 crc kubenswrapper[4631]: I1128 13:31:59.960744 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-5cpds" event={"ID":"15c5a33d-b3c0-45aa-963f-b6d9aa639edd","Type":"ContainerStarted","Data":"e7d004bc8b8179e3e6e72a18c7b75e8b8920250ac0309832f8f80bd3cf588a16"} Nov 28 13:32:00 crc kubenswrapper[4631]: I1128 13:32:00.980130 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-5cpds" event={"ID":"15c5a33d-b3c0-45aa-963f-b6d9aa639edd","Type":"ContainerStarted","Data":"3f2e4fba8d595e8af64861ce6eb7b321ffbfd7b30d21a0ff0b68866a315375a4"} Nov 28 13:32:00 crc kubenswrapper[4631]: I1128 13:32:00.980536 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-5cpds" event={"ID":"15c5a33d-b3c0-45aa-963f-b6d9aa639edd","Type":"ContainerStarted","Data":"89e0519e5c809d53443193eb4dc5e830269a1aff2ad08e1c79b08fe305248efa"} Nov 28 13:32:00 crc kubenswrapper[4631]: I1128 13:32:00.980546 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-5cpds" event={"ID":"15c5a33d-b3c0-45aa-963f-b6d9aa639edd","Type":"ContainerStarted","Data":"dd9b450c26d165b3afb716e689f8d748728d4a2fcef4934548e811e39ac63180"} Nov 28 13:32:00 crc kubenswrapper[4631]: I1128 13:32:00.980559 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-5cpds" Nov 28 13:32:01 crc kubenswrapper[4631]: I1128 13:32:01.003669 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-5cpds" podStartSLOduration=8.668722936 podStartE2EDuration="17.003648829s" podCreationTimestamp="2025-11-28 13:31:44 +0000 UTC" firstStartedPulling="2025-11-28 13:31:46.725327273 +0000 UTC m=+663.532630617" lastFinishedPulling="2025-11-28 13:31:55.060253166 +0000 UTC m=+671.867556510" observedRunningTime="2025-11-28 13:32:01.002367547 +0000 UTC m=+677.809670891" watchObservedRunningTime="2025-11-28 13:32:01.003648829 +0000 UTC m=+677.810952173" Nov 28 13:32:01 crc kubenswrapper[4631]: I1128 13:32:01.908061 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-n8h2h"] Nov 28 13:32:01 crc kubenswrapper[4631]: I1128 13:32:01.908996 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-n8h2h" Nov 28 13:32:01 crc kubenswrapper[4631]: I1128 13:32:01.918163 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Nov 28 13:32:01 crc kubenswrapper[4631]: I1128 13:32:01.919736 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-sbsvs" Nov 28 13:32:01 crc kubenswrapper[4631]: I1128 13:32:01.920669 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Nov 28 13:32:01 crc kubenswrapper[4631]: I1128 13:32:01.932727 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-n8h2h"] Nov 28 13:32:02 crc kubenswrapper[4631]: I1128 13:32:02.065688 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5drk7\" (UniqueName: \"kubernetes.io/projected/96db9727-d349-49a4-8dc4-08cc7b68284d-kube-api-access-5drk7\") pod \"openstack-operator-index-n8h2h\" (UID: \"96db9727-d349-49a4-8dc4-08cc7b68284d\") " pod="openstack-operators/openstack-operator-index-n8h2h" Nov 28 13:32:02 crc kubenswrapper[4631]: I1128 13:32:02.167602 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5drk7\" (UniqueName: \"kubernetes.io/projected/96db9727-d349-49a4-8dc4-08cc7b68284d-kube-api-access-5drk7\") pod \"openstack-operator-index-n8h2h\" (UID: \"96db9727-d349-49a4-8dc4-08cc7b68284d\") " pod="openstack-operators/openstack-operator-index-n8h2h" Nov 28 13:32:02 crc kubenswrapper[4631]: I1128 13:32:02.191359 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5drk7\" (UniqueName: \"kubernetes.io/projected/96db9727-d349-49a4-8dc4-08cc7b68284d-kube-api-access-5drk7\") pod \"openstack-operator-index-n8h2h\" (UID: \"96db9727-d349-49a4-8dc4-08cc7b68284d\") " pod="openstack-operators/openstack-operator-index-n8h2h" Nov 28 13:32:02 crc kubenswrapper[4631]: I1128 13:32:02.236723 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-n8h2h" Nov 28 13:32:02 crc kubenswrapper[4631]: I1128 13:32:02.480815 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-n8h2h"] Nov 28 13:32:02 crc kubenswrapper[4631]: I1128 13:32:02.994696 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-n8h2h" event={"ID":"96db9727-d349-49a4-8dc4-08cc7b68284d","Type":"ContainerStarted","Data":"784ddd843213a1200b8d83eb91e8c3a76b9e8c6d01472b412b74fb8d6731c9f0"} Nov 28 13:32:05 crc kubenswrapper[4631]: I1128 13:32:05.083773 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-n8h2h"] Nov 28 13:32:05 crc kubenswrapper[4631]: I1128 13:32:05.701323 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-qclsp"] Nov 28 13:32:05 crc kubenswrapper[4631]: I1128 13:32:05.702870 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-qclsp" Nov 28 13:32:05 crc kubenswrapper[4631]: I1128 13:32:05.728310 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-qclsp"] Nov 28 13:32:05 crc kubenswrapper[4631]: I1128 13:32:05.813156 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-f8648f98b-fpkw8" Nov 28 13:32:05 crc kubenswrapper[4631]: I1128 13:32:05.838890 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h9k2z\" (UniqueName: \"kubernetes.io/projected/8c8c3b22-0767-4d9f-9ec4-bf42ddcd8433-kube-api-access-h9k2z\") pod \"openstack-operator-index-qclsp\" (UID: \"8c8c3b22-0767-4d9f-9ec4-bf42ddcd8433\") " pod="openstack-operators/openstack-operator-index-qclsp" Nov 28 13:32:05 crc kubenswrapper[4631]: I1128 13:32:05.887088 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-9vxk7" Nov 28 13:32:05 crc kubenswrapper[4631]: I1128 13:32:05.893933 4631 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-5cpds" Nov 28 13:32:05 crc kubenswrapper[4631]: I1128 13:32:05.935962 4631 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-5cpds" Nov 28 13:32:05 crc kubenswrapper[4631]: I1128 13:32:05.941549 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h9k2z\" (UniqueName: \"kubernetes.io/projected/8c8c3b22-0767-4d9f-9ec4-bf42ddcd8433-kube-api-access-h9k2z\") pod \"openstack-operator-index-qclsp\" (UID: \"8c8c3b22-0767-4d9f-9ec4-bf42ddcd8433\") " pod="openstack-operators/openstack-operator-index-qclsp" Nov 28 13:32:05 crc kubenswrapper[4631]: I1128 13:32:05.967243 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h9k2z\" (UniqueName: \"kubernetes.io/projected/8c8c3b22-0767-4d9f-9ec4-bf42ddcd8433-kube-api-access-h9k2z\") pod \"openstack-operator-index-qclsp\" (UID: \"8c8c3b22-0767-4d9f-9ec4-bf42ddcd8433\") " pod="openstack-operators/openstack-operator-index-qclsp" Nov 28 13:32:06 crc kubenswrapper[4631]: I1128 13:32:06.045789 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-qclsp" Nov 28 13:32:09 crc kubenswrapper[4631]: I1128 13:32:09.586559 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-qclsp"] Nov 28 13:32:09 crc kubenswrapper[4631]: W1128 13:32:09.590193 4631 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8c8c3b22_0767_4d9f_9ec4_bf42ddcd8433.slice/crio-38942bf55d784a2927573e09f4d755900cc4e823d4d1ae12a16bcd46632ee411 WatchSource:0}: Error finding container 38942bf55d784a2927573e09f4d755900cc4e823d4d1ae12a16bcd46632ee411: Status 404 returned error can't find the container with id 38942bf55d784a2927573e09f4d755900cc4e823d4d1ae12a16bcd46632ee411 Nov 28 13:32:10 crc kubenswrapper[4631]: I1128 13:32:10.054698 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-qclsp" event={"ID":"8c8c3b22-0767-4d9f-9ec4-bf42ddcd8433","Type":"ContainerStarted","Data":"38942bf55d784a2927573e09f4d755900cc4e823d4d1ae12a16bcd46632ee411"} Nov 28 13:32:11 crc kubenswrapper[4631]: I1128 13:32:11.063435 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-qclsp" event={"ID":"8c8c3b22-0767-4d9f-9ec4-bf42ddcd8433","Type":"ContainerStarted","Data":"2f291adc71aeda5bbe3bf535729b38d2ecb5522cd44ff1524a629f8cd78ad0e6"} Nov 28 13:32:11 crc kubenswrapper[4631]: I1128 13:32:11.066748 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-n8h2h" event={"ID":"96db9727-d349-49a4-8dc4-08cc7b68284d","Type":"ContainerStarted","Data":"11f7a50e8713335cefefde196132bfffd5058ff5f457104f9173e589b3a70ff8"} Nov 28 13:32:11 crc kubenswrapper[4631]: I1128 13:32:11.066919 4631 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/openstack-operator-index-n8h2h" podUID="96db9727-d349-49a4-8dc4-08cc7b68284d" containerName="registry-server" containerID="cri-o://11f7a50e8713335cefefde196132bfffd5058ff5f457104f9173e589b3a70ff8" gracePeriod=2 Nov 28 13:32:11 crc kubenswrapper[4631]: I1128 13:32:11.093243 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-qclsp" podStartSLOduration=5.515849353 podStartE2EDuration="6.0932142s" podCreationTimestamp="2025-11-28 13:32:05 +0000 UTC" firstStartedPulling="2025-11-28 13:32:09.593504784 +0000 UTC m=+686.400808148" lastFinishedPulling="2025-11-28 13:32:10.170869641 +0000 UTC m=+686.978172995" observedRunningTime="2025-11-28 13:32:11.091638511 +0000 UTC m=+687.898941895" watchObservedRunningTime="2025-11-28 13:32:11.0932142 +0000 UTC m=+687.900517574" Nov 28 13:32:11 crc kubenswrapper[4631]: I1128 13:32:11.120383 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-n8h2h" podStartSLOduration=2.442777927 podStartE2EDuration="10.120356652s" podCreationTimestamp="2025-11-28 13:32:01 +0000 UTC" firstStartedPulling="2025-11-28 13:32:02.495354327 +0000 UTC m=+679.302657671" lastFinishedPulling="2025-11-28 13:32:10.172933052 +0000 UTC m=+686.980236396" observedRunningTime="2025-11-28 13:32:11.11662693 +0000 UTC m=+687.923930334" watchObservedRunningTime="2025-11-28 13:32:11.120356652 +0000 UTC m=+687.927660026" Nov 28 13:32:12 crc kubenswrapper[4631]: I1128 13:32:12.078953 4631 generic.go:334] "Generic (PLEG): container finished" podID="96db9727-d349-49a4-8dc4-08cc7b68284d" containerID="11f7a50e8713335cefefde196132bfffd5058ff5f457104f9173e589b3a70ff8" exitCode=0 Nov 28 13:32:12 crc kubenswrapper[4631]: I1128 13:32:12.079076 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-n8h2h" event={"ID":"96db9727-d349-49a4-8dc4-08cc7b68284d","Type":"ContainerDied","Data":"11f7a50e8713335cefefde196132bfffd5058ff5f457104f9173e589b3a70ff8"} Nov 28 13:32:12 crc kubenswrapper[4631]: I1128 13:32:12.236881 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-n8h2h" Nov 28 13:32:13 crc kubenswrapper[4631]: I1128 13:32:13.554885 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-n8h2h" Nov 28 13:32:13 crc kubenswrapper[4631]: I1128 13:32:13.702904 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5drk7\" (UniqueName: \"kubernetes.io/projected/96db9727-d349-49a4-8dc4-08cc7b68284d-kube-api-access-5drk7\") pod \"96db9727-d349-49a4-8dc4-08cc7b68284d\" (UID: \"96db9727-d349-49a4-8dc4-08cc7b68284d\") " Nov 28 13:32:13 crc kubenswrapper[4631]: I1128 13:32:13.713641 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96db9727-d349-49a4-8dc4-08cc7b68284d-kube-api-access-5drk7" (OuterVolumeSpecName: "kube-api-access-5drk7") pod "96db9727-d349-49a4-8dc4-08cc7b68284d" (UID: "96db9727-d349-49a4-8dc4-08cc7b68284d"). InnerVolumeSpecName "kube-api-access-5drk7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:32:13 crc kubenswrapper[4631]: I1128 13:32:13.805542 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5drk7\" (UniqueName: \"kubernetes.io/projected/96db9727-d349-49a4-8dc4-08cc7b68284d-kube-api-access-5drk7\") on node \"crc\" DevicePath \"\"" Nov 28 13:32:14 crc kubenswrapper[4631]: I1128 13:32:14.100696 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-n8h2h" event={"ID":"96db9727-d349-49a4-8dc4-08cc7b68284d","Type":"ContainerDied","Data":"784ddd843213a1200b8d83eb91e8c3a76b9e8c6d01472b412b74fb8d6731c9f0"} Nov 28 13:32:14 crc kubenswrapper[4631]: I1128 13:32:14.100794 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-n8h2h" Nov 28 13:32:14 crc kubenswrapper[4631]: I1128 13:32:14.100829 4631 scope.go:117] "RemoveContainer" containerID="11f7a50e8713335cefefde196132bfffd5058ff5f457104f9173e589b3a70ff8" Nov 28 13:32:14 crc kubenswrapper[4631]: I1128 13:32:14.142452 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-n8h2h"] Nov 28 13:32:14 crc kubenswrapper[4631]: I1128 13:32:14.146456 4631 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/openstack-operator-index-n8h2h"] Nov 28 13:32:15 crc kubenswrapper[4631]: I1128 13:32:15.530330 4631 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96db9727-d349-49a4-8dc4-08cc7b68284d" path="/var/lib/kubelet/pods/96db9727-d349-49a4-8dc4-08cc7b68284d/volumes" Nov 28 13:32:15 crc kubenswrapper[4631]: I1128 13:32:15.897508 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-5cpds" Nov 28 13:32:16 crc kubenswrapper[4631]: I1128 13:32:16.046128 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-qclsp" Nov 28 13:32:16 crc kubenswrapper[4631]: I1128 13:32:16.046221 4631 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-qclsp" Nov 28 13:32:16 crc kubenswrapper[4631]: I1128 13:32:16.082969 4631 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-qclsp" Nov 28 13:32:16 crc kubenswrapper[4631]: I1128 13:32:16.156307 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-qclsp" Nov 28 13:32:23 crc kubenswrapper[4631]: I1128 13:32:23.371360 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/dee8d58050df5e5c4d52f5078d12ac93857da536112fc471b2066b915e9wbcx"] Nov 28 13:32:23 crc kubenswrapper[4631]: E1128 13:32:23.371721 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="96db9727-d349-49a4-8dc4-08cc7b68284d" containerName="registry-server" Nov 28 13:32:23 crc kubenswrapper[4631]: I1128 13:32:23.371739 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="96db9727-d349-49a4-8dc4-08cc7b68284d" containerName="registry-server" Nov 28 13:32:23 crc kubenswrapper[4631]: I1128 13:32:23.371899 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="96db9727-d349-49a4-8dc4-08cc7b68284d" containerName="registry-server" Nov 28 13:32:23 crc kubenswrapper[4631]: I1128 13:32:23.373033 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/dee8d58050df5e5c4d52f5078d12ac93857da536112fc471b2066b915e9wbcx" Nov 28 13:32:23 crc kubenswrapper[4631]: I1128 13:32:23.375585 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-7g29l" Nov 28 13:32:23 crc kubenswrapper[4631]: I1128 13:32:23.394448 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/dee8d58050df5e5c4d52f5078d12ac93857da536112fc471b2066b915e9wbcx"] Nov 28 13:32:23 crc kubenswrapper[4631]: I1128 13:32:23.448402 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/57ea9be9-7467-4c95-b00b-14f7df44eae4-util\") pod \"dee8d58050df5e5c4d52f5078d12ac93857da536112fc471b2066b915e9wbcx\" (UID: \"57ea9be9-7467-4c95-b00b-14f7df44eae4\") " pod="openstack-operators/dee8d58050df5e5c4d52f5078d12ac93857da536112fc471b2066b915e9wbcx" Nov 28 13:32:23 crc kubenswrapper[4631]: I1128 13:32:23.449583 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/57ea9be9-7467-4c95-b00b-14f7df44eae4-bundle\") pod \"dee8d58050df5e5c4d52f5078d12ac93857da536112fc471b2066b915e9wbcx\" (UID: \"57ea9be9-7467-4c95-b00b-14f7df44eae4\") " pod="openstack-operators/dee8d58050df5e5c4d52f5078d12ac93857da536112fc471b2066b915e9wbcx" Nov 28 13:32:23 crc kubenswrapper[4631]: I1128 13:32:23.449831 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xkw9t\" (UniqueName: \"kubernetes.io/projected/57ea9be9-7467-4c95-b00b-14f7df44eae4-kube-api-access-xkw9t\") pod \"dee8d58050df5e5c4d52f5078d12ac93857da536112fc471b2066b915e9wbcx\" (UID: \"57ea9be9-7467-4c95-b00b-14f7df44eae4\") " pod="openstack-operators/dee8d58050df5e5c4d52f5078d12ac93857da536112fc471b2066b915e9wbcx" Nov 28 13:32:23 crc kubenswrapper[4631]: I1128 13:32:23.551131 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xkw9t\" (UniqueName: \"kubernetes.io/projected/57ea9be9-7467-4c95-b00b-14f7df44eae4-kube-api-access-xkw9t\") pod \"dee8d58050df5e5c4d52f5078d12ac93857da536112fc471b2066b915e9wbcx\" (UID: \"57ea9be9-7467-4c95-b00b-14f7df44eae4\") " pod="openstack-operators/dee8d58050df5e5c4d52f5078d12ac93857da536112fc471b2066b915e9wbcx" Nov 28 13:32:23 crc kubenswrapper[4631]: I1128 13:32:23.551191 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/57ea9be9-7467-4c95-b00b-14f7df44eae4-util\") pod \"dee8d58050df5e5c4d52f5078d12ac93857da536112fc471b2066b915e9wbcx\" (UID: \"57ea9be9-7467-4c95-b00b-14f7df44eae4\") " pod="openstack-operators/dee8d58050df5e5c4d52f5078d12ac93857da536112fc471b2066b915e9wbcx" Nov 28 13:32:23 crc kubenswrapper[4631]: I1128 13:32:23.551214 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/57ea9be9-7467-4c95-b00b-14f7df44eae4-bundle\") pod \"dee8d58050df5e5c4d52f5078d12ac93857da536112fc471b2066b915e9wbcx\" (UID: \"57ea9be9-7467-4c95-b00b-14f7df44eae4\") " pod="openstack-operators/dee8d58050df5e5c4d52f5078d12ac93857da536112fc471b2066b915e9wbcx" Nov 28 13:32:23 crc kubenswrapper[4631]: I1128 13:32:23.551750 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/57ea9be9-7467-4c95-b00b-14f7df44eae4-util\") pod \"dee8d58050df5e5c4d52f5078d12ac93857da536112fc471b2066b915e9wbcx\" (UID: \"57ea9be9-7467-4c95-b00b-14f7df44eae4\") " pod="openstack-operators/dee8d58050df5e5c4d52f5078d12ac93857da536112fc471b2066b915e9wbcx" Nov 28 13:32:23 crc kubenswrapper[4631]: I1128 13:32:23.551769 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/57ea9be9-7467-4c95-b00b-14f7df44eae4-bundle\") pod \"dee8d58050df5e5c4d52f5078d12ac93857da536112fc471b2066b915e9wbcx\" (UID: \"57ea9be9-7467-4c95-b00b-14f7df44eae4\") " pod="openstack-operators/dee8d58050df5e5c4d52f5078d12ac93857da536112fc471b2066b915e9wbcx" Nov 28 13:32:23 crc kubenswrapper[4631]: I1128 13:32:23.573570 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xkw9t\" (UniqueName: \"kubernetes.io/projected/57ea9be9-7467-4c95-b00b-14f7df44eae4-kube-api-access-xkw9t\") pod \"dee8d58050df5e5c4d52f5078d12ac93857da536112fc471b2066b915e9wbcx\" (UID: \"57ea9be9-7467-4c95-b00b-14f7df44eae4\") " pod="openstack-operators/dee8d58050df5e5c4d52f5078d12ac93857da536112fc471b2066b915e9wbcx" Nov 28 13:32:23 crc kubenswrapper[4631]: I1128 13:32:23.708491 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/dee8d58050df5e5c4d52f5078d12ac93857da536112fc471b2066b915e9wbcx" Nov 28 13:32:24 crc kubenswrapper[4631]: I1128 13:32:24.153936 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/dee8d58050df5e5c4d52f5078d12ac93857da536112fc471b2066b915e9wbcx"] Nov 28 13:32:24 crc kubenswrapper[4631]: I1128 13:32:24.871276 4631 generic.go:334] "Generic (PLEG): container finished" podID="57ea9be9-7467-4c95-b00b-14f7df44eae4" containerID="9bfb7897cedb785811f8f95acd1cbb617ec57daccd63531a81abd2f546caea26" exitCode=0 Nov 28 13:32:24 crc kubenswrapper[4631]: I1128 13:32:24.871431 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/dee8d58050df5e5c4d52f5078d12ac93857da536112fc471b2066b915e9wbcx" event={"ID":"57ea9be9-7467-4c95-b00b-14f7df44eae4","Type":"ContainerDied","Data":"9bfb7897cedb785811f8f95acd1cbb617ec57daccd63531a81abd2f546caea26"} Nov 28 13:32:24 crc kubenswrapper[4631]: I1128 13:32:24.871887 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/dee8d58050df5e5c4d52f5078d12ac93857da536112fc471b2066b915e9wbcx" event={"ID":"57ea9be9-7467-4c95-b00b-14f7df44eae4","Type":"ContainerStarted","Data":"ed9d1c126ee4b486e2ac9460e3b668308d96e54db1138abc13a4dc2af2afc108"} Nov 28 13:32:25 crc kubenswrapper[4631]: I1128 13:32:25.886160 4631 generic.go:334] "Generic (PLEG): container finished" podID="57ea9be9-7467-4c95-b00b-14f7df44eae4" containerID="3b1f4c542b61e9a30e960ff55f4483082bf60b09b51c72ee4b3d541f1f6b5fe2" exitCode=0 Nov 28 13:32:25 crc kubenswrapper[4631]: I1128 13:32:25.886227 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/dee8d58050df5e5c4d52f5078d12ac93857da536112fc471b2066b915e9wbcx" event={"ID":"57ea9be9-7467-4c95-b00b-14f7df44eae4","Type":"ContainerDied","Data":"3b1f4c542b61e9a30e960ff55f4483082bf60b09b51c72ee4b3d541f1f6b5fe2"} Nov 28 13:32:26 crc kubenswrapper[4631]: I1128 13:32:26.897593 4631 generic.go:334] "Generic (PLEG): container finished" podID="57ea9be9-7467-4c95-b00b-14f7df44eae4" containerID="fdd42b823dbbcda53ae8b06f4a857a9dfcdbfa0adfcb2ef52d8cdc47855ff655" exitCode=0 Nov 28 13:32:26 crc kubenswrapper[4631]: I1128 13:32:26.897662 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/dee8d58050df5e5c4d52f5078d12ac93857da536112fc471b2066b915e9wbcx" event={"ID":"57ea9be9-7467-4c95-b00b-14f7df44eae4","Type":"ContainerDied","Data":"fdd42b823dbbcda53ae8b06f4a857a9dfcdbfa0adfcb2ef52d8cdc47855ff655"} Nov 28 13:32:28 crc kubenswrapper[4631]: I1128 13:32:28.229502 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/dee8d58050df5e5c4d52f5078d12ac93857da536112fc471b2066b915e9wbcx" Nov 28 13:32:28 crc kubenswrapper[4631]: I1128 13:32:28.239330 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xkw9t\" (UniqueName: \"kubernetes.io/projected/57ea9be9-7467-4c95-b00b-14f7df44eae4-kube-api-access-xkw9t\") pod \"57ea9be9-7467-4c95-b00b-14f7df44eae4\" (UID: \"57ea9be9-7467-4c95-b00b-14f7df44eae4\") " Nov 28 13:32:28 crc kubenswrapper[4631]: I1128 13:32:28.239568 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/57ea9be9-7467-4c95-b00b-14f7df44eae4-bundle\") pod \"57ea9be9-7467-4c95-b00b-14f7df44eae4\" (UID: \"57ea9be9-7467-4c95-b00b-14f7df44eae4\") " Nov 28 13:32:28 crc kubenswrapper[4631]: I1128 13:32:28.239701 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/57ea9be9-7467-4c95-b00b-14f7df44eae4-util\") pod \"57ea9be9-7467-4c95-b00b-14f7df44eae4\" (UID: \"57ea9be9-7467-4c95-b00b-14f7df44eae4\") " Nov 28 13:32:28 crc kubenswrapper[4631]: I1128 13:32:28.240580 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57ea9be9-7467-4c95-b00b-14f7df44eae4-bundle" (OuterVolumeSpecName: "bundle") pod "57ea9be9-7467-4c95-b00b-14f7df44eae4" (UID: "57ea9be9-7467-4c95-b00b-14f7df44eae4"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 13:32:28 crc kubenswrapper[4631]: I1128 13:32:28.251569 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57ea9be9-7467-4c95-b00b-14f7df44eae4-kube-api-access-xkw9t" (OuterVolumeSpecName: "kube-api-access-xkw9t") pod "57ea9be9-7467-4c95-b00b-14f7df44eae4" (UID: "57ea9be9-7467-4c95-b00b-14f7df44eae4"). InnerVolumeSpecName "kube-api-access-xkw9t". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:32:28 crc kubenswrapper[4631]: I1128 13:32:28.274776 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57ea9be9-7467-4c95-b00b-14f7df44eae4-util" (OuterVolumeSpecName: "util") pod "57ea9be9-7467-4c95-b00b-14f7df44eae4" (UID: "57ea9be9-7467-4c95-b00b-14f7df44eae4"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 13:32:28 crc kubenswrapper[4631]: I1128 13:32:28.342116 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xkw9t\" (UniqueName: \"kubernetes.io/projected/57ea9be9-7467-4c95-b00b-14f7df44eae4-kube-api-access-xkw9t\") on node \"crc\" DevicePath \"\"" Nov 28 13:32:28 crc kubenswrapper[4631]: I1128 13:32:28.342172 4631 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/57ea9be9-7467-4c95-b00b-14f7df44eae4-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 13:32:28 crc kubenswrapper[4631]: I1128 13:32:28.342185 4631 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/57ea9be9-7467-4c95-b00b-14f7df44eae4-util\") on node \"crc\" DevicePath \"\"" Nov 28 13:32:28 crc kubenswrapper[4631]: I1128 13:32:28.919461 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/dee8d58050df5e5c4d52f5078d12ac93857da536112fc471b2066b915e9wbcx" event={"ID":"57ea9be9-7467-4c95-b00b-14f7df44eae4","Type":"ContainerDied","Data":"ed9d1c126ee4b486e2ac9460e3b668308d96e54db1138abc13a4dc2af2afc108"} Nov 28 13:32:28 crc kubenswrapper[4631]: I1128 13:32:28.919521 4631 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ed9d1c126ee4b486e2ac9460e3b668308d96e54db1138abc13a4dc2af2afc108" Nov 28 13:32:28 crc kubenswrapper[4631]: I1128 13:32:28.919553 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/dee8d58050df5e5c4d52f5078d12ac93857da536112fc471b2066b915e9wbcx" Nov 28 13:32:36 crc kubenswrapper[4631]: I1128 13:32:36.136429 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-operator-5bcf64b564-zsrq9"] Nov 28 13:32:36 crc kubenswrapper[4631]: E1128 13:32:36.137371 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="57ea9be9-7467-4c95-b00b-14f7df44eae4" containerName="pull" Nov 28 13:32:36 crc kubenswrapper[4631]: I1128 13:32:36.137386 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="57ea9be9-7467-4c95-b00b-14f7df44eae4" containerName="pull" Nov 28 13:32:36 crc kubenswrapper[4631]: E1128 13:32:36.137397 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="57ea9be9-7467-4c95-b00b-14f7df44eae4" containerName="extract" Nov 28 13:32:36 crc kubenswrapper[4631]: I1128 13:32:36.137403 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="57ea9be9-7467-4c95-b00b-14f7df44eae4" containerName="extract" Nov 28 13:32:36 crc kubenswrapper[4631]: E1128 13:32:36.137410 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="57ea9be9-7467-4c95-b00b-14f7df44eae4" containerName="util" Nov 28 13:32:36 crc kubenswrapper[4631]: I1128 13:32:36.137417 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="57ea9be9-7467-4c95-b00b-14f7df44eae4" containerName="util" Nov 28 13:32:36 crc kubenswrapper[4631]: I1128 13:32:36.137513 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="57ea9be9-7467-4c95-b00b-14f7df44eae4" containerName="extract" Nov 28 13:32:36 crc kubenswrapper[4631]: I1128 13:32:36.138076 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-5bcf64b564-zsrq9" Nov 28 13:32:36 crc kubenswrapper[4631]: I1128 13:32:36.142689 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-operator-dockercfg-rfq8t" Nov 28 13:32:36 crc kubenswrapper[4631]: I1128 13:32:36.180068 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-5bcf64b564-zsrq9"] Nov 28 13:32:36 crc kubenswrapper[4631]: I1128 13:32:36.180948 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-497ds\" (UniqueName: \"kubernetes.io/projected/1104dd12-c822-4ffa-9634-d82898f8f575-kube-api-access-497ds\") pod \"openstack-operator-controller-operator-5bcf64b564-zsrq9\" (UID: \"1104dd12-c822-4ffa-9634-d82898f8f575\") " pod="openstack-operators/openstack-operator-controller-operator-5bcf64b564-zsrq9" Nov 28 13:32:36 crc kubenswrapper[4631]: I1128 13:32:36.282039 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-497ds\" (UniqueName: \"kubernetes.io/projected/1104dd12-c822-4ffa-9634-d82898f8f575-kube-api-access-497ds\") pod \"openstack-operator-controller-operator-5bcf64b564-zsrq9\" (UID: \"1104dd12-c822-4ffa-9634-d82898f8f575\") " pod="openstack-operators/openstack-operator-controller-operator-5bcf64b564-zsrq9" Nov 28 13:32:36 crc kubenswrapper[4631]: I1128 13:32:36.302769 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-497ds\" (UniqueName: \"kubernetes.io/projected/1104dd12-c822-4ffa-9634-d82898f8f575-kube-api-access-497ds\") pod \"openstack-operator-controller-operator-5bcf64b564-zsrq9\" (UID: \"1104dd12-c822-4ffa-9634-d82898f8f575\") " pod="openstack-operators/openstack-operator-controller-operator-5bcf64b564-zsrq9" Nov 28 13:32:36 crc kubenswrapper[4631]: I1128 13:32:36.456809 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-5bcf64b564-zsrq9" Nov 28 13:32:36 crc kubenswrapper[4631]: I1128 13:32:36.718940 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-5bcf64b564-zsrq9"] Nov 28 13:32:36 crc kubenswrapper[4631]: W1128 13:32:36.737388 4631 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1104dd12_c822_4ffa_9634_d82898f8f575.slice/crio-4df135687d45e52746f82c4fd050c728ce9a3add11b5cfb166dff79686865da4 WatchSource:0}: Error finding container 4df135687d45e52746f82c4fd050c728ce9a3add11b5cfb166dff79686865da4: Status 404 returned error can't find the container with id 4df135687d45e52746f82c4fd050c728ce9a3add11b5cfb166dff79686865da4 Nov 28 13:32:36 crc kubenswrapper[4631]: I1128 13:32:36.998728 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-5bcf64b564-zsrq9" event={"ID":"1104dd12-c822-4ffa-9634-d82898f8f575","Type":"ContainerStarted","Data":"4df135687d45e52746f82c4fd050c728ce9a3add11b5cfb166dff79686865da4"} Nov 28 13:32:45 crc kubenswrapper[4631]: I1128 13:32:45.062837 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-5bcf64b564-zsrq9" event={"ID":"1104dd12-c822-4ffa-9634-d82898f8f575","Type":"ContainerStarted","Data":"a2d1c3f5897b48eecd91a2214be07bbb2fb23ae254fcadddffb30aa1a18b263e"} Nov 28 13:32:45 crc kubenswrapper[4631]: I1128 13:32:45.063400 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-operator-5bcf64b564-zsrq9" Nov 28 13:32:45 crc kubenswrapper[4631]: I1128 13:32:45.118130 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-operator-5bcf64b564-zsrq9" podStartSLOduration=1.874325102 podStartE2EDuration="9.118108361s" podCreationTimestamp="2025-11-28 13:32:36 +0000 UTC" firstStartedPulling="2025-11-28 13:32:36.740216447 +0000 UTC m=+713.547519791" lastFinishedPulling="2025-11-28 13:32:43.983999706 +0000 UTC m=+720.791303050" observedRunningTime="2025-11-28 13:32:45.110699267 +0000 UTC m=+721.918002611" watchObservedRunningTime="2025-11-28 13:32:45.118108361 +0000 UTC m=+721.925411695" Nov 28 13:32:56 crc kubenswrapper[4631]: I1128 13:32:56.462413 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-operator-5bcf64b564-zsrq9" Nov 28 13:33:19 crc kubenswrapper[4631]: I1128 13:33:19.634594 4631 patch_prober.go:28] interesting pod/machine-config-daemon-47sxw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 28 13:33:19 crc kubenswrapper[4631]: I1128 13:33:19.635219 4631 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" podUID="8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 28 13:33:23 crc kubenswrapper[4631]: I1128 13:33:23.685243 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7b64f4fb85-hdr72"] Nov 28 13:33:23 crc kubenswrapper[4631]: I1128 13:33:23.686718 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-7b64f4fb85-hdr72" Nov 28 13:33:23 crc kubenswrapper[4631]: I1128 13:33:23.692121 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-bfdg6" Nov 28 13:33:23 crc kubenswrapper[4631]: I1128 13:33:23.714904 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-6b7f75547b-rmlgq"] Nov 28 13:33:23 crc kubenswrapper[4631]: I1128 13:33:23.716077 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-6b7f75547b-rmlgq" Nov 28 13:33:23 crc kubenswrapper[4631]: I1128 13:33:23.719139 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-gwtpl" Nov 28 13:33:23 crc kubenswrapper[4631]: I1128 13:33:23.719523 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-955677c94-6bt2f"] Nov 28 13:33:23 crc kubenswrapper[4631]: I1128 13:33:23.720639 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-955677c94-6bt2f" Nov 28 13:33:23 crc kubenswrapper[4631]: I1128 13:33:23.728826 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7b64f4fb85-hdr72"] Nov 28 13:33:23 crc kubenswrapper[4631]: I1128 13:33:23.729942 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-5bmnl" Nov 28 13:33:23 crc kubenswrapper[4631]: I1128 13:33:23.751099 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-955677c94-6bt2f"] Nov 28 13:33:23 crc kubenswrapper[4631]: I1128 13:33:23.760501 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-6b7f75547b-rmlgq"] Nov 28 13:33:23 crc kubenswrapper[4631]: I1128 13:33:23.795626 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-589cbd6b5b-npdtc"] Nov 28 13:33:23 crc kubenswrapper[4631]: I1128 13:33:23.796242 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-plw2d\" (UniqueName: \"kubernetes.io/projected/00a81f80-35b9-4eb4-a508-3f104ca426be-kube-api-access-plw2d\") pod \"barbican-operator-controller-manager-7b64f4fb85-hdr72\" (UID: \"00a81f80-35b9-4eb4-a508-3f104ca426be\") " pod="openstack-operators/barbican-operator-controller-manager-7b64f4fb85-hdr72" Nov 28 13:33:23 crc kubenswrapper[4631]: I1128 13:33:23.806277 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-589cbd6b5b-npdtc"] Nov 28 13:33:23 crc kubenswrapper[4631]: I1128 13:33:23.806437 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-589cbd6b5b-npdtc" Nov 28 13:33:23 crc kubenswrapper[4631]: I1128 13:33:23.808830 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-jjm8d" Nov 28 13:33:23 crc kubenswrapper[4631]: I1128 13:33:23.846727 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-5b77f656f-8pvq9"] Nov 28 13:33:23 crc kubenswrapper[4631]: I1128 13:33:23.855913 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-5b77f656f-8pvq9" Nov 28 13:33:23 crc kubenswrapper[4631]: I1128 13:33:23.858745 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-chdkg" Nov 28 13:33:23 crc kubenswrapper[4631]: I1128 13:33:23.897958 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-plw2d\" (UniqueName: \"kubernetes.io/projected/00a81f80-35b9-4eb4-a508-3f104ca426be-kube-api-access-plw2d\") pod \"barbican-operator-controller-manager-7b64f4fb85-hdr72\" (UID: \"00a81f80-35b9-4eb4-a508-3f104ca426be\") " pod="openstack-operators/barbican-operator-controller-manager-7b64f4fb85-hdr72" Nov 28 13:33:23 crc kubenswrapper[4631]: I1128 13:33:23.898044 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cvp2c\" (UniqueName: \"kubernetes.io/projected/8028d28d-2414-4ffe-9c5d-4854f77e980f-kube-api-access-cvp2c\") pod \"cinder-operator-controller-manager-6b7f75547b-rmlgq\" (UID: \"8028d28d-2414-4ffe-9c5d-4854f77e980f\") " pod="openstack-operators/cinder-operator-controller-manager-6b7f75547b-rmlgq" Nov 28 13:33:23 crc kubenswrapper[4631]: I1128 13:33:23.898079 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l59zp\" (UniqueName: \"kubernetes.io/projected/06994d92-3dcb-461c-83ac-c7e65d4bd640-kube-api-access-l59zp\") pod \"designate-operator-controller-manager-955677c94-6bt2f\" (UID: \"06994d92-3dcb-461c-83ac-c7e65d4bd640\") " pod="openstack-operators/designate-operator-controller-manager-955677c94-6bt2f" Nov 28 13:33:23 crc kubenswrapper[4631]: I1128 13:33:23.903586 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-5d494799bf-2mq7v"] Nov 28 13:33:23 crc kubenswrapper[4631]: I1128 13:33:23.904828 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-5d494799bf-2mq7v" Nov 28 13:33:23 crc kubenswrapper[4631]: I1128 13:33:23.913488 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-5qx2m" Nov 28 13:33:23 crc kubenswrapper[4631]: I1128 13:33:23.921069 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-5b77f656f-8pvq9"] Nov 28 13:33:23 crc kubenswrapper[4631]: I1128 13:33:23.936380 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-5d494799bf-2mq7v"] Nov 28 13:33:23 crc kubenswrapper[4631]: I1128 13:33:23.954703 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-57548d458d-h7npd"] Nov 28 13:33:23 crc kubenswrapper[4631]: I1128 13:33:23.956469 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-57548d458d-h7npd" Nov 28 13:33:23 crc kubenswrapper[4631]: I1128 13:33:23.968351 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Nov 28 13:33:23 crc kubenswrapper[4631]: I1128 13:33:23.968593 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-cgmfv" Nov 28 13:33:23 crc kubenswrapper[4631]: I1128 13:33:23.991330 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-plw2d\" (UniqueName: \"kubernetes.io/projected/00a81f80-35b9-4eb4-a508-3f104ca426be-kube-api-access-plw2d\") pod \"barbican-operator-controller-manager-7b64f4fb85-hdr72\" (UID: \"00a81f80-35b9-4eb4-a508-3f104ca426be\") " pod="openstack-operators/barbican-operator-controller-manager-7b64f4fb85-hdr72" Nov 28 13:33:23 crc kubenswrapper[4631]: I1128 13:33:23.994996 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-67cb4dc6d4-db8kg"] Nov 28 13:33:23 crc kubenswrapper[4631]: I1128 13:33:23.996215 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-67cb4dc6d4-db8kg" Nov 28 13:33:24 crc kubenswrapper[4631]: I1128 13:33:24.002084 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vvzpr\" (UniqueName: \"kubernetes.io/projected/9cc92c5c-285e-4604-bc60-7637938a78c8-kube-api-access-vvzpr\") pod \"heat-operator-controller-manager-5b77f656f-8pvq9\" (UID: \"9cc92c5c-285e-4604-bc60-7637938a78c8\") " pod="openstack-operators/heat-operator-controller-manager-5b77f656f-8pvq9" Nov 28 13:33:24 crc kubenswrapper[4631]: I1128 13:33:24.002258 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l59zp\" (UniqueName: \"kubernetes.io/projected/06994d92-3dcb-461c-83ac-c7e65d4bd640-kube-api-access-l59zp\") pod \"designate-operator-controller-manager-955677c94-6bt2f\" (UID: \"06994d92-3dcb-461c-83ac-c7e65d4bd640\") " pod="openstack-operators/designate-operator-controller-manager-955677c94-6bt2f" Nov 28 13:33:24 crc kubenswrapper[4631]: I1128 13:33:24.002400 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ddkv7\" (UniqueName: \"kubernetes.io/projected/59ee25ad-f169-47b9-9d5c-f7a7a253a591-kube-api-access-ddkv7\") pod \"glance-operator-controller-manager-589cbd6b5b-npdtc\" (UID: \"59ee25ad-f169-47b9-9d5c-f7a7a253a591\") " pod="openstack-operators/glance-operator-controller-manager-589cbd6b5b-npdtc" Nov 28 13:33:24 crc kubenswrapper[4631]: I1128 13:33:24.002544 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cvp2c\" (UniqueName: \"kubernetes.io/projected/8028d28d-2414-4ffe-9c5d-4854f77e980f-kube-api-access-cvp2c\") pod \"cinder-operator-controller-manager-6b7f75547b-rmlgq\" (UID: \"8028d28d-2414-4ffe-9c5d-4854f77e980f\") " pod="openstack-operators/cinder-operator-controller-manager-6b7f75547b-rmlgq" Nov 28 13:33:24 crc kubenswrapper[4631]: I1128 13:33:24.006087 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-7b64f4fb85-hdr72" Nov 28 13:33:24 crc kubenswrapper[4631]: I1128 13:33:24.020779 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-prl9x" Nov 28 13:33:24 crc kubenswrapper[4631]: I1128 13:33:24.029369 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-57548d458d-h7npd"] Nov 28 13:33:24 crc kubenswrapper[4631]: I1128 13:33:24.047430 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l59zp\" (UniqueName: \"kubernetes.io/projected/06994d92-3dcb-461c-83ac-c7e65d4bd640-kube-api-access-l59zp\") pod \"designate-operator-controller-manager-955677c94-6bt2f\" (UID: \"06994d92-3dcb-461c-83ac-c7e65d4bd640\") " pod="openstack-operators/designate-operator-controller-manager-955677c94-6bt2f" Nov 28 13:33:24 crc kubenswrapper[4631]: I1128 13:33:24.064847 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cvp2c\" (UniqueName: \"kubernetes.io/projected/8028d28d-2414-4ffe-9c5d-4854f77e980f-kube-api-access-cvp2c\") pod \"cinder-operator-controller-manager-6b7f75547b-rmlgq\" (UID: \"8028d28d-2414-4ffe-9c5d-4854f77e980f\") " pod="openstack-operators/cinder-operator-controller-manager-6b7f75547b-rmlgq" Nov 28 13:33:24 crc kubenswrapper[4631]: I1128 13:33:24.080638 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-67cb4dc6d4-db8kg"] Nov 28 13:33:24 crc kubenswrapper[4631]: I1128 13:33:24.082948 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-6b49bfffcd-5ks5f"] Nov 28 13:33:24 crc kubenswrapper[4631]: I1128 13:33:24.084124 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-6b49bfffcd-5ks5f" Nov 28 13:33:24 crc kubenswrapper[4631]: I1128 13:33:24.089636 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-5h5r6" Nov 28 13:33:24 crc kubenswrapper[4631]: I1128 13:33:24.105224 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zhh2d\" (UniqueName: \"kubernetes.io/projected/aafff68d-b989-4edc-ab1c-b9d33edb8ee2-kube-api-access-zhh2d\") pod \"horizon-operator-controller-manager-5d494799bf-2mq7v\" (UID: \"aafff68d-b989-4edc-ab1c-b9d33edb8ee2\") " pod="openstack-operators/horizon-operator-controller-manager-5d494799bf-2mq7v" Nov 28 13:33:24 crc kubenswrapper[4631]: I1128 13:33:24.105309 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b44dl\" (UniqueName: \"kubernetes.io/projected/0079ab2d-c5a8-45eb-b6e8-9574158de32f-kube-api-access-b44dl\") pod \"infra-operator-controller-manager-57548d458d-h7npd\" (UID: \"0079ab2d-c5a8-45eb-b6e8-9574158de32f\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-h7npd" Nov 28 13:33:24 crc kubenswrapper[4631]: I1128 13:33:24.105338 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zn6mw\" (UniqueName: \"kubernetes.io/projected/ed17d71f-b72c-4f12-b5f7-d5cb3bb130dc-kube-api-access-zn6mw\") pod \"ironic-operator-controller-manager-67cb4dc6d4-db8kg\" (UID: \"ed17d71f-b72c-4f12-b5f7-d5cb3bb130dc\") " pod="openstack-operators/ironic-operator-controller-manager-67cb4dc6d4-db8kg" Nov 28 13:33:24 crc kubenswrapper[4631]: I1128 13:33:24.105374 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vvzpr\" (UniqueName: \"kubernetes.io/projected/9cc92c5c-285e-4604-bc60-7637938a78c8-kube-api-access-vvzpr\") pod \"heat-operator-controller-manager-5b77f656f-8pvq9\" (UID: \"9cc92c5c-285e-4604-bc60-7637938a78c8\") " pod="openstack-operators/heat-operator-controller-manager-5b77f656f-8pvq9" Nov 28 13:33:24 crc kubenswrapper[4631]: I1128 13:33:24.105410 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ddkv7\" (UniqueName: \"kubernetes.io/projected/59ee25ad-f169-47b9-9d5c-f7a7a253a591-kube-api-access-ddkv7\") pod \"glance-operator-controller-manager-589cbd6b5b-npdtc\" (UID: \"59ee25ad-f169-47b9-9d5c-f7a7a253a591\") " pod="openstack-operators/glance-operator-controller-manager-589cbd6b5b-npdtc" Nov 28 13:33:24 crc kubenswrapper[4631]: I1128 13:33:24.105455 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/0079ab2d-c5a8-45eb-b6e8-9574158de32f-cert\") pod \"infra-operator-controller-manager-57548d458d-h7npd\" (UID: \"0079ab2d-c5a8-45eb-b6e8-9574158de32f\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-h7npd" Nov 28 13:33:24 crc kubenswrapper[4631]: I1128 13:33:24.126251 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-6b49bfffcd-5ks5f"] Nov 28 13:33:24 crc kubenswrapper[4631]: I1128 13:33:24.132387 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-66f4dd4bc7-m9t96"] Nov 28 13:33:24 crc kubenswrapper[4631]: I1128 13:33:24.133760 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-66f4dd4bc7-m9t96" Nov 28 13:33:24 crc kubenswrapper[4631]: I1128 13:33:24.192095 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-5d499bf58b-dck2d"] Nov 28 13:33:24 crc kubenswrapper[4631]: I1128 13:33:24.193878 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vvzpr\" (UniqueName: \"kubernetes.io/projected/9cc92c5c-285e-4604-bc60-7637938a78c8-kube-api-access-vvzpr\") pod \"heat-operator-controller-manager-5b77f656f-8pvq9\" (UID: \"9cc92c5c-285e-4604-bc60-7637938a78c8\") " pod="openstack-operators/heat-operator-controller-manager-5b77f656f-8pvq9" Nov 28 13:33:24 crc kubenswrapper[4631]: I1128 13:33:24.195331 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ddkv7\" (UniqueName: \"kubernetes.io/projected/59ee25ad-f169-47b9-9d5c-f7a7a253a591-kube-api-access-ddkv7\") pod \"glance-operator-controller-manager-589cbd6b5b-npdtc\" (UID: \"59ee25ad-f169-47b9-9d5c-f7a7a253a591\") " pod="openstack-operators/glance-operator-controller-manager-589cbd6b5b-npdtc" Nov 28 13:33:24 crc kubenswrapper[4631]: I1128 13:33:24.215320 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-wwffb" Nov 28 13:33:24 crc kubenswrapper[4631]: I1128 13:33:24.235650 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/0079ab2d-c5a8-45eb-b6e8-9574158de32f-cert\") pod \"infra-operator-controller-manager-57548d458d-h7npd\" (UID: \"0079ab2d-c5a8-45eb-b6e8-9574158de32f\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-h7npd" Nov 28 13:33:24 crc kubenswrapper[4631]: E1128 13:33:24.235844 4631 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Nov 28 13:33:24 crc kubenswrapper[4631]: E1128 13:33:24.235969 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0079ab2d-c5a8-45eb-b6e8-9574158de32f-cert podName:0079ab2d-c5a8-45eb-b6e8-9574158de32f nodeName:}" failed. No retries permitted until 2025-11-28 13:33:24.735943202 +0000 UTC m=+761.543246536 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/0079ab2d-c5a8-45eb-b6e8-9574158de32f-cert") pod "infra-operator-controller-manager-57548d458d-h7npd" (UID: "0079ab2d-c5a8-45eb-b6e8-9574158de32f") : secret "infra-operator-webhook-server-cert" not found Nov 28 13:33:24 crc kubenswrapper[4631]: I1128 13:33:24.236160 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zhh2d\" (UniqueName: \"kubernetes.io/projected/aafff68d-b989-4edc-ab1c-b9d33edb8ee2-kube-api-access-zhh2d\") pod \"horizon-operator-controller-manager-5d494799bf-2mq7v\" (UID: \"aafff68d-b989-4edc-ab1c-b9d33edb8ee2\") " pod="openstack-operators/horizon-operator-controller-manager-5d494799bf-2mq7v" Nov 28 13:33:24 crc kubenswrapper[4631]: I1128 13:33:24.236195 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b44dl\" (UniqueName: \"kubernetes.io/projected/0079ab2d-c5a8-45eb-b6e8-9574158de32f-kube-api-access-b44dl\") pod \"infra-operator-controller-manager-57548d458d-h7npd\" (UID: \"0079ab2d-c5a8-45eb-b6e8-9574158de32f\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-h7npd" Nov 28 13:33:24 crc kubenswrapper[4631]: I1128 13:33:24.236222 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8pxds\" (UniqueName: \"kubernetes.io/projected/f3e12436-8a74-4443-94a4-3b9717105b05-kube-api-access-8pxds\") pod \"keystone-operator-controller-manager-6b49bfffcd-5ks5f\" (UID: \"f3e12436-8a74-4443-94a4-3b9717105b05\") " pod="openstack-operators/keystone-operator-controller-manager-6b49bfffcd-5ks5f" Nov 28 13:33:24 crc kubenswrapper[4631]: I1128 13:33:24.236245 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zn6mw\" (UniqueName: \"kubernetes.io/projected/ed17d71f-b72c-4f12-b5f7-d5cb3bb130dc-kube-api-access-zn6mw\") pod \"ironic-operator-controller-manager-67cb4dc6d4-db8kg\" (UID: \"ed17d71f-b72c-4f12-b5f7-d5cb3bb130dc\") " pod="openstack-operators/ironic-operator-controller-manager-67cb4dc6d4-db8kg" Nov 28 13:33:24 crc kubenswrapper[4631]: I1128 13:33:24.252121 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-66f4dd4bc7-m9t96"] Nov 28 13:33:24 crc kubenswrapper[4631]: I1128 13:33:24.252251 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-5d499bf58b-dck2d" Nov 28 13:33:24 crc kubenswrapper[4631]: I1128 13:33:24.255341 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-6fdcddb789-msc6n"] Nov 28 13:33:24 crc kubenswrapper[4631]: I1128 13:33:24.256425 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-6fdcddb789-msc6n" Nov 28 13:33:24 crc kubenswrapper[4631]: I1128 13:33:24.264399 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-jbzlt" Nov 28 13:33:24 crc kubenswrapper[4631]: I1128 13:33:24.267727 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-99rds" Nov 28 13:33:24 crc kubenswrapper[4631]: I1128 13:33:24.279346 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-5d499bf58b-dck2d"] Nov 28 13:33:24 crc kubenswrapper[4631]: I1128 13:33:24.279404 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-6fdcddb789-msc6n"] Nov 28 13:33:24 crc kubenswrapper[4631]: I1128 13:33:24.291485 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-79556f57fc-tpbhk"] Nov 28 13:33:24 crc kubenswrapper[4631]: I1128 13:33:24.293187 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-tpbhk" Nov 28 13:33:24 crc kubenswrapper[4631]: I1128 13:33:24.295438 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zhh2d\" (UniqueName: \"kubernetes.io/projected/aafff68d-b989-4edc-ab1c-b9d33edb8ee2-kube-api-access-zhh2d\") pod \"horizon-operator-controller-manager-5d494799bf-2mq7v\" (UID: \"aafff68d-b989-4edc-ab1c-b9d33edb8ee2\") " pod="openstack-operators/horizon-operator-controller-manager-5d494799bf-2mq7v" Nov 28 13:33:24 crc kubenswrapper[4631]: I1128 13:33:24.297931 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-s7fbj" Nov 28 13:33:24 crc kubenswrapper[4631]: I1128 13:33:24.311936 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-79556f57fc-tpbhk"] Nov 28 13:33:24 crc kubenswrapper[4631]: I1128 13:33:24.324374 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b44dl\" (UniqueName: \"kubernetes.io/projected/0079ab2d-c5a8-45eb-b6e8-9574158de32f-kube-api-access-b44dl\") pod \"infra-operator-controller-manager-57548d458d-h7npd\" (UID: \"0079ab2d-c5a8-45eb-b6e8-9574158de32f\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-h7npd" Nov 28 13:33:24 crc kubenswrapper[4631]: I1128 13:33:24.354812 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-955677c94-6bt2f" Nov 28 13:33:24 crc kubenswrapper[4631]: I1128 13:33:24.357364 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zn6mw\" (UniqueName: \"kubernetes.io/projected/ed17d71f-b72c-4f12-b5f7-d5cb3bb130dc-kube-api-access-zn6mw\") pod \"ironic-operator-controller-manager-67cb4dc6d4-db8kg\" (UID: \"ed17d71f-b72c-4f12-b5f7-d5cb3bb130dc\") " pod="openstack-operators/ironic-operator-controller-manager-67cb4dc6d4-db8kg" Nov 28 13:33:24 crc kubenswrapper[4631]: I1128 13:33:24.363188 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wfnvp\" (UniqueName: \"kubernetes.io/projected/cefc6050-9552-4dda-9864-39ac05626785-kube-api-access-wfnvp\") pod \"mariadb-operator-controller-manager-66f4dd4bc7-m9t96\" (UID: \"cefc6050-9552-4dda-9864-39ac05626785\") " pod="openstack-operators/mariadb-operator-controller-manager-66f4dd4bc7-m9t96" Nov 28 13:33:24 crc kubenswrapper[4631]: I1128 13:33:24.363232 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9w9vj\" (UniqueName: \"kubernetes.io/projected/dbbf77e7-3c97-480f-83b1-0eb8bb8f386d-kube-api-access-9w9vj\") pod \"manila-operator-controller-manager-5d499bf58b-dck2d\" (UID: \"dbbf77e7-3c97-480f-83b1-0eb8bb8f386d\") " pod="openstack-operators/manila-operator-controller-manager-5d499bf58b-dck2d" Nov 28 13:33:24 crc kubenswrapper[4631]: I1128 13:33:24.363338 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fh7fb\" (UniqueName: \"kubernetes.io/projected/bb908ace-bb62-4f55-8131-90e48da7585b-kube-api-access-fh7fb\") pod \"nova-operator-controller-manager-79556f57fc-tpbhk\" (UID: \"bb908ace-bb62-4f55-8131-90e48da7585b\") " pod="openstack-operators/nova-operator-controller-manager-79556f57fc-tpbhk" Nov 28 13:33:24 crc kubenswrapper[4631]: I1128 13:33:24.363401 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8pxds\" (UniqueName: \"kubernetes.io/projected/f3e12436-8a74-4443-94a4-3b9717105b05-kube-api-access-8pxds\") pod \"keystone-operator-controller-manager-6b49bfffcd-5ks5f\" (UID: \"f3e12436-8a74-4443-94a4-3b9717105b05\") " pod="openstack-operators/keystone-operator-controller-manager-6b49bfffcd-5ks5f" Nov 28 13:33:24 crc kubenswrapper[4631]: I1128 13:33:24.363504 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8v2vt\" (UniqueName: \"kubernetes.io/projected/2dab52af-5e82-4119-b4d5-56fb31abcb4d-kube-api-access-8v2vt\") pod \"neutron-operator-controller-manager-6fdcddb789-msc6n\" (UID: \"2dab52af-5e82-4119-b4d5-56fb31abcb4d\") " pod="openstack-operators/neutron-operator-controller-manager-6fdcddb789-msc6n" Nov 28 13:33:24 crc kubenswrapper[4631]: I1128 13:33:24.363742 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-6b7f75547b-rmlgq" Nov 28 13:33:24 crc kubenswrapper[4631]: I1128 13:33:24.420970 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-67cb4dc6d4-db8kg" Nov 28 13:33:24 crc kubenswrapper[4631]: I1128 13:33:24.448267 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-589cbd6b5b-npdtc" Nov 28 13:33:24 crc kubenswrapper[4631]: I1128 13:33:24.453268 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8pxds\" (UniqueName: \"kubernetes.io/projected/f3e12436-8a74-4443-94a4-3b9717105b05-kube-api-access-8pxds\") pod \"keystone-operator-controller-manager-6b49bfffcd-5ks5f\" (UID: \"f3e12436-8a74-4443-94a4-3b9717105b05\") " pod="openstack-operators/keystone-operator-controller-manager-6b49bfffcd-5ks5f" Nov 28 13:33:24 crc kubenswrapper[4631]: I1128 13:33:24.467122 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wfnvp\" (UniqueName: \"kubernetes.io/projected/cefc6050-9552-4dda-9864-39ac05626785-kube-api-access-wfnvp\") pod \"mariadb-operator-controller-manager-66f4dd4bc7-m9t96\" (UID: \"cefc6050-9552-4dda-9864-39ac05626785\") " pod="openstack-operators/mariadb-operator-controller-manager-66f4dd4bc7-m9t96" Nov 28 13:33:24 crc kubenswrapper[4631]: I1128 13:33:24.467178 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9w9vj\" (UniqueName: \"kubernetes.io/projected/dbbf77e7-3c97-480f-83b1-0eb8bb8f386d-kube-api-access-9w9vj\") pod \"manila-operator-controller-manager-5d499bf58b-dck2d\" (UID: \"dbbf77e7-3c97-480f-83b1-0eb8bb8f386d\") " pod="openstack-operators/manila-operator-controller-manager-5d499bf58b-dck2d" Nov 28 13:33:24 crc kubenswrapper[4631]: I1128 13:33:24.467211 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fh7fb\" (UniqueName: \"kubernetes.io/projected/bb908ace-bb62-4f55-8131-90e48da7585b-kube-api-access-fh7fb\") pod \"nova-operator-controller-manager-79556f57fc-tpbhk\" (UID: \"bb908ace-bb62-4f55-8131-90e48da7585b\") " pod="openstack-operators/nova-operator-controller-manager-79556f57fc-tpbhk" Nov 28 13:33:24 crc kubenswrapper[4631]: I1128 13:33:24.467298 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8v2vt\" (UniqueName: \"kubernetes.io/projected/2dab52af-5e82-4119-b4d5-56fb31abcb4d-kube-api-access-8v2vt\") pod \"neutron-operator-controller-manager-6fdcddb789-msc6n\" (UID: \"2dab52af-5e82-4119-b4d5-56fb31abcb4d\") " pod="openstack-operators/neutron-operator-controller-manager-6fdcddb789-msc6n" Nov 28 13:33:24 crc kubenswrapper[4631]: I1128 13:33:24.473393 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-64cdc6ff96-4qzz7"] Nov 28 13:33:24 crc kubenswrapper[4631]: I1128 13:33:24.474537 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-64cdc6ff96-4qzz7" Nov 28 13:33:24 crc kubenswrapper[4631]: I1128 13:33:24.476182 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-5b77f656f-8pvq9" Nov 28 13:33:24 crc kubenswrapper[4631]: I1128 13:33:24.496759 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-fn2wh" Nov 28 13:33:24 crc kubenswrapper[4631]: I1128 13:33:24.514666 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-5fcdb54b6b2f9vf"] Nov 28 13:33:24 crc kubenswrapper[4631]: I1128 13:33:24.525526 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5fcdb54b6b2f9vf" Nov 28 13:33:24 crc kubenswrapper[4631]: I1128 13:33:24.540401 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-64cdc6ff96-4qzz7"] Nov 28 13:33:24 crc kubenswrapper[4631]: I1128 13:33:24.548855 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-5d494799bf-2mq7v" Nov 28 13:33:24 crc kubenswrapper[4631]: I1128 13:33:24.571326 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-774wx\" (UniqueName: \"kubernetes.io/projected/ca18d816-8a63-4dcd-ac27-b1e7ad9ce4f6-kube-api-access-774wx\") pod \"openstack-baremetal-operator-controller-manager-5fcdb54b6b2f9vf\" (UID: \"ca18d816-8a63-4dcd-ac27-b1e7ad9ce4f6\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5fcdb54b6b2f9vf" Nov 28 13:33:24 crc kubenswrapper[4631]: I1128 13:33:24.571372 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-czcr5\" (UniqueName: \"kubernetes.io/projected/ecfd67ad-5136-4e6e-9e41-3d933d10ea2c-kube-api-access-czcr5\") pod \"octavia-operator-controller-manager-64cdc6ff96-4qzz7\" (UID: \"ecfd67ad-5136-4e6e-9e41-3d933d10ea2c\") " pod="openstack-operators/octavia-operator-controller-manager-64cdc6ff96-4qzz7" Nov 28 13:33:24 crc kubenswrapper[4631]: I1128 13:33:24.571401 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ca18d816-8a63-4dcd-ac27-b1e7ad9ce4f6-cert\") pod \"openstack-baremetal-operator-controller-manager-5fcdb54b6b2f9vf\" (UID: \"ca18d816-8a63-4dcd-ac27-b1e7ad9ce4f6\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5fcdb54b6b2f9vf" Nov 28 13:33:24 crc kubenswrapper[4631]: I1128 13:33:24.589493 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-5fcdb54b6b2f9vf"] Nov 28 13:33:24 crc kubenswrapper[4631]: I1128 13:33:24.616030 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-v4w8x" Nov 28 13:33:24 crc kubenswrapper[4631]: I1128 13:33:24.616518 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Nov 28 13:33:24 crc kubenswrapper[4631]: I1128 13:33:24.646303 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wfnvp\" (UniqueName: \"kubernetes.io/projected/cefc6050-9552-4dda-9864-39ac05626785-kube-api-access-wfnvp\") pod \"mariadb-operator-controller-manager-66f4dd4bc7-m9t96\" (UID: \"cefc6050-9552-4dda-9864-39ac05626785\") " pod="openstack-operators/mariadb-operator-controller-manager-66f4dd4bc7-m9t96" Nov 28 13:33:24 crc kubenswrapper[4631]: I1128 13:33:24.655487 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-56897c768d-wkpql"] Nov 28 13:33:24 crc kubenswrapper[4631]: I1128 13:33:24.658447 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8v2vt\" (UniqueName: \"kubernetes.io/projected/2dab52af-5e82-4119-b4d5-56fb31abcb4d-kube-api-access-8v2vt\") pod \"neutron-operator-controller-manager-6fdcddb789-msc6n\" (UID: \"2dab52af-5e82-4119-b4d5-56fb31abcb4d\") " pod="openstack-operators/neutron-operator-controller-manager-6fdcddb789-msc6n" Nov 28 13:33:24 crc kubenswrapper[4631]: I1128 13:33:24.659076 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fh7fb\" (UniqueName: \"kubernetes.io/projected/bb908ace-bb62-4f55-8131-90e48da7585b-kube-api-access-fh7fb\") pod \"nova-operator-controller-manager-79556f57fc-tpbhk\" (UID: \"bb908ace-bb62-4f55-8131-90e48da7585b\") " pod="openstack-operators/nova-operator-controller-manager-79556f57fc-tpbhk" Nov 28 13:33:24 crc kubenswrapper[4631]: I1128 13:33:24.660376 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-56897c768d-wkpql" Nov 28 13:33:24 crc kubenswrapper[4631]: I1128 13:33:24.668427 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9w9vj\" (UniqueName: \"kubernetes.io/projected/dbbf77e7-3c97-480f-83b1-0eb8bb8f386d-kube-api-access-9w9vj\") pod \"manila-operator-controller-manager-5d499bf58b-dck2d\" (UID: \"dbbf77e7-3c97-480f-83b1-0eb8bb8f386d\") " pod="openstack-operators/manila-operator-controller-manager-5d499bf58b-dck2d" Nov 28 13:33:24 crc kubenswrapper[4631]: I1128 13:33:24.668728 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-57988cc5b5-wqpdp"] Nov 28 13:33:24 crc kubenswrapper[4631]: I1128 13:33:24.672579 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-57988cc5b5-wqpdp" Nov 28 13:33:24 crc kubenswrapper[4631]: I1128 13:33:24.673249 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-774wx\" (UniqueName: \"kubernetes.io/projected/ca18d816-8a63-4dcd-ac27-b1e7ad9ce4f6-kube-api-access-774wx\") pod \"openstack-baremetal-operator-controller-manager-5fcdb54b6b2f9vf\" (UID: \"ca18d816-8a63-4dcd-ac27-b1e7ad9ce4f6\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5fcdb54b6b2f9vf" Nov 28 13:33:24 crc kubenswrapper[4631]: I1128 13:33:24.681255 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-999ds" Nov 28 13:33:24 crc kubenswrapper[4631]: I1128 13:33:24.683787 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-czcr5\" (UniqueName: \"kubernetes.io/projected/ecfd67ad-5136-4e6e-9e41-3d933d10ea2c-kube-api-access-czcr5\") pod \"octavia-operator-controller-manager-64cdc6ff96-4qzz7\" (UID: \"ecfd67ad-5136-4e6e-9e41-3d933d10ea2c\") " pod="openstack-operators/octavia-operator-controller-manager-64cdc6ff96-4qzz7" Nov 28 13:33:24 crc kubenswrapper[4631]: I1128 13:33:24.683864 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ca18d816-8a63-4dcd-ac27-b1e7ad9ce4f6-cert\") pod \"openstack-baremetal-operator-controller-manager-5fcdb54b6b2f9vf\" (UID: \"ca18d816-8a63-4dcd-ac27-b1e7ad9ce4f6\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5fcdb54b6b2f9vf" Nov 28 13:33:24 crc kubenswrapper[4631]: E1128 13:33:24.684112 4631 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Nov 28 13:33:24 crc kubenswrapper[4631]: E1128 13:33:24.684160 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ca18d816-8a63-4dcd-ac27-b1e7ad9ce4f6-cert podName:ca18d816-8a63-4dcd-ac27-b1e7ad9ce4f6 nodeName:}" failed. No retries permitted until 2025-11-28 13:33:25.184144648 +0000 UTC m=+761.991447992 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/ca18d816-8a63-4dcd-ac27-b1e7ad9ce4f6-cert") pod "openstack-baremetal-operator-controller-manager-5fcdb54b6b2f9vf" (UID: "ca18d816-8a63-4dcd-ac27-b1e7ad9ce4f6") : secret "openstack-baremetal-operator-webhook-server-cert" not found Nov 28 13:33:24 crc kubenswrapper[4631]: I1128 13:33:24.688915 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-56897c768d-wkpql"] Nov 28 13:33:24 crc kubenswrapper[4631]: I1128 13:33:24.690731 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-tpbhk" Nov 28 13:33:24 crc kubenswrapper[4631]: I1128 13:33:24.694966 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-t68zq" Nov 28 13:33:24 crc kubenswrapper[4631]: I1128 13:33:24.730726 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-6b49bfffcd-5ks5f" Nov 28 13:33:24 crc kubenswrapper[4631]: I1128 13:33:24.749846 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-57988cc5b5-wqpdp"] Nov 28 13:33:24 crc kubenswrapper[4631]: I1128 13:33:24.766184 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-d77b94747-2g446"] Nov 28 13:33:24 crc kubenswrapper[4631]: I1128 13:33:24.767549 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-d77b94747-2g446" Nov 28 13:33:24 crc kubenswrapper[4631]: I1128 13:33:24.787301 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/0079ab2d-c5a8-45eb-b6e8-9574158de32f-cert\") pod \"infra-operator-controller-manager-57548d458d-h7npd\" (UID: \"0079ab2d-c5a8-45eb-b6e8-9574158de32f\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-h7npd" Nov 28 13:33:24 crc kubenswrapper[4631]: I1128 13:33:24.787442 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-khlhq\" (UniqueName: \"kubernetes.io/projected/05a6dcba-27b4-4b20-9275-f434b8943be9-kube-api-access-khlhq\") pod \"placement-operator-controller-manager-57988cc5b5-wqpdp\" (UID: \"05a6dcba-27b4-4b20-9275-f434b8943be9\") " pod="openstack-operators/placement-operator-controller-manager-57988cc5b5-wqpdp" Nov 28 13:33:24 crc kubenswrapper[4631]: I1128 13:33:24.787540 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m2twv\" (UniqueName: \"kubernetes.io/projected/d09fac54-dd24-47df-83a1-23cae3102e7a-kube-api-access-m2twv\") pod \"ovn-operator-controller-manager-56897c768d-wkpql\" (UID: \"d09fac54-dd24-47df-83a1-23cae3102e7a\") " pod="openstack-operators/ovn-operator-controller-manager-56897c768d-wkpql" Nov 28 13:33:24 crc kubenswrapper[4631]: E1128 13:33:24.787770 4631 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Nov 28 13:33:24 crc kubenswrapper[4631]: E1128 13:33:24.787850 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0079ab2d-c5a8-45eb-b6e8-9574158de32f-cert podName:0079ab2d-c5a8-45eb-b6e8-9574158de32f nodeName:}" failed. No retries permitted until 2025-11-28 13:33:25.787828482 +0000 UTC m=+762.595131826 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/0079ab2d-c5a8-45eb-b6e8-9574158de32f-cert") pod "infra-operator-controller-manager-57548d458d-h7npd" (UID: "0079ab2d-c5a8-45eb-b6e8-9574158de32f") : secret "infra-operator-webhook-server-cert" not found Nov 28 13:33:24 crc kubenswrapper[4631]: I1128 13:33:24.804984 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-774wx\" (UniqueName: \"kubernetes.io/projected/ca18d816-8a63-4dcd-ac27-b1e7ad9ce4f6-kube-api-access-774wx\") pod \"openstack-baremetal-operator-controller-manager-5fcdb54b6b2f9vf\" (UID: \"ca18d816-8a63-4dcd-ac27-b1e7ad9ce4f6\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5fcdb54b6b2f9vf" Nov 28 13:33:24 crc kubenswrapper[4631]: I1128 13:33:24.805004 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-czcr5\" (UniqueName: \"kubernetes.io/projected/ecfd67ad-5136-4e6e-9e41-3d933d10ea2c-kube-api-access-czcr5\") pod \"octavia-operator-controller-manager-64cdc6ff96-4qzz7\" (UID: \"ecfd67ad-5136-4e6e-9e41-3d933d10ea2c\") " pod="openstack-operators/octavia-operator-controller-manager-64cdc6ff96-4qzz7" Nov 28 13:33:24 crc kubenswrapper[4631]: I1128 13:33:24.805445 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-r94wn"] Nov 28 13:33:24 crc kubenswrapper[4631]: I1128 13:33:24.806721 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-r94wn" Nov 28 13:33:24 crc kubenswrapper[4631]: I1128 13:33:24.829815 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-66f4dd4bc7-m9t96" Nov 28 13:33:24 crc kubenswrapper[4631]: I1128 13:33:24.832373 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-d77b94747-2g446"] Nov 28 13:33:24 crc kubenswrapper[4631]: I1128 13:33:24.832693 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-w7kq4" Nov 28 13:33:24 crc kubenswrapper[4631]: I1128 13:33:24.870698 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-pb5gw" Nov 28 13:33:24 crc kubenswrapper[4631]: I1128 13:33:24.871840 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-64cdc6ff96-4qzz7" Nov 28 13:33:24 crc kubenswrapper[4631]: I1128 13:33:24.882196 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-5d499bf58b-dck2d" Nov 28 13:33:24 crc kubenswrapper[4631]: I1128 13:33:24.883022 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-r94wn"] Nov 28 13:33:24 crc kubenswrapper[4631]: I1128 13:33:24.889114 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xhk5t\" (UniqueName: \"kubernetes.io/projected/2d59108a-2b92-421b-beb8-157c2eca09ac-kube-api-access-xhk5t\") pod \"swift-operator-controller-manager-d77b94747-2g446\" (UID: \"2d59108a-2b92-421b-beb8-157c2eca09ac\") " pod="openstack-operators/swift-operator-controller-manager-d77b94747-2g446" Nov 28 13:33:24 crc kubenswrapper[4631]: I1128 13:33:24.889162 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-khlhq\" (UniqueName: \"kubernetes.io/projected/05a6dcba-27b4-4b20-9275-f434b8943be9-kube-api-access-khlhq\") pod \"placement-operator-controller-manager-57988cc5b5-wqpdp\" (UID: \"05a6dcba-27b4-4b20-9275-f434b8943be9\") " pod="openstack-operators/placement-operator-controller-manager-57988cc5b5-wqpdp" Nov 28 13:33:24 crc kubenswrapper[4631]: I1128 13:33:24.889219 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m2twv\" (UniqueName: \"kubernetes.io/projected/d09fac54-dd24-47df-83a1-23cae3102e7a-kube-api-access-m2twv\") pod \"ovn-operator-controller-manager-56897c768d-wkpql\" (UID: \"d09fac54-dd24-47df-83a1-23cae3102e7a\") " pod="openstack-operators/ovn-operator-controller-manager-56897c768d-wkpql" Nov 28 13:33:24 crc kubenswrapper[4631]: I1128 13:33:24.967939 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-khlhq\" (UniqueName: \"kubernetes.io/projected/05a6dcba-27b4-4b20-9275-f434b8943be9-kube-api-access-khlhq\") pod \"placement-operator-controller-manager-57988cc5b5-wqpdp\" (UID: \"05a6dcba-27b4-4b20-9275-f434b8943be9\") " pod="openstack-operators/placement-operator-controller-manager-57988cc5b5-wqpdp" Nov 28 13:33:24 crc kubenswrapper[4631]: I1128 13:33:24.969518 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m2twv\" (UniqueName: \"kubernetes.io/projected/d09fac54-dd24-47df-83a1-23cae3102e7a-kube-api-access-m2twv\") pod \"ovn-operator-controller-manager-56897c768d-wkpql\" (UID: \"d09fac54-dd24-47df-83a1-23cae3102e7a\") " pod="openstack-operators/ovn-operator-controller-manager-56897c768d-wkpql" Nov 28 13:33:25 crc kubenswrapper[4631]: I1128 13:33:25.008544 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-6fdcddb789-msc6n" Nov 28 13:33:25 crc kubenswrapper[4631]: I1128 13:33:25.009852 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2whr6\" (UniqueName: \"kubernetes.io/projected/f2aea9a5-85cf-464f-8eba-382b3db8d71d-kube-api-access-2whr6\") pod \"telemetry-operator-controller-manager-76cc84c6bb-r94wn\" (UID: \"f2aea9a5-85cf-464f-8eba-382b3db8d71d\") " pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-r94wn" Nov 28 13:33:25 crc kubenswrapper[4631]: I1128 13:33:25.009914 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xhk5t\" (UniqueName: \"kubernetes.io/projected/2d59108a-2b92-421b-beb8-157c2eca09ac-kube-api-access-xhk5t\") pod \"swift-operator-controller-manager-d77b94747-2g446\" (UID: \"2d59108a-2b92-421b-beb8-157c2eca09ac\") " pod="openstack-operators/swift-operator-controller-manager-d77b94747-2g446" Nov 28 13:33:25 crc kubenswrapper[4631]: I1128 13:33:25.012499 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-56897c768d-wkpql" Nov 28 13:33:25 crc kubenswrapper[4631]: I1128 13:33:25.036700 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-57988cc5b5-wqpdp" Nov 28 13:33:25 crc kubenswrapper[4631]: I1128 13:33:25.119208 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2whr6\" (UniqueName: \"kubernetes.io/projected/f2aea9a5-85cf-464f-8eba-382b3db8d71d-kube-api-access-2whr6\") pod \"telemetry-operator-controller-manager-76cc84c6bb-r94wn\" (UID: \"f2aea9a5-85cf-464f-8eba-382b3db8d71d\") " pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-r94wn" Nov 28 13:33:25 crc kubenswrapper[4631]: I1128 13:33:25.131889 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-656dcb59d4-wh9v9"] Nov 28 13:33:25 crc kubenswrapper[4631]: I1128 13:33:25.133497 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-656dcb59d4-wh9v9" Nov 28 13:33:25 crc kubenswrapper[4631]: I1128 13:33:25.154601 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xhk5t\" (UniqueName: \"kubernetes.io/projected/2d59108a-2b92-421b-beb8-157c2eca09ac-kube-api-access-xhk5t\") pod \"swift-operator-controller-manager-d77b94747-2g446\" (UID: \"2d59108a-2b92-421b-beb8-157c2eca09ac\") " pod="openstack-operators/swift-operator-controller-manager-d77b94747-2g446" Nov 28 13:33:25 crc kubenswrapper[4631]: I1128 13:33:25.190804 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-s2d6l" Nov 28 13:33:25 crc kubenswrapper[4631]: I1128 13:33:25.223917 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-656dcb59d4-wh9v9"] Nov 28 13:33:25 crc kubenswrapper[4631]: I1128 13:33:25.251112 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s25km\" (UniqueName: \"kubernetes.io/projected/826f27ac-a938-4292-a357-ff54c4117986-kube-api-access-s25km\") pod \"watcher-operator-controller-manager-656dcb59d4-wh9v9\" (UID: \"826f27ac-a938-4292-a357-ff54c4117986\") " pod="openstack-operators/watcher-operator-controller-manager-656dcb59d4-wh9v9" Nov 28 13:33:25 crc kubenswrapper[4631]: I1128 13:33:25.251270 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ca18d816-8a63-4dcd-ac27-b1e7ad9ce4f6-cert\") pod \"openstack-baremetal-operator-controller-manager-5fcdb54b6b2f9vf\" (UID: \"ca18d816-8a63-4dcd-ac27-b1e7ad9ce4f6\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5fcdb54b6b2f9vf" Nov 28 13:33:25 crc kubenswrapper[4631]: E1128 13:33:25.251526 4631 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Nov 28 13:33:25 crc kubenswrapper[4631]: E1128 13:33:25.251941 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ca18d816-8a63-4dcd-ac27-b1e7ad9ce4f6-cert podName:ca18d816-8a63-4dcd-ac27-b1e7ad9ce4f6 nodeName:}" failed. No retries permitted until 2025-11-28 13:33:26.251895623 +0000 UTC m=+763.059198967 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/ca18d816-8a63-4dcd-ac27-b1e7ad9ce4f6-cert") pod "openstack-baremetal-operator-controller-manager-5fcdb54b6b2f9vf" (UID: "ca18d816-8a63-4dcd-ac27-b1e7ad9ce4f6") : secret "openstack-baremetal-operator-webhook-server-cert" not found Nov 28 13:33:25 crc kubenswrapper[4631]: I1128 13:33:25.284897 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7b64f4fb85-hdr72"] Nov 28 13:33:25 crc kubenswrapper[4631]: I1128 13:33:25.287076 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2whr6\" (UniqueName: \"kubernetes.io/projected/f2aea9a5-85cf-464f-8eba-382b3db8d71d-kube-api-access-2whr6\") pod \"telemetry-operator-controller-manager-76cc84c6bb-r94wn\" (UID: \"f2aea9a5-85cf-464f-8eba-382b3db8d71d\") " pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-r94wn" Nov 28 13:33:25 crc kubenswrapper[4631]: I1128 13:33:25.360039 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s25km\" (UniqueName: \"kubernetes.io/projected/826f27ac-a938-4292-a357-ff54c4117986-kube-api-access-s25km\") pod \"watcher-operator-controller-manager-656dcb59d4-wh9v9\" (UID: \"826f27ac-a938-4292-a357-ff54c4117986\") " pod="openstack-operators/watcher-operator-controller-manager-656dcb59d4-wh9v9" Nov 28 13:33:25 crc kubenswrapper[4631]: I1128 13:33:25.411224 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7b64f4fb85-hdr72" event={"ID":"00a81f80-35b9-4eb4-a508-3f104ca426be","Type":"ContainerStarted","Data":"0e13d63e51a29dbc861ac4b4daa6747d6fe3688b63da62ffa926447218e6b6d0"} Nov 28 13:33:25 crc kubenswrapper[4631]: I1128 13:33:25.415365 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-5cd6c7f4c8-jwd25"] Nov 28 13:33:25 crc kubenswrapper[4631]: I1128 13:33:25.416861 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5cd6c7f4c8-jwd25" Nov 28 13:33:25 crc kubenswrapper[4631]: I1128 13:33:25.437743 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-d77b94747-2g446" Nov 28 13:33:25 crc kubenswrapper[4631]: I1128 13:33:25.460533 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-zsg6r" Nov 28 13:33:25 crc kubenswrapper[4631]: I1128 13:33:25.472945 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-r94wn" Nov 28 13:33:25 crc kubenswrapper[4631]: I1128 13:33:25.473910 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s25km\" (UniqueName: \"kubernetes.io/projected/826f27ac-a938-4292-a357-ff54c4117986-kube-api-access-s25km\") pod \"watcher-operator-controller-manager-656dcb59d4-wh9v9\" (UID: \"826f27ac-a938-4292-a357-ff54c4117986\") " pod="openstack-operators/watcher-operator-controller-manager-656dcb59d4-wh9v9" Nov 28 13:33:25 crc kubenswrapper[4631]: I1128 13:33:25.497037 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5cd6c7f4c8-jwd25"] Nov 28 13:33:25 crc kubenswrapper[4631]: I1128 13:33:25.562719 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sx296\" (UniqueName: \"kubernetes.io/projected/fa36877c-8fc2-4074-a75d-135c2006807a-kube-api-access-sx296\") pod \"test-operator-controller-manager-5cd6c7f4c8-jwd25\" (UID: \"fa36877c-8fc2-4074-a75d-135c2006807a\") " pod="openstack-operators/test-operator-controller-manager-5cd6c7f4c8-jwd25" Nov 28 13:33:25 crc kubenswrapper[4631]: I1128 13:33:25.574245 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-656dcb59d4-wh9v9" Nov 28 13:33:25 crc kubenswrapper[4631]: I1128 13:33:25.664783 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sx296\" (UniqueName: \"kubernetes.io/projected/fa36877c-8fc2-4074-a75d-135c2006807a-kube-api-access-sx296\") pod \"test-operator-controller-manager-5cd6c7f4c8-jwd25\" (UID: \"fa36877c-8fc2-4074-a75d-135c2006807a\") " pod="openstack-operators/test-operator-controller-manager-5cd6c7f4c8-jwd25" Nov 28 13:33:25 crc kubenswrapper[4631]: I1128 13:33:25.671722 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-854b9f67c6-hfb26"] Nov 28 13:33:25 crc kubenswrapper[4631]: I1128 13:33:25.672916 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-854b9f67c6-hfb26" Nov 28 13:33:25 crc kubenswrapper[4631]: I1128 13:33:25.676957 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"metrics-server-cert" Nov 28 13:33:25 crc kubenswrapper[4631]: I1128 13:33:25.677196 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Nov 28 13:33:25 crc kubenswrapper[4631]: I1128 13:33:25.681600 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-8b4p4" Nov 28 13:33:25 crc kubenswrapper[4631]: I1128 13:33:25.726455 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sx296\" (UniqueName: \"kubernetes.io/projected/fa36877c-8fc2-4074-a75d-135c2006807a-kube-api-access-sx296\") pod \"test-operator-controller-manager-5cd6c7f4c8-jwd25\" (UID: \"fa36877c-8fc2-4074-a75d-135c2006807a\") " pod="openstack-operators/test-operator-controller-manager-5cd6c7f4c8-jwd25" Nov 28 13:33:25 crc kubenswrapper[4631]: I1128 13:33:25.756749 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-854b9f67c6-hfb26"] Nov 28 13:33:25 crc kubenswrapper[4631]: I1128 13:33:25.772323 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4bc38647-eefa-4ad6-bef5-6a7ed535b794-metrics-certs\") pod \"openstack-operator-controller-manager-854b9f67c6-hfb26\" (UID: \"4bc38647-eefa-4ad6-bef5-6a7ed535b794\") " pod="openstack-operators/openstack-operator-controller-manager-854b9f67c6-hfb26" Nov 28 13:33:25 crc kubenswrapper[4631]: I1128 13:33:25.772384 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6rw5k\" (UniqueName: \"kubernetes.io/projected/4bc38647-eefa-4ad6-bef5-6a7ed535b794-kube-api-access-6rw5k\") pod \"openstack-operator-controller-manager-854b9f67c6-hfb26\" (UID: \"4bc38647-eefa-4ad6-bef5-6a7ed535b794\") " pod="openstack-operators/openstack-operator-controller-manager-854b9f67c6-hfb26" Nov 28 13:33:25 crc kubenswrapper[4631]: I1128 13:33:25.772415 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5cd6c7f4c8-jwd25" Nov 28 13:33:25 crc kubenswrapper[4631]: I1128 13:33:25.772422 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/4bc38647-eefa-4ad6-bef5-6a7ed535b794-webhook-certs\") pod \"openstack-operator-controller-manager-854b9f67c6-hfb26\" (UID: \"4bc38647-eefa-4ad6-bef5-6a7ed535b794\") " pod="openstack-operators/openstack-operator-controller-manager-854b9f67c6-hfb26" Nov 28 13:33:25 crc kubenswrapper[4631]: I1128 13:33:25.874117 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4bc38647-eefa-4ad6-bef5-6a7ed535b794-metrics-certs\") pod \"openstack-operator-controller-manager-854b9f67c6-hfb26\" (UID: \"4bc38647-eefa-4ad6-bef5-6a7ed535b794\") " pod="openstack-operators/openstack-operator-controller-manager-854b9f67c6-hfb26" Nov 28 13:33:25 crc kubenswrapper[4631]: I1128 13:33:25.874161 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6rw5k\" (UniqueName: \"kubernetes.io/projected/4bc38647-eefa-4ad6-bef5-6a7ed535b794-kube-api-access-6rw5k\") pod \"openstack-operator-controller-manager-854b9f67c6-hfb26\" (UID: \"4bc38647-eefa-4ad6-bef5-6a7ed535b794\") " pod="openstack-operators/openstack-operator-controller-manager-854b9f67c6-hfb26" Nov 28 13:33:25 crc kubenswrapper[4631]: I1128 13:33:25.874191 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/4bc38647-eefa-4ad6-bef5-6a7ed535b794-webhook-certs\") pod \"openstack-operator-controller-manager-854b9f67c6-hfb26\" (UID: \"4bc38647-eefa-4ad6-bef5-6a7ed535b794\") " pod="openstack-operators/openstack-operator-controller-manager-854b9f67c6-hfb26" Nov 28 13:33:25 crc kubenswrapper[4631]: I1128 13:33:25.874229 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/0079ab2d-c5a8-45eb-b6e8-9574158de32f-cert\") pod \"infra-operator-controller-manager-57548d458d-h7npd\" (UID: \"0079ab2d-c5a8-45eb-b6e8-9574158de32f\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-h7npd" Nov 28 13:33:25 crc kubenswrapper[4631]: E1128 13:33:25.874442 4631 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Nov 28 13:33:25 crc kubenswrapper[4631]: E1128 13:33:25.874502 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0079ab2d-c5a8-45eb-b6e8-9574158de32f-cert podName:0079ab2d-c5a8-45eb-b6e8-9574158de32f nodeName:}" failed. No retries permitted until 2025-11-28 13:33:27.874486129 +0000 UTC m=+764.681789473 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/0079ab2d-c5a8-45eb-b6e8-9574158de32f-cert") pod "infra-operator-controller-manager-57548d458d-h7npd" (UID: "0079ab2d-c5a8-45eb-b6e8-9574158de32f") : secret "infra-operator-webhook-server-cert" not found Nov 28 13:33:25 crc kubenswrapper[4631]: E1128 13:33:25.874553 4631 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Nov 28 13:33:25 crc kubenswrapper[4631]: E1128 13:33:25.874574 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4bc38647-eefa-4ad6-bef5-6a7ed535b794-metrics-certs podName:4bc38647-eefa-4ad6-bef5-6a7ed535b794 nodeName:}" failed. No retries permitted until 2025-11-28 13:33:26.374566131 +0000 UTC m=+763.181869475 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/4bc38647-eefa-4ad6-bef5-6a7ed535b794-metrics-certs") pod "openstack-operator-controller-manager-854b9f67c6-hfb26" (UID: "4bc38647-eefa-4ad6-bef5-6a7ed535b794") : secret "metrics-server-cert" not found Nov 28 13:33:25 crc kubenswrapper[4631]: E1128 13:33:25.875551 4631 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Nov 28 13:33:25 crc kubenswrapper[4631]: E1128 13:33:25.875665 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4bc38647-eefa-4ad6-bef5-6a7ed535b794-webhook-certs podName:4bc38647-eefa-4ad6-bef5-6a7ed535b794 nodeName:}" failed. No retries permitted until 2025-11-28 13:33:26.375638648 +0000 UTC m=+763.182942062 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/4bc38647-eefa-4ad6-bef5-6a7ed535b794-webhook-certs") pod "openstack-operator-controller-manager-854b9f67c6-hfb26" (UID: "4bc38647-eefa-4ad6-bef5-6a7ed535b794") : secret "webhook-server-cert" not found Nov 28 13:33:25 crc kubenswrapper[4631]: I1128 13:33:25.930443 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-nrpfx"] Nov 28 13:33:25 crc kubenswrapper[4631]: I1128 13:33:25.931633 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-nrpfx" Nov 28 13:33:25 crc kubenswrapper[4631]: I1128 13:33:25.934441 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6rw5k\" (UniqueName: \"kubernetes.io/projected/4bc38647-eefa-4ad6-bef5-6a7ed535b794-kube-api-access-6rw5k\") pod \"openstack-operator-controller-manager-854b9f67c6-hfb26\" (UID: \"4bc38647-eefa-4ad6-bef5-6a7ed535b794\") " pod="openstack-operators/openstack-operator-controller-manager-854b9f67c6-hfb26" Nov 28 13:33:25 crc kubenswrapper[4631]: I1128 13:33:25.946109 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-rb2zm" Nov 28 13:33:25 crc kubenswrapper[4631]: I1128 13:33:25.974989 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-nrpfx"] Nov 28 13:33:26 crc kubenswrapper[4631]: I1128 13:33:26.087146 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-llbd2\" (UniqueName: \"kubernetes.io/projected/10a95153-0898-4016-92a6-d56633555348-kube-api-access-llbd2\") pod \"rabbitmq-cluster-operator-manager-668c99d594-nrpfx\" (UID: \"10a95153-0898-4016-92a6-d56633555348\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-nrpfx" Nov 28 13:33:26 crc kubenswrapper[4631]: I1128 13:33:26.125754 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-67cb4dc6d4-db8kg"] Nov 28 13:33:26 crc kubenswrapper[4631]: I1128 13:33:26.165777 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-6b7f75547b-rmlgq"] Nov 28 13:33:26 crc kubenswrapper[4631]: I1128 13:33:26.189225 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-llbd2\" (UniqueName: \"kubernetes.io/projected/10a95153-0898-4016-92a6-d56633555348-kube-api-access-llbd2\") pod \"rabbitmq-cluster-operator-manager-668c99d594-nrpfx\" (UID: \"10a95153-0898-4016-92a6-d56633555348\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-nrpfx" Nov 28 13:33:26 crc kubenswrapper[4631]: I1128 13:33:26.296621 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ca18d816-8a63-4dcd-ac27-b1e7ad9ce4f6-cert\") pod \"openstack-baremetal-operator-controller-manager-5fcdb54b6b2f9vf\" (UID: \"ca18d816-8a63-4dcd-ac27-b1e7ad9ce4f6\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5fcdb54b6b2f9vf" Nov 28 13:33:26 crc kubenswrapper[4631]: E1128 13:33:26.296798 4631 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Nov 28 13:33:26 crc kubenswrapper[4631]: E1128 13:33:26.297318 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ca18d816-8a63-4dcd-ac27-b1e7ad9ce4f6-cert podName:ca18d816-8a63-4dcd-ac27-b1e7ad9ce4f6 nodeName:}" failed. No retries permitted until 2025-11-28 13:33:28.297251584 +0000 UTC m=+765.104554928 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/ca18d816-8a63-4dcd-ac27-b1e7ad9ce4f6-cert") pod "openstack-baremetal-operator-controller-manager-5fcdb54b6b2f9vf" (UID: "ca18d816-8a63-4dcd-ac27-b1e7ad9ce4f6") : secret "openstack-baremetal-operator-webhook-server-cert" not found Nov 28 13:33:26 crc kubenswrapper[4631]: I1128 13:33:26.349962 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-llbd2\" (UniqueName: \"kubernetes.io/projected/10a95153-0898-4016-92a6-d56633555348-kube-api-access-llbd2\") pod \"rabbitmq-cluster-operator-manager-668c99d594-nrpfx\" (UID: \"10a95153-0898-4016-92a6-d56633555348\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-nrpfx" Nov 28 13:33:26 crc kubenswrapper[4631]: I1128 13:33:26.374709 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-5b77f656f-8pvq9"] Nov 28 13:33:26 crc kubenswrapper[4631]: I1128 13:33:26.398927 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4bc38647-eefa-4ad6-bef5-6a7ed535b794-metrics-certs\") pod \"openstack-operator-controller-manager-854b9f67c6-hfb26\" (UID: \"4bc38647-eefa-4ad6-bef5-6a7ed535b794\") " pod="openstack-operators/openstack-operator-controller-manager-854b9f67c6-hfb26" Nov 28 13:33:26 crc kubenswrapper[4631]: I1128 13:33:26.398978 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/4bc38647-eefa-4ad6-bef5-6a7ed535b794-webhook-certs\") pod \"openstack-operator-controller-manager-854b9f67c6-hfb26\" (UID: \"4bc38647-eefa-4ad6-bef5-6a7ed535b794\") " pod="openstack-operators/openstack-operator-controller-manager-854b9f67c6-hfb26" Nov 28 13:33:26 crc kubenswrapper[4631]: E1128 13:33:26.399175 4631 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Nov 28 13:33:26 crc kubenswrapper[4631]: E1128 13:33:26.399230 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4bc38647-eefa-4ad6-bef5-6a7ed535b794-webhook-certs podName:4bc38647-eefa-4ad6-bef5-6a7ed535b794 nodeName:}" failed. No retries permitted until 2025-11-28 13:33:27.399214525 +0000 UTC m=+764.206517869 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/4bc38647-eefa-4ad6-bef5-6a7ed535b794-webhook-certs") pod "openstack-operator-controller-manager-854b9f67c6-hfb26" (UID: "4bc38647-eefa-4ad6-bef5-6a7ed535b794") : secret "webhook-server-cert" not found Nov 28 13:33:26 crc kubenswrapper[4631]: E1128 13:33:26.399675 4631 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Nov 28 13:33:26 crc kubenswrapper[4631]: E1128 13:33:26.399703 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4bc38647-eefa-4ad6-bef5-6a7ed535b794-metrics-certs podName:4bc38647-eefa-4ad6-bef5-6a7ed535b794 nodeName:}" failed. No retries permitted until 2025-11-28 13:33:27.399695277 +0000 UTC m=+764.206998621 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/4bc38647-eefa-4ad6-bef5-6a7ed535b794-metrics-certs") pod "openstack-operator-controller-manager-854b9f67c6-hfb26" (UID: "4bc38647-eefa-4ad6-bef5-6a7ed535b794") : secret "metrics-server-cert" not found Nov 28 13:33:26 crc kubenswrapper[4631]: I1128 13:33:26.438871 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-6b7f75547b-rmlgq" event={"ID":"8028d28d-2414-4ffe-9c5d-4854f77e980f","Type":"ContainerStarted","Data":"2aef427c43f14987110691729ab4c93fc5216af245f8902511090ff6f55c08a9"} Nov 28 13:33:26 crc kubenswrapper[4631]: I1128 13:33:26.444879 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5b77f656f-8pvq9" event={"ID":"9cc92c5c-285e-4604-bc60-7637938a78c8","Type":"ContainerStarted","Data":"be9535cb6ff82f8e373050c4e4b15cf4e6b92f661312c4c9a8f4d7ddacb0999a"} Nov 28 13:33:26 crc kubenswrapper[4631]: I1128 13:33:26.453785 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-67cb4dc6d4-db8kg" event={"ID":"ed17d71f-b72c-4f12-b5f7-d5cb3bb130dc","Type":"ContainerStarted","Data":"6bf434543ba24a3be982dee66ecb00f25e55f323b3b517089c280bfa2ace0573"} Nov 28 13:33:26 crc kubenswrapper[4631]: I1128 13:33:26.599013 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-nrpfx" Nov 28 13:33:26 crc kubenswrapper[4631]: I1128 13:33:26.663492 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-955677c94-6bt2f"] Nov 28 13:33:26 crc kubenswrapper[4631]: I1128 13:33:26.691592 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-79556f57fc-tpbhk"] Nov 28 13:33:26 crc kubenswrapper[4631]: I1128 13:33:26.777500 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-589cbd6b5b-npdtc"] Nov 28 13:33:26 crc kubenswrapper[4631]: I1128 13:33:26.884257 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-5d499bf58b-dck2d"] Nov 28 13:33:26 crc kubenswrapper[4631]: W1128 13:33:26.895672 4631 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddbbf77e7_3c97_480f_83b1_0eb8bb8f386d.slice/crio-41fd316a8e4d5c301729faa1dc168824ea04368449a56ff9222168159b94a532 WatchSource:0}: Error finding container 41fd316a8e4d5c301729faa1dc168824ea04368449a56ff9222168159b94a532: Status 404 returned error can't find the container with id 41fd316a8e4d5c301729faa1dc168824ea04368449a56ff9222168159b94a532 Nov 28 13:33:26 crc kubenswrapper[4631]: I1128 13:33:26.919253 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-6b49bfffcd-5ks5f"] Nov 28 13:33:26 crc kubenswrapper[4631]: I1128 13:33:26.959824 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-66f4dd4bc7-m9t96"] Nov 28 13:33:26 crc kubenswrapper[4631]: W1128 13:33:26.962549 4631 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcefc6050_9552_4dda_9864_39ac05626785.slice/crio-16cc15c32a5060dfdf2f5fc11e4015a88c6943817ee42674fc8f286b36d185ec WatchSource:0}: Error finding container 16cc15c32a5060dfdf2f5fc11e4015a88c6943817ee42674fc8f286b36d185ec: Status 404 returned error can't find the container with id 16cc15c32a5060dfdf2f5fc11e4015a88c6943817ee42674fc8f286b36d185ec Nov 28 13:33:27 crc kubenswrapper[4631]: I1128 13:33:27.100565 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-5d494799bf-2mq7v"] Nov 28 13:33:27 crc kubenswrapper[4631]: I1128 13:33:27.170203 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-656dcb59d4-wh9v9"] Nov 28 13:33:27 crc kubenswrapper[4631]: I1128 13:33:27.195947 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-57988cc5b5-wqpdp"] Nov 28 13:33:27 crc kubenswrapper[4631]: W1128 13:33:27.197500 4631 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod05a6dcba_27b4_4b20_9275_f434b8943be9.slice/crio-5ae2f72e60b7c3b6f9aa4f1b96552b002b875a26e1cf40cddddc79792dfe418a WatchSource:0}: Error finding container 5ae2f72e60b7c3b6f9aa4f1b96552b002b875a26e1cf40cddddc79792dfe418a: Status 404 returned error can't find the container with id 5ae2f72e60b7c3b6f9aa4f1b96552b002b875a26e1cf40cddddc79792dfe418a Nov 28 13:33:27 crc kubenswrapper[4631]: I1128 13:33:27.231075 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-64cdc6ff96-4qzz7"] Nov 28 13:33:27 crc kubenswrapper[4631]: W1128 13:33:27.234841 4631 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podecfd67ad_5136_4e6e_9e41_3d933d10ea2c.slice/crio-4be88051cc9ec68fb0bf0b228d22e0dcc7e6506d72d24136900fd3c70e737680 WatchSource:0}: Error finding container 4be88051cc9ec68fb0bf0b228d22e0dcc7e6506d72d24136900fd3c70e737680: Status 404 returned error can't find the container with id 4be88051cc9ec68fb0bf0b228d22e0dcc7e6506d72d24136900fd3c70e737680 Nov 28 13:33:27 crc kubenswrapper[4631]: I1128 13:33:27.348564 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5cd6c7f4c8-jwd25"] Nov 28 13:33:27 crc kubenswrapper[4631]: W1128 13:33:27.359551 4631 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd09fac54_dd24_47df_83a1_23cae3102e7a.slice/crio-02c41ec53ed599861add2af60a2a6dc216f4b55e0fe6d13370212273d71bf4b3 WatchSource:0}: Error finding container 02c41ec53ed599861add2af60a2a6dc216f4b55e0fe6d13370212273d71bf4b3: Status 404 returned error can't find the container with id 02c41ec53ed599861add2af60a2a6dc216f4b55e0fe6d13370212273d71bf4b3 Nov 28 13:33:27 crc kubenswrapper[4631]: W1128 13:33:27.362163 4631 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2dab52af_5e82_4119_b4d5_56fb31abcb4d.slice/crio-edbfc790bf4a0318884cc93fa6333c0550867b33c5b29833c017a2f310122eed WatchSource:0}: Error finding container edbfc790bf4a0318884cc93fa6333c0550867b33c5b29833c017a2f310122eed: Status 404 returned error can't find the container with id edbfc790bf4a0318884cc93fa6333c0550867b33c5b29833c017a2f310122eed Nov 28 13:33:27 crc kubenswrapper[4631]: W1128 13:33:27.372990 4631 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfa36877c_8fc2_4074_a75d_135c2006807a.slice/crio-7bdd7f0e0ffd83054c54ef5524be037201be7a603fbede5cf3c4058b980d5133 WatchSource:0}: Error finding container 7bdd7f0e0ffd83054c54ef5524be037201be7a603fbede5cf3c4058b980d5133: Status 404 returned error can't find the container with id 7bdd7f0e0ffd83054c54ef5524be037201be7a603fbede5cf3c4058b980d5133 Nov 28 13:33:27 crc kubenswrapper[4631]: I1128 13:33:27.376823 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-6fdcddb789-msc6n"] Nov 28 13:33:27 crc kubenswrapper[4631]: E1128 13:33:27.379216 4631 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/swift-operator@sha256:72236301580ff9080f7e311b832d7ba66666a9afeda51f969745229624ff26e4,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-xhk5t,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-d77b94747-2g446_openstack-operators(2d59108a-2b92-421b-beb8-157c2eca09ac): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 28 13:33:27 crc kubenswrapper[4631]: E1128 13:33:27.385541 4631 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-xhk5t,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-d77b94747-2g446_openstack-operators(2d59108a-2b92-421b-beb8-157c2eca09ac): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 28 13:33:27 crc kubenswrapper[4631]: E1128 13:33:27.386802 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/swift-operator-controller-manager-d77b94747-2g446" podUID="2d59108a-2b92-421b-beb8-157c2eca09ac" Nov 28 13:33:27 crc kubenswrapper[4631]: E1128 13:33:27.389385 4631 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/test-operator@sha256:210517b918e30df1c95fc7d961c8e57e9a9d1cc2b9fe7eb4dad2034dd53a90aa,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-sx296,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-5cd6c7f4c8-jwd25_openstack-operators(fa36877c-8fc2-4074-a75d-135c2006807a): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 28 13:33:27 crc kubenswrapper[4631]: W1128 13:33:27.389697 4631 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf2aea9a5_85cf_464f_8eba_382b3db8d71d.slice/crio-bc092d72a41917672567656c92c9f21ee7efec7fa7ec4c2dd693f556b1f3faae WatchSource:0}: Error finding container bc092d72a41917672567656c92c9f21ee7efec7fa7ec4c2dd693f556b1f3faae: Status 404 returned error can't find the container with id bc092d72a41917672567656c92c9f21ee7efec7fa7ec4c2dd693f556b1f3faae Nov 28 13:33:27 crc kubenswrapper[4631]: E1128 13:33:27.391642 4631 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-sx296,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-5cd6c7f4c8-jwd25_openstack-operators(fa36877c-8fc2-4074-a75d-135c2006807a): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 28 13:33:27 crc kubenswrapper[4631]: E1128 13:33:27.392851 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/test-operator-controller-manager-5cd6c7f4c8-jwd25" podUID="fa36877c-8fc2-4074-a75d-135c2006807a" Nov 28 13:33:27 crc kubenswrapper[4631]: I1128 13:33:27.394853 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-56897c768d-wkpql"] Nov 28 13:33:27 crc kubenswrapper[4631]: E1128 13:33:27.398727 4631 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/telemetry-operator@sha256:7d66757c0af67104f0389e851a7cc0daa44443ad202d157417bd86bbb57cc385,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-2whr6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-76cc84c6bb-r94wn_openstack-operators(f2aea9a5-85cf-464f-8eba-382b3db8d71d): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 28 13:33:27 crc kubenswrapper[4631]: E1128 13:33:27.402347 4631 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-2whr6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-76cc84c6bb-r94wn_openstack-operators(f2aea9a5-85cf-464f-8eba-382b3db8d71d): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 28 13:33:27 crc kubenswrapper[4631]: E1128 13:33:27.403436 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-r94wn" podUID="f2aea9a5-85cf-464f-8eba-382b3db8d71d" Nov 28 13:33:27 crc kubenswrapper[4631]: I1128 13:33:27.406982 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-r94wn"] Nov 28 13:33:27 crc kubenswrapper[4631]: I1128 13:33:27.416098 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-d77b94747-2g446"] Nov 28 13:33:27 crc kubenswrapper[4631]: W1128 13:33:27.419704 4631 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod10a95153_0898_4016_92a6_d56633555348.slice/crio-bb9a930ac64d36590036a5dd8865b8a001058acf2c62c9e246d678716ee9dc34 WatchSource:0}: Error finding container bb9a930ac64d36590036a5dd8865b8a001058acf2c62c9e246d678716ee9dc34: Status 404 returned error can't find the container with id bb9a930ac64d36590036a5dd8865b8a001058acf2c62c9e246d678716ee9dc34 Nov 28 13:33:27 crc kubenswrapper[4631]: E1128 13:33:27.422565 4631 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-llbd2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-668c99d594-nrpfx_openstack-operators(10a95153-0898-4016-92a6-d56633555348): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 28 13:33:27 crc kubenswrapper[4631]: I1128 13:33:27.422936 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-nrpfx"] Nov 28 13:33:27 crc kubenswrapper[4631]: E1128 13:33:27.424353 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-nrpfx" podUID="10a95153-0898-4016-92a6-d56633555348" Nov 28 13:33:27 crc kubenswrapper[4631]: I1128 13:33:27.459321 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4bc38647-eefa-4ad6-bef5-6a7ed535b794-metrics-certs\") pod \"openstack-operator-controller-manager-854b9f67c6-hfb26\" (UID: \"4bc38647-eefa-4ad6-bef5-6a7ed535b794\") " pod="openstack-operators/openstack-operator-controller-manager-854b9f67c6-hfb26" Nov 28 13:33:27 crc kubenswrapper[4631]: I1128 13:33:27.459381 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/4bc38647-eefa-4ad6-bef5-6a7ed535b794-webhook-certs\") pod \"openstack-operator-controller-manager-854b9f67c6-hfb26\" (UID: \"4bc38647-eefa-4ad6-bef5-6a7ed535b794\") " pod="openstack-operators/openstack-operator-controller-manager-854b9f67c6-hfb26" Nov 28 13:33:27 crc kubenswrapper[4631]: E1128 13:33:27.459574 4631 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Nov 28 13:33:27 crc kubenswrapper[4631]: E1128 13:33:27.459664 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4bc38647-eefa-4ad6-bef5-6a7ed535b794-metrics-certs podName:4bc38647-eefa-4ad6-bef5-6a7ed535b794 nodeName:}" failed. No retries permitted until 2025-11-28 13:33:29.459643821 +0000 UTC m=+766.266947165 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/4bc38647-eefa-4ad6-bef5-6a7ed535b794-metrics-certs") pod "openstack-operator-controller-manager-854b9f67c6-hfb26" (UID: "4bc38647-eefa-4ad6-bef5-6a7ed535b794") : secret "metrics-server-cert" not found Nov 28 13:33:27 crc kubenswrapper[4631]: E1128 13:33:27.459589 4631 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Nov 28 13:33:27 crc kubenswrapper[4631]: E1128 13:33:27.459700 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4bc38647-eefa-4ad6-bef5-6a7ed535b794-webhook-certs podName:4bc38647-eefa-4ad6-bef5-6a7ed535b794 nodeName:}" failed. No retries permitted until 2025-11-28 13:33:29.459694112 +0000 UTC m=+766.266997456 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/4bc38647-eefa-4ad6-bef5-6a7ed535b794-webhook-certs") pod "openstack-operator-controller-manager-854b9f67c6-hfb26" (UID: "4bc38647-eefa-4ad6-bef5-6a7ed535b794") : secret "webhook-server-cert" not found Nov 28 13:33:27 crc kubenswrapper[4631]: I1128 13:33:27.464560 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-589cbd6b5b-npdtc" event={"ID":"59ee25ad-f169-47b9-9d5c-f7a7a253a591","Type":"ContainerStarted","Data":"01876fec49b707fcdf2fd7fc750a79241e0a2127bacc1d1ee5efad9c57898169"} Nov 28 13:33:27 crc kubenswrapper[4631]: I1128 13:33:27.467858 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-6fdcddb789-msc6n" event={"ID":"2dab52af-5e82-4119-b4d5-56fb31abcb4d","Type":"ContainerStarted","Data":"edbfc790bf4a0318884cc93fa6333c0550867b33c5b29833c017a2f310122eed"} Nov 28 13:33:27 crc kubenswrapper[4631]: I1128 13:33:27.469591 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-66f4dd4bc7-m9t96" event={"ID":"cefc6050-9552-4dda-9864-39ac05626785","Type":"ContainerStarted","Data":"16cc15c32a5060dfdf2f5fc11e4015a88c6943817ee42674fc8f286b36d185ec"} Nov 28 13:33:27 crc kubenswrapper[4631]: I1128 13:33:27.471729 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-656dcb59d4-wh9v9" event={"ID":"826f27ac-a938-4292-a357-ff54c4117986","Type":"ContainerStarted","Data":"a10db68e2486d206b3b236ddfd0f565d06ddf59e2de12a662d9f645a07712bf9"} Nov 28 13:33:27 crc kubenswrapper[4631]: I1128 13:33:27.474471 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-955677c94-6bt2f" event={"ID":"06994d92-3dcb-461c-83ac-c7e65d4bd640","Type":"ContainerStarted","Data":"c37b22ad3a90e0b8f6e86c02840824a760e063695058f594d49809ba80db32ba"} Nov 28 13:33:27 crc kubenswrapper[4631]: I1128 13:33:27.475883 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-56897c768d-wkpql" event={"ID":"d09fac54-dd24-47df-83a1-23cae3102e7a","Type":"ContainerStarted","Data":"02c41ec53ed599861add2af60a2a6dc216f4b55e0fe6d13370212273d71bf4b3"} Nov 28 13:33:27 crc kubenswrapper[4631]: I1128 13:33:27.477297 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-5d494799bf-2mq7v" event={"ID":"aafff68d-b989-4edc-ab1c-b9d33edb8ee2","Type":"ContainerStarted","Data":"e271603c56422e913b188002ddc457cfd1620b8be8d8ad2b29499c763fe0e334"} Nov 28 13:33:27 crc kubenswrapper[4631]: I1128 13:33:27.478713 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-r94wn" event={"ID":"f2aea9a5-85cf-464f-8eba-382b3db8d71d","Type":"ContainerStarted","Data":"bc092d72a41917672567656c92c9f21ee7efec7fa7ec4c2dd693f556b1f3faae"} Nov 28 13:33:27 crc kubenswrapper[4631]: I1128 13:33:27.482720 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-d77b94747-2g446" event={"ID":"2d59108a-2b92-421b-beb8-157c2eca09ac","Type":"ContainerStarted","Data":"a406c5b18f17a1e8ead0a61c5cc994ff313de652d597cb191d0eaa4be242c040"} Nov 28 13:33:27 crc kubenswrapper[4631]: E1128 13:33:27.483152 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/telemetry-operator@sha256:7d66757c0af67104f0389e851a7cc0daa44443ad202d157417bd86bbb57cc385\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-r94wn" podUID="f2aea9a5-85cf-464f-8eba-382b3db8d71d" Nov 28 13:33:27 crc kubenswrapper[4631]: I1128 13:33:27.486996 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-tpbhk" event={"ID":"bb908ace-bb62-4f55-8131-90e48da7585b","Type":"ContainerStarted","Data":"e62cc05763d3dd20a6298af431bd4ff6568633e2a218fa1f31f5455119bcf6d6"} Nov 28 13:33:27 crc kubenswrapper[4631]: E1128 13:33:27.487748 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:72236301580ff9080f7e311b832d7ba66666a9afeda51f969745229624ff26e4\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/swift-operator-controller-manager-d77b94747-2g446" podUID="2d59108a-2b92-421b-beb8-157c2eca09ac" Nov 28 13:33:27 crc kubenswrapper[4631]: I1128 13:33:27.492899 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-57988cc5b5-wqpdp" event={"ID":"05a6dcba-27b4-4b20-9275-f434b8943be9","Type":"ContainerStarted","Data":"5ae2f72e60b7c3b6f9aa4f1b96552b002b875a26e1cf40cddddc79792dfe418a"} Nov 28 13:33:27 crc kubenswrapper[4631]: I1128 13:33:27.501992 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-5d499bf58b-dck2d" event={"ID":"dbbf77e7-3c97-480f-83b1-0eb8bb8f386d","Type":"ContainerStarted","Data":"41fd316a8e4d5c301729faa1dc168824ea04368449a56ff9222168159b94a532"} Nov 28 13:33:27 crc kubenswrapper[4631]: I1128 13:33:27.510778 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-6b49bfffcd-5ks5f" event={"ID":"f3e12436-8a74-4443-94a4-3b9717105b05","Type":"ContainerStarted","Data":"3e36344e376a5bd9efde6cc0d57165c8f7ef05211db746883df6cfca3d2db265"} Nov 28 13:33:27 crc kubenswrapper[4631]: I1128 13:33:27.528631 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-64cdc6ff96-4qzz7" event={"ID":"ecfd67ad-5136-4e6e-9e41-3d933d10ea2c","Type":"ContainerStarted","Data":"4be88051cc9ec68fb0bf0b228d22e0dcc7e6506d72d24136900fd3c70e737680"} Nov 28 13:33:27 crc kubenswrapper[4631]: I1128 13:33:27.542548 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-nrpfx" event={"ID":"10a95153-0898-4016-92a6-d56633555348","Type":"ContainerStarted","Data":"bb9a930ac64d36590036a5dd8865b8a001058acf2c62c9e246d678716ee9dc34"} Nov 28 13:33:27 crc kubenswrapper[4631]: E1128 13:33:27.545145 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-nrpfx" podUID="10a95153-0898-4016-92a6-d56633555348" Nov 28 13:33:27 crc kubenswrapper[4631]: I1128 13:33:27.548960 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5cd6c7f4c8-jwd25" event={"ID":"fa36877c-8fc2-4074-a75d-135c2006807a","Type":"ContainerStarted","Data":"7bdd7f0e0ffd83054c54ef5524be037201be7a603fbede5cf3c4058b980d5133"} Nov 28 13:33:27 crc kubenswrapper[4631]: E1128 13:33:27.556507 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:210517b918e30df1c95fc7d961c8e57e9a9d1cc2b9fe7eb4dad2034dd53a90aa\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/test-operator-controller-manager-5cd6c7f4c8-jwd25" podUID="fa36877c-8fc2-4074-a75d-135c2006807a" Nov 28 13:33:27 crc kubenswrapper[4631]: I1128 13:33:27.971105 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/0079ab2d-c5a8-45eb-b6e8-9574158de32f-cert\") pod \"infra-operator-controller-manager-57548d458d-h7npd\" (UID: \"0079ab2d-c5a8-45eb-b6e8-9574158de32f\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-h7npd" Nov 28 13:33:27 crc kubenswrapper[4631]: E1128 13:33:27.971319 4631 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Nov 28 13:33:27 crc kubenswrapper[4631]: E1128 13:33:27.971421 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0079ab2d-c5a8-45eb-b6e8-9574158de32f-cert podName:0079ab2d-c5a8-45eb-b6e8-9574158de32f nodeName:}" failed. No retries permitted until 2025-11-28 13:33:31.971392486 +0000 UTC m=+768.778695830 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/0079ab2d-c5a8-45eb-b6e8-9574158de32f-cert") pod "infra-operator-controller-manager-57548d458d-h7npd" (UID: "0079ab2d-c5a8-45eb-b6e8-9574158de32f") : secret "infra-operator-webhook-server-cert" not found Nov 28 13:33:28 crc kubenswrapper[4631]: I1128 13:33:28.380824 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ca18d816-8a63-4dcd-ac27-b1e7ad9ce4f6-cert\") pod \"openstack-baremetal-operator-controller-manager-5fcdb54b6b2f9vf\" (UID: \"ca18d816-8a63-4dcd-ac27-b1e7ad9ce4f6\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5fcdb54b6b2f9vf" Nov 28 13:33:28 crc kubenswrapper[4631]: E1128 13:33:28.381072 4631 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Nov 28 13:33:28 crc kubenswrapper[4631]: E1128 13:33:28.381169 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ca18d816-8a63-4dcd-ac27-b1e7ad9ce4f6-cert podName:ca18d816-8a63-4dcd-ac27-b1e7ad9ce4f6 nodeName:}" failed. No retries permitted until 2025-11-28 13:33:32.381148048 +0000 UTC m=+769.188451392 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/ca18d816-8a63-4dcd-ac27-b1e7ad9ce4f6-cert") pod "openstack-baremetal-operator-controller-manager-5fcdb54b6b2f9vf" (UID: "ca18d816-8a63-4dcd-ac27-b1e7ad9ce4f6") : secret "openstack-baremetal-operator-webhook-server-cert" not found Nov 28 13:33:28 crc kubenswrapper[4631]: E1128 13:33:28.558278 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-nrpfx" podUID="10a95153-0898-4016-92a6-d56633555348" Nov 28 13:33:28 crc kubenswrapper[4631]: E1128 13:33:28.558748 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:72236301580ff9080f7e311b832d7ba66666a9afeda51f969745229624ff26e4\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/swift-operator-controller-manager-d77b94747-2g446" podUID="2d59108a-2b92-421b-beb8-157c2eca09ac" Nov 28 13:33:28 crc kubenswrapper[4631]: E1128 13:33:28.559575 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/telemetry-operator@sha256:7d66757c0af67104f0389e851a7cc0daa44443ad202d157417bd86bbb57cc385\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-r94wn" podUID="f2aea9a5-85cf-464f-8eba-382b3db8d71d" Nov 28 13:33:28 crc kubenswrapper[4631]: E1128 13:33:28.564132 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:210517b918e30df1c95fc7d961c8e57e9a9d1cc2b9fe7eb4dad2034dd53a90aa\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/test-operator-controller-manager-5cd6c7f4c8-jwd25" podUID="fa36877c-8fc2-4074-a75d-135c2006807a" Nov 28 13:33:29 crc kubenswrapper[4631]: I1128 13:33:29.499024 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4bc38647-eefa-4ad6-bef5-6a7ed535b794-metrics-certs\") pod \"openstack-operator-controller-manager-854b9f67c6-hfb26\" (UID: \"4bc38647-eefa-4ad6-bef5-6a7ed535b794\") " pod="openstack-operators/openstack-operator-controller-manager-854b9f67c6-hfb26" Nov 28 13:33:29 crc kubenswrapper[4631]: I1128 13:33:29.499099 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/4bc38647-eefa-4ad6-bef5-6a7ed535b794-webhook-certs\") pod \"openstack-operator-controller-manager-854b9f67c6-hfb26\" (UID: \"4bc38647-eefa-4ad6-bef5-6a7ed535b794\") " pod="openstack-operators/openstack-operator-controller-manager-854b9f67c6-hfb26" Nov 28 13:33:29 crc kubenswrapper[4631]: E1128 13:33:29.499278 4631 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Nov 28 13:33:29 crc kubenswrapper[4631]: E1128 13:33:29.499305 4631 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Nov 28 13:33:29 crc kubenswrapper[4631]: E1128 13:33:29.499360 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4bc38647-eefa-4ad6-bef5-6a7ed535b794-webhook-certs podName:4bc38647-eefa-4ad6-bef5-6a7ed535b794 nodeName:}" failed. No retries permitted until 2025-11-28 13:33:33.499342438 +0000 UTC m=+770.306645782 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/4bc38647-eefa-4ad6-bef5-6a7ed535b794-webhook-certs") pod "openstack-operator-controller-manager-854b9f67c6-hfb26" (UID: "4bc38647-eefa-4ad6-bef5-6a7ed535b794") : secret "webhook-server-cert" not found Nov 28 13:33:29 crc kubenswrapper[4631]: E1128 13:33:29.499402 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4bc38647-eefa-4ad6-bef5-6a7ed535b794-metrics-certs podName:4bc38647-eefa-4ad6-bef5-6a7ed535b794 nodeName:}" failed. No retries permitted until 2025-11-28 13:33:33.499380739 +0000 UTC m=+770.306684073 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/4bc38647-eefa-4ad6-bef5-6a7ed535b794-metrics-certs") pod "openstack-operator-controller-manager-854b9f67c6-hfb26" (UID: "4bc38647-eefa-4ad6-bef5-6a7ed535b794") : secret "metrics-server-cert" not found Nov 28 13:33:32 crc kubenswrapper[4631]: I1128 13:33:32.056720 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/0079ab2d-c5a8-45eb-b6e8-9574158de32f-cert\") pod \"infra-operator-controller-manager-57548d458d-h7npd\" (UID: \"0079ab2d-c5a8-45eb-b6e8-9574158de32f\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-h7npd" Nov 28 13:33:32 crc kubenswrapper[4631]: E1128 13:33:32.057225 4631 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Nov 28 13:33:32 crc kubenswrapper[4631]: E1128 13:33:32.057369 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0079ab2d-c5a8-45eb-b6e8-9574158de32f-cert podName:0079ab2d-c5a8-45eb-b6e8-9574158de32f nodeName:}" failed. No retries permitted until 2025-11-28 13:33:40.057347341 +0000 UTC m=+776.864650685 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/0079ab2d-c5a8-45eb-b6e8-9574158de32f-cert") pod "infra-operator-controller-manager-57548d458d-h7npd" (UID: "0079ab2d-c5a8-45eb-b6e8-9574158de32f") : secret "infra-operator-webhook-server-cert" not found Nov 28 13:33:32 crc kubenswrapper[4631]: I1128 13:33:32.465179 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ca18d816-8a63-4dcd-ac27-b1e7ad9ce4f6-cert\") pod \"openstack-baremetal-operator-controller-manager-5fcdb54b6b2f9vf\" (UID: \"ca18d816-8a63-4dcd-ac27-b1e7ad9ce4f6\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5fcdb54b6b2f9vf" Nov 28 13:33:32 crc kubenswrapper[4631]: E1128 13:33:32.465475 4631 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Nov 28 13:33:32 crc kubenswrapper[4631]: E1128 13:33:32.465598 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ca18d816-8a63-4dcd-ac27-b1e7ad9ce4f6-cert podName:ca18d816-8a63-4dcd-ac27-b1e7ad9ce4f6 nodeName:}" failed. No retries permitted until 2025-11-28 13:33:40.465570866 +0000 UTC m=+777.272874380 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/ca18d816-8a63-4dcd-ac27-b1e7ad9ce4f6-cert") pod "openstack-baremetal-operator-controller-manager-5fcdb54b6b2f9vf" (UID: "ca18d816-8a63-4dcd-ac27-b1e7ad9ce4f6") : secret "openstack-baremetal-operator-webhook-server-cert" not found Nov 28 13:33:33 crc kubenswrapper[4631]: I1128 13:33:33.588236 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4bc38647-eefa-4ad6-bef5-6a7ed535b794-metrics-certs\") pod \"openstack-operator-controller-manager-854b9f67c6-hfb26\" (UID: \"4bc38647-eefa-4ad6-bef5-6a7ed535b794\") " pod="openstack-operators/openstack-operator-controller-manager-854b9f67c6-hfb26" Nov 28 13:33:33 crc kubenswrapper[4631]: I1128 13:33:33.588805 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/4bc38647-eefa-4ad6-bef5-6a7ed535b794-webhook-certs\") pod \"openstack-operator-controller-manager-854b9f67c6-hfb26\" (UID: \"4bc38647-eefa-4ad6-bef5-6a7ed535b794\") " pod="openstack-operators/openstack-operator-controller-manager-854b9f67c6-hfb26" Nov 28 13:33:33 crc kubenswrapper[4631]: E1128 13:33:33.589045 4631 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Nov 28 13:33:33 crc kubenswrapper[4631]: E1128 13:33:33.589177 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4bc38647-eefa-4ad6-bef5-6a7ed535b794-webhook-certs podName:4bc38647-eefa-4ad6-bef5-6a7ed535b794 nodeName:}" failed. No retries permitted until 2025-11-28 13:33:41.589101178 +0000 UTC m=+778.396404512 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/4bc38647-eefa-4ad6-bef5-6a7ed535b794-webhook-certs") pod "openstack-operator-controller-manager-854b9f67c6-hfb26" (UID: "4bc38647-eefa-4ad6-bef5-6a7ed535b794") : secret "webhook-server-cert" not found Nov 28 13:33:33 crc kubenswrapper[4631]: E1128 13:33:33.590718 4631 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Nov 28 13:33:33 crc kubenswrapper[4631]: E1128 13:33:33.590755 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4bc38647-eefa-4ad6-bef5-6a7ed535b794-metrics-certs podName:4bc38647-eefa-4ad6-bef5-6a7ed535b794 nodeName:}" failed. No retries permitted until 2025-11-28 13:33:41.590744228 +0000 UTC m=+778.398047572 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/4bc38647-eefa-4ad6-bef5-6a7ed535b794-metrics-certs") pod "openstack-operator-controller-manager-854b9f67c6-hfb26" (UID: "4bc38647-eefa-4ad6-bef5-6a7ed535b794") : secret "metrics-server-cert" not found Nov 28 13:33:39 crc kubenswrapper[4631]: E1128 13:33:39.259835 4631 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/nova-operator@sha256:c053e34316044f14929e16e4f0d97f9f1b24cb68b5e22b925ca74c66aaaed0a7" Nov 28 13:33:39 crc kubenswrapper[4631]: E1128 13:33:39.260861 4631 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/nova-operator@sha256:c053e34316044f14929e16e4f0d97f9f1b24cb68b5e22b925ca74c66aaaed0a7,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-fh7fb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod nova-operator-controller-manager-79556f57fc-tpbhk_openstack-operators(bb908ace-bb62-4f55-8131-90e48da7585b): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 28 13:33:39 crc kubenswrapper[4631]: E1128 13:33:39.801561 4631 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/glance-operator@sha256:45ae665ce2ea81aef212ee402cb02693ee49001a7c88c40c9598ff2859b838a2" Nov 28 13:33:39 crc kubenswrapper[4631]: E1128 13:33:39.801792 4631 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/glance-operator@sha256:45ae665ce2ea81aef212ee402cb02693ee49001a7c88c40c9598ff2859b838a2,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-ddkv7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod glance-operator-controller-manager-589cbd6b5b-npdtc_openstack-operators(59ee25ad-f169-47b9-9d5c-f7a7a253a591): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 28 13:33:40 crc kubenswrapper[4631]: I1128 13:33:40.107509 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/0079ab2d-c5a8-45eb-b6e8-9574158de32f-cert\") pod \"infra-operator-controller-manager-57548d458d-h7npd\" (UID: \"0079ab2d-c5a8-45eb-b6e8-9574158de32f\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-h7npd" Nov 28 13:33:40 crc kubenswrapper[4631]: I1128 13:33:40.115026 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/0079ab2d-c5a8-45eb-b6e8-9574158de32f-cert\") pod \"infra-operator-controller-manager-57548d458d-h7npd\" (UID: \"0079ab2d-c5a8-45eb-b6e8-9574158de32f\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-h7npd" Nov 28 13:33:40 crc kubenswrapper[4631]: I1128 13:33:40.231580 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-57548d458d-h7npd" Nov 28 13:33:40 crc kubenswrapper[4631]: I1128 13:33:40.517244 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ca18d816-8a63-4dcd-ac27-b1e7ad9ce4f6-cert\") pod \"openstack-baremetal-operator-controller-manager-5fcdb54b6b2f9vf\" (UID: \"ca18d816-8a63-4dcd-ac27-b1e7ad9ce4f6\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5fcdb54b6b2f9vf" Nov 28 13:33:40 crc kubenswrapper[4631]: I1128 13:33:40.521465 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ca18d816-8a63-4dcd-ac27-b1e7ad9ce4f6-cert\") pod \"openstack-baremetal-operator-controller-manager-5fcdb54b6b2f9vf\" (UID: \"ca18d816-8a63-4dcd-ac27-b1e7ad9ce4f6\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5fcdb54b6b2f9vf" Nov 28 13:33:40 crc kubenswrapper[4631]: I1128 13:33:40.782235 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5fcdb54b6b2f9vf" Nov 28 13:33:41 crc kubenswrapper[4631]: I1128 13:33:41.644598 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4bc38647-eefa-4ad6-bef5-6a7ed535b794-metrics-certs\") pod \"openstack-operator-controller-manager-854b9f67c6-hfb26\" (UID: \"4bc38647-eefa-4ad6-bef5-6a7ed535b794\") " pod="openstack-operators/openstack-operator-controller-manager-854b9f67c6-hfb26" Nov 28 13:33:41 crc kubenswrapper[4631]: I1128 13:33:41.644704 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/4bc38647-eefa-4ad6-bef5-6a7ed535b794-webhook-certs\") pod \"openstack-operator-controller-manager-854b9f67c6-hfb26\" (UID: \"4bc38647-eefa-4ad6-bef5-6a7ed535b794\") " pod="openstack-operators/openstack-operator-controller-manager-854b9f67c6-hfb26" Nov 28 13:33:41 crc kubenswrapper[4631]: I1128 13:33:41.839245 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4bc38647-eefa-4ad6-bef5-6a7ed535b794-metrics-certs\") pod \"openstack-operator-controller-manager-854b9f67c6-hfb26\" (UID: \"4bc38647-eefa-4ad6-bef5-6a7ed535b794\") " pod="openstack-operators/openstack-operator-controller-manager-854b9f67c6-hfb26" Nov 28 13:33:41 crc kubenswrapper[4631]: I1128 13:33:41.841273 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/4bc38647-eefa-4ad6-bef5-6a7ed535b794-webhook-certs\") pod \"openstack-operator-controller-manager-854b9f67c6-hfb26\" (UID: \"4bc38647-eefa-4ad6-bef5-6a7ed535b794\") " pod="openstack-operators/openstack-operator-controller-manager-854b9f67c6-hfb26" Nov 28 13:33:41 crc kubenswrapper[4631]: I1128 13:33:41.912231 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-854b9f67c6-hfb26" Nov 28 13:33:41 crc kubenswrapper[4631]: E1128 13:33:41.986209 4631 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/heat-operator@sha256:2ee37ff474bee3203447df4f326a9279a515e770573153338296dd074722c677" Nov 28 13:33:41 crc kubenswrapper[4631]: E1128 13:33:41.986587 4631 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/heat-operator@sha256:2ee37ff474bee3203447df4f326a9279a515e770573153338296dd074722c677,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-vvzpr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod heat-operator-controller-manager-5b77f656f-8pvq9_openstack-operators(9cc92c5c-285e-4604-bc60-7637938a78c8): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 28 13:33:46 crc kubenswrapper[4631]: I1128 13:33:46.704378 4631 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-nmstate/nmstate-handler-9t5sn" podUID="6dd14cda-0401-4840-8e08-aebcb536ffb9" containerName="nmstate-handler" probeResult="failure" output="command timed out" Nov 28 13:33:46 crc kubenswrapper[4631]: I1128 13:33:46.935617 4631 prober.go:107] "Probe failed" probeType="Liveness" pod="metallb-system/frr-k8s-5cpds" podUID="15c5a33d-b3c0-45aa-963f-b6d9aa639edd" containerName="frr" probeResult="failure" output="Get \"http://127.0.0.1:7573/livez\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 28 13:33:48 crc kubenswrapper[4631]: E1128 13:33:48.577525 4631 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/mariadb-operator@sha256:888edf6f432e52eaa5fc3caeae616fe38a3302b006bbba0e38885b2beba9f0f2" Nov 28 13:33:48 crc kubenswrapper[4631]: E1128 13:33:48.577946 4631 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/mariadb-operator@sha256:888edf6f432e52eaa5fc3caeae616fe38a3302b006bbba0e38885b2beba9f0f2,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-wfnvp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod mariadb-operator-controller-manager-66f4dd4bc7-m9t96_openstack-operators(cefc6050-9552-4dda-9864-39ac05626785): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 28 13:33:49 crc kubenswrapper[4631]: E1128 13:33:49.122781 4631 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/horizon-operator@sha256:9413ed1bc2ae1a6bd28c59b1c7f7e91e1638de7b2a7d4729ed3fa2135182465d" Nov 28 13:33:49 crc kubenswrapper[4631]: E1128 13:33:49.123041 4631 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/horizon-operator@sha256:9413ed1bc2ae1a6bd28c59b1c7f7e91e1638de7b2a7d4729ed3fa2135182465d,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-zhh2d,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-operator-controller-manager-5d494799bf-2mq7v_openstack-operators(aafff68d-b989-4edc-ab1c-b9d33edb8ee2): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 28 13:33:49 crc kubenswrapper[4631]: I1128 13:33:49.635415 4631 patch_prober.go:28] interesting pod/machine-config-daemon-47sxw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 28 13:33:49 crc kubenswrapper[4631]: I1128 13:33:49.635538 4631 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" podUID="8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 28 13:33:55 crc kubenswrapper[4631]: E1128 13:33:55.326730 4631 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/manila-operator@sha256:89910bc3ecceb7590d3207ac294eb7354de358cf39ef03c72323b26c598e50e6" Nov 28 13:33:55 crc kubenswrapper[4631]: E1128 13:33:55.327488 4631 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/manila-operator@sha256:89910bc3ecceb7590d3207ac294eb7354de358cf39ef03c72323b26c598e50e6,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-9w9vj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod manila-operator-controller-manager-5d499bf58b-dck2d_openstack-operators(dbbf77e7-3c97-480f-83b1-0eb8bb8f386d): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 28 13:33:55 crc kubenswrapper[4631]: I1128 13:33:55.330062 4631 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 28 13:33:56 crc kubenswrapper[4631]: E1128 13:33:56.097187 4631 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/placement-operator@sha256:225958f250a1075b69439d776a13acc45c78695c21abda23600fb53ca1640423" Nov 28 13:33:56 crc kubenswrapper[4631]: E1128 13:33:56.097514 4631 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/placement-operator@sha256:225958f250a1075b69439d776a13acc45c78695c21abda23600fb53ca1640423,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-khlhq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-operator-controller-manager-57988cc5b5-wqpdp_openstack-operators(05a6dcba-27b4-4b20-9275-f434b8943be9): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 28 13:33:56 crc kubenswrapper[4631]: E1128 13:33:56.618845 4631 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/watcher-operator@sha256:6bed55b172b9ee8ccc3952cbfc543d8bd44e2690f6db94348a754152fd78f4cf" Nov 28 13:33:56 crc kubenswrapper[4631]: E1128 13:33:56.619100 4631 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/watcher-operator@sha256:6bed55b172b9ee8ccc3952cbfc543d8bd44e2690f6db94348a754152fd78f4cf,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-s25km,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-656dcb59d4-wh9v9_openstack-operators(826f27ac-a938-4292-a357-ff54c4117986): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 28 13:34:00 crc kubenswrapper[4631]: E1128 13:34:00.391750 4631 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/octavia-operator@sha256:ddc8a82f05930db8ee7a8d6d189b5a66373060656e4baf71ac302f89c477da4c" Nov 28 13:34:00 crc kubenswrapper[4631]: E1128 13:34:00.392427 4631 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/octavia-operator@sha256:ddc8a82f05930db8ee7a8d6d189b5a66373060656e4baf71ac302f89c477da4c,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-czcr5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod octavia-operator-controller-manager-64cdc6ff96-4qzz7_openstack-operators(ecfd67ad-5136-4e6e-9e41-3d933d10ea2c): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 28 13:34:00 crc kubenswrapper[4631]: E1128 13:34:00.993897 4631 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/ovn-operator@sha256:bbb543d2d67c73e5df5d6357c3251363eb34a99575c5bf10416edd45dbdae2f6" Nov 28 13:34:00 crc kubenswrapper[4631]: E1128 13:34:00.994126 4631 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/ovn-operator@sha256:bbb543d2d67c73e5df5d6357c3251363eb34a99575c5bf10416edd45dbdae2f6,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-m2twv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovn-operator-controller-manager-56897c768d-wkpql_openstack-operators(d09fac54-dd24-47df-83a1-23cae3102e7a): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 28 13:34:01 crc kubenswrapper[4631]: E1128 13:34:01.573046 4631 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/neutron-operator@sha256:e00a9ed0ab26c5b745bd804ab1fe6b22428d026f17ea05a05f045e060342f46c" Nov 28 13:34:01 crc kubenswrapper[4631]: E1128 13:34:01.573691 4631 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/neutron-operator@sha256:e00a9ed0ab26c5b745bd804ab1fe6b22428d026f17ea05a05f045e060342f46c,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-8v2vt,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod neutron-operator-controller-manager-6fdcddb789-msc6n_openstack-operators(2dab52af-5e82-4119-b4d5-56fb31abcb4d): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 28 13:34:05 crc kubenswrapper[4631]: E1128 13:34:05.889547 4631 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.82:5001/openstack-k8s-operators/keystone-operator:1f28d89d658c10cc8240a92d3e619e8c07fd9268" Nov 28 13:34:05 crc kubenswrapper[4631]: E1128 13:34:05.889836 4631 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.82:5001/openstack-k8s-operators/keystone-operator:1f28d89d658c10cc8240a92d3e619e8c07fd9268" Nov 28 13:34:05 crc kubenswrapper[4631]: E1128 13:34:05.890028 4631 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:38.102.83.82:5001/openstack-k8s-operators/keystone-operator:1f28d89d658c10cc8240a92d3e619e8c07fd9268,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-8pxds,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod keystone-operator-controller-manager-6b49bfffcd-5ks5f_openstack-operators(f3e12436-8a74-4443-94a4-3b9717105b05): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 28 13:34:07 crc kubenswrapper[4631]: I1128 13:34:07.534676 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-hcvqs"] Nov 28 13:34:07 crc kubenswrapper[4631]: I1128 13:34:07.539000 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hcvqs" Nov 28 13:34:07 crc kubenswrapper[4631]: I1128 13:34:07.547844 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-hcvqs"] Nov 28 13:34:07 crc kubenswrapper[4631]: I1128 13:34:07.696757 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bf6aea40-3632-4c7c-90ea-5584dd081ef2-utilities\") pod \"redhat-operators-hcvqs\" (UID: \"bf6aea40-3632-4c7c-90ea-5584dd081ef2\") " pod="openshift-marketplace/redhat-operators-hcvqs" Nov 28 13:34:07 crc kubenswrapper[4631]: I1128 13:34:07.696876 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bf6aea40-3632-4c7c-90ea-5584dd081ef2-catalog-content\") pod \"redhat-operators-hcvqs\" (UID: \"bf6aea40-3632-4c7c-90ea-5584dd081ef2\") " pod="openshift-marketplace/redhat-operators-hcvqs" Nov 28 13:34:07 crc kubenswrapper[4631]: I1128 13:34:07.698193 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-76t8n\" (UniqueName: \"kubernetes.io/projected/bf6aea40-3632-4c7c-90ea-5584dd081ef2-kube-api-access-76t8n\") pod \"redhat-operators-hcvqs\" (UID: \"bf6aea40-3632-4c7c-90ea-5584dd081ef2\") " pod="openshift-marketplace/redhat-operators-hcvqs" Nov 28 13:34:07 crc kubenswrapper[4631]: I1128 13:34:07.801513 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bf6aea40-3632-4c7c-90ea-5584dd081ef2-utilities\") pod \"redhat-operators-hcvqs\" (UID: \"bf6aea40-3632-4c7c-90ea-5584dd081ef2\") " pod="openshift-marketplace/redhat-operators-hcvqs" Nov 28 13:34:07 crc kubenswrapper[4631]: I1128 13:34:07.801582 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bf6aea40-3632-4c7c-90ea-5584dd081ef2-catalog-content\") pod \"redhat-operators-hcvqs\" (UID: \"bf6aea40-3632-4c7c-90ea-5584dd081ef2\") " pod="openshift-marketplace/redhat-operators-hcvqs" Nov 28 13:34:07 crc kubenswrapper[4631]: I1128 13:34:07.801653 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-76t8n\" (UniqueName: \"kubernetes.io/projected/bf6aea40-3632-4c7c-90ea-5584dd081ef2-kube-api-access-76t8n\") pod \"redhat-operators-hcvqs\" (UID: \"bf6aea40-3632-4c7c-90ea-5584dd081ef2\") " pod="openshift-marketplace/redhat-operators-hcvqs" Nov 28 13:34:07 crc kubenswrapper[4631]: I1128 13:34:07.802677 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bf6aea40-3632-4c7c-90ea-5584dd081ef2-catalog-content\") pod \"redhat-operators-hcvqs\" (UID: \"bf6aea40-3632-4c7c-90ea-5584dd081ef2\") " pod="openshift-marketplace/redhat-operators-hcvqs" Nov 28 13:34:07 crc kubenswrapper[4631]: I1128 13:34:07.802949 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bf6aea40-3632-4c7c-90ea-5584dd081ef2-utilities\") pod \"redhat-operators-hcvqs\" (UID: \"bf6aea40-3632-4c7c-90ea-5584dd081ef2\") " pod="openshift-marketplace/redhat-operators-hcvqs" Nov 28 13:34:07 crc kubenswrapper[4631]: I1128 13:34:07.828653 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-76t8n\" (UniqueName: \"kubernetes.io/projected/bf6aea40-3632-4c7c-90ea-5584dd081ef2-kube-api-access-76t8n\") pod \"redhat-operators-hcvqs\" (UID: \"bf6aea40-3632-4c7c-90ea-5584dd081ef2\") " pod="openshift-marketplace/redhat-operators-hcvqs" Nov 28 13:34:07 crc kubenswrapper[4631]: I1128 13:34:07.860792 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hcvqs" Nov 28 13:34:09 crc kubenswrapper[4631]: I1128 13:34:09.892649 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-zfd5w"] Nov 28 13:34:09 crc kubenswrapper[4631]: I1128 13:34:09.894101 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zfd5w" Nov 28 13:34:09 crc kubenswrapper[4631]: I1128 13:34:09.915139 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-zfd5w"] Nov 28 13:34:10 crc kubenswrapper[4631]: I1128 13:34:10.040546 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h87rf\" (UniqueName: \"kubernetes.io/projected/8206d20e-c1fb-46c2-b9ab-b13a38922d29-kube-api-access-h87rf\") pod \"redhat-marketplace-zfd5w\" (UID: \"8206d20e-c1fb-46c2-b9ab-b13a38922d29\") " pod="openshift-marketplace/redhat-marketplace-zfd5w" Nov 28 13:34:10 crc kubenswrapper[4631]: I1128 13:34:10.040594 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8206d20e-c1fb-46c2-b9ab-b13a38922d29-catalog-content\") pod \"redhat-marketplace-zfd5w\" (UID: \"8206d20e-c1fb-46c2-b9ab-b13a38922d29\") " pod="openshift-marketplace/redhat-marketplace-zfd5w" Nov 28 13:34:10 crc kubenswrapper[4631]: I1128 13:34:10.040658 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8206d20e-c1fb-46c2-b9ab-b13a38922d29-utilities\") pod \"redhat-marketplace-zfd5w\" (UID: \"8206d20e-c1fb-46c2-b9ab-b13a38922d29\") " pod="openshift-marketplace/redhat-marketplace-zfd5w" Nov 28 13:34:10 crc kubenswrapper[4631]: I1128 13:34:10.142030 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8206d20e-c1fb-46c2-b9ab-b13a38922d29-utilities\") pod \"redhat-marketplace-zfd5w\" (UID: \"8206d20e-c1fb-46c2-b9ab-b13a38922d29\") " pod="openshift-marketplace/redhat-marketplace-zfd5w" Nov 28 13:34:10 crc kubenswrapper[4631]: I1128 13:34:10.142143 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h87rf\" (UniqueName: \"kubernetes.io/projected/8206d20e-c1fb-46c2-b9ab-b13a38922d29-kube-api-access-h87rf\") pod \"redhat-marketplace-zfd5w\" (UID: \"8206d20e-c1fb-46c2-b9ab-b13a38922d29\") " pod="openshift-marketplace/redhat-marketplace-zfd5w" Nov 28 13:34:10 crc kubenswrapper[4631]: I1128 13:34:10.142171 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8206d20e-c1fb-46c2-b9ab-b13a38922d29-catalog-content\") pod \"redhat-marketplace-zfd5w\" (UID: \"8206d20e-c1fb-46c2-b9ab-b13a38922d29\") " pod="openshift-marketplace/redhat-marketplace-zfd5w" Nov 28 13:34:10 crc kubenswrapper[4631]: I1128 13:34:10.142721 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8206d20e-c1fb-46c2-b9ab-b13a38922d29-utilities\") pod \"redhat-marketplace-zfd5w\" (UID: \"8206d20e-c1fb-46c2-b9ab-b13a38922d29\") " pod="openshift-marketplace/redhat-marketplace-zfd5w" Nov 28 13:34:10 crc kubenswrapper[4631]: I1128 13:34:10.142767 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8206d20e-c1fb-46c2-b9ab-b13a38922d29-catalog-content\") pod \"redhat-marketplace-zfd5w\" (UID: \"8206d20e-c1fb-46c2-b9ab-b13a38922d29\") " pod="openshift-marketplace/redhat-marketplace-zfd5w" Nov 28 13:34:10 crc kubenswrapper[4631]: I1128 13:34:10.168576 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h87rf\" (UniqueName: \"kubernetes.io/projected/8206d20e-c1fb-46c2-b9ab-b13a38922d29-kube-api-access-h87rf\") pod \"redhat-marketplace-zfd5w\" (UID: \"8206d20e-c1fb-46c2-b9ab-b13a38922d29\") " pod="openshift-marketplace/redhat-marketplace-zfd5w" Nov 28 13:34:10 crc kubenswrapper[4631]: I1128 13:34:10.229648 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zfd5w" Nov 28 13:34:12 crc kubenswrapper[4631]: E1128 13:34:12.483388 4631 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2" Nov 28 13:34:12 crc kubenswrapper[4631]: E1128 13:34:12.483997 4631 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-llbd2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-668c99d594-nrpfx_openstack-operators(10a95153-0898-4016-92a6-d56633555348): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 28 13:34:12 crc kubenswrapper[4631]: E1128 13:34:12.485373 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-nrpfx" podUID="10a95153-0898-4016-92a6-d56633555348" Nov 28 13:34:13 crc kubenswrapper[4631]: I1128 13:34:13.032839 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-57548d458d-h7npd"] Nov 28 13:34:13 crc kubenswrapper[4631]: I1128 13:34:13.119946 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-854b9f67c6-hfb26"] Nov 28 13:34:13 crc kubenswrapper[4631]: I1128 13:34:13.194184 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-5fcdb54b6b2f9vf"] Nov 28 13:34:13 crc kubenswrapper[4631]: W1128 13:34:13.968138 4631 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0079ab2d_c5a8_45eb_b6e8_9574158de32f.slice/crio-c5585619452cea1b49170771725312f6e0a1658dd583ccdb2dbc3ec324041e23 WatchSource:0}: Error finding container c5585619452cea1b49170771725312f6e0a1658dd583ccdb2dbc3ec324041e23: Status 404 returned error can't find the container with id c5585619452cea1b49170771725312f6e0a1658dd583ccdb2dbc3ec324041e23 Nov 28 13:34:13 crc kubenswrapper[4631]: W1128 13:34:13.973415 4631 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4bc38647_eefa_4ad6_bef5_6a7ed535b794.slice/crio-8c0afbb29a6d73c202a6478e705a006c01798978ca3b9c6a40e9a152edfe19ad WatchSource:0}: Error finding container 8c0afbb29a6d73c202a6478e705a006c01798978ca3b9c6a40e9a152edfe19ad: Status 404 returned error can't find the container with id 8c0afbb29a6d73c202a6478e705a006c01798978ca3b9c6a40e9a152edfe19ad Nov 28 13:34:14 crc kubenswrapper[4631]: I1128 13:34:14.402843 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-hcvqs"] Nov 28 13:34:14 crc kubenswrapper[4631]: I1128 13:34:14.895070 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-zfd5w"] Nov 28 13:34:14 crc kubenswrapper[4631]: I1128 13:34:14.980914 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hcvqs" event={"ID":"bf6aea40-3632-4c7c-90ea-5584dd081ef2","Type":"ContainerStarted","Data":"6e5c15dab80a930be599aa068a59bc4e2407286fc1851689ae98fd0c777da5df"} Nov 28 13:34:14 crc kubenswrapper[4631]: I1128 13:34:14.987530 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5cd6c7f4c8-jwd25" event={"ID":"fa36877c-8fc2-4074-a75d-135c2006807a","Type":"ContainerStarted","Data":"22354639bc7ddfef5c210b8d97eca2c6962be5b5e7701cfa19474d31b56641ff"} Nov 28 13:34:14 crc kubenswrapper[4631]: I1128 13:34:14.990836 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-854b9f67c6-hfb26" event={"ID":"4bc38647-eefa-4ad6-bef5-6a7ed535b794","Type":"ContainerStarted","Data":"8c0afbb29a6d73c202a6478e705a006c01798978ca3b9c6a40e9a152edfe19ad"} Nov 28 13:34:14 crc kubenswrapper[4631]: I1128 13:34:14.995091 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-57548d458d-h7npd" event={"ID":"0079ab2d-c5a8-45eb-b6e8-9574158de32f","Type":"ContainerStarted","Data":"c5585619452cea1b49170771725312f6e0a1658dd583ccdb2dbc3ec324041e23"} Nov 28 13:34:14 crc kubenswrapper[4631]: I1128 13:34:14.998781 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-6b7f75547b-rmlgq" event={"ID":"8028d28d-2414-4ffe-9c5d-4854f77e980f","Type":"ContainerStarted","Data":"367c5c6a3c417e1871e652442a00b20ee04fff44268535ffa11f778c0f034b6e"} Nov 28 13:34:15 crc kubenswrapper[4631]: I1128 13:34:15.000122 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5fcdb54b6b2f9vf" event={"ID":"ca18d816-8a63-4dcd-ac27-b1e7ad9ce4f6","Type":"ContainerStarted","Data":"8e785f799d19282cebdc276c3f19a62d2c0f228f96f95c0eed57a17de467025b"} Nov 28 13:34:15 crc kubenswrapper[4631]: I1128 13:34:15.006474 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7b64f4fb85-hdr72" event={"ID":"00a81f80-35b9-4eb4-a508-3f104ca426be","Type":"ContainerStarted","Data":"a252d3700d39856f17cdb0cca594d6084241ee095f3e5c65c289651faaa6c00a"} Nov 28 13:34:15 crc kubenswrapper[4631]: I1128 13:34:15.033233 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-955677c94-6bt2f" event={"ID":"06994d92-3dcb-461c-83ac-c7e65d4bd640","Type":"ContainerStarted","Data":"7dd6d580051146cf40047d2babbe1124c1f3656819f68efa7f7c41f51f2acc0f"} Nov 28 13:34:15 crc kubenswrapper[4631]: I1128 13:34:15.062498 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-67cb4dc6d4-db8kg" event={"ID":"ed17d71f-b72c-4f12-b5f7-d5cb3bb130dc","Type":"ContainerStarted","Data":"f5625aa292aa7356e0cf3b6b3b256b9f6c6ecb6e0883ec786d02d5daf6e00812"} Nov 28 13:34:15 crc kubenswrapper[4631]: W1128 13:34:15.174971 4631 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8206d20e_c1fb_46c2_b9ab_b13a38922d29.slice/crio-6c49094ec9c124d09271e55d9afae45554dcc2a31a1a915cd4c670db01a6b489 WatchSource:0}: Error finding container 6c49094ec9c124d09271e55d9afae45554dcc2a31a1a915cd4c670db01a6b489: Status 404 returned error can't find the container with id 6c49094ec9c124d09271e55d9afae45554dcc2a31a1a915cd4c670db01a6b489 Nov 28 13:34:16 crc kubenswrapper[4631]: I1128 13:34:16.085696 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-r94wn" event={"ID":"f2aea9a5-85cf-464f-8eba-382b3db8d71d","Type":"ContainerStarted","Data":"377739abd3b9398cfed515762ac429d7fc790c3ef40605e28fb42ab448ab304e"} Nov 28 13:34:16 crc kubenswrapper[4631]: I1128 13:34:16.107630 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-d77b94747-2g446" event={"ID":"2d59108a-2b92-421b-beb8-157c2eca09ac","Type":"ContainerStarted","Data":"f7ed941fe43aa49b31b9c9ee1343b4af38395aca34e10152da632a6292aa42cd"} Nov 28 13:34:16 crc kubenswrapper[4631]: I1128 13:34:16.114936 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zfd5w" event={"ID":"8206d20e-c1fb-46c2-b9ab-b13a38922d29","Type":"ContainerStarted","Data":"6c49094ec9c124d09271e55d9afae45554dcc2a31a1a915cd4c670db01a6b489"} Nov 28 13:34:18 crc kubenswrapper[4631]: E1128 13:34:18.026248 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/horizon-operator-controller-manager-5d494799bf-2mq7v" podUID="aafff68d-b989-4edc-ab1c-b9d33edb8ee2" Nov 28 13:34:18 crc kubenswrapper[4631]: I1128 13:34:18.137545 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7b64f4fb85-hdr72" event={"ID":"00a81f80-35b9-4eb4-a508-3f104ca426be","Type":"ContainerStarted","Data":"99c393d02ff6199efcc9cf96cc6f51736058458678ccf272a0c99cfc1ea47ee9"} Nov 28 13:34:18 crc kubenswrapper[4631]: I1128 13:34:18.137699 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-7b64f4fb85-hdr72" Nov 28 13:34:18 crc kubenswrapper[4631]: I1128 13:34:18.139760 4631 generic.go:334] "Generic (PLEG): container finished" podID="bf6aea40-3632-4c7c-90ea-5584dd081ef2" containerID="128147b0b653091336f64aea110f6a9a5219757366275397a467ddcdb05f1ccf" exitCode=0 Nov 28 13:34:18 crc kubenswrapper[4631]: I1128 13:34:18.139926 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hcvqs" event={"ID":"bf6aea40-3632-4c7c-90ea-5584dd081ef2","Type":"ContainerDied","Data":"128147b0b653091336f64aea110f6a9a5219757366275397a467ddcdb05f1ccf"} Nov 28 13:34:18 crc kubenswrapper[4631]: I1128 13:34:18.143970 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-854b9f67c6-hfb26" event={"ID":"4bc38647-eefa-4ad6-bef5-6a7ed535b794","Type":"ContainerStarted","Data":"8274314bd98fb986ace8a13cc5a8186c8d89ccbd4b08f53914d8d6a7d0389fc3"} Nov 28 13:34:18 crc kubenswrapper[4631]: I1128 13:34:18.144819 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-854b9f67c6-hfb26" Nov 28 13:34:18 crc kubenswrapper[4631]: I1128 13:34:18.146790 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-5d494799bf-2mq7v" event={"ID":"aafff68d-b989-4edc-ab1c-b9d33edb8ee2","Type":"ContainerStarted","Data":"ca302137ef9775f24a68dca304f6a1ff2d7709b52295d4835005cf67bc2a9a0c"} Nov 28 13:34:18 crc kubenswrapper[4631]: I1128 13:34:18.149442 4631 generic.go:334] "Generic (PLEG): container finished" podID="8206d20e-c1fb-46c2-b9ab-b13a38922d29" containerID="9200a264305d5c2092cb9c6e312b5db7863ca4b3d226a9427048aad7767fcb88" exitCode=0 Nov 28 13:34:18 crc kubenswrapper[4631]: I1128 13:34:18.149472 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zfd5w" event={"ID":"8206d20e-c1fb-46c2-b9ab-b13a38922d29","Type":"ContainerDied","Data":"9200a264305d5c2092cb9c6e312b5db7863ca4b3d226a9427048aad7767fcb88"} Nov 28 13:34:18 crc kubenswrapper[4631]: I1128 13:34:18.164885 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-7b64f4fb85-hdr72" podStartSLOduration=2.888186869 podStartE2EDuration="55.164861093s" podCreationTimestamp="2025-11-28 13:33:23 +0000 UTC" firstStartedPulling="2025-11-28 13:33:25.380913026 +0000 UTC m=+762.188216370" lastFinishedPulling="2025-11-28 13:34:17.65758725 +0000 UTC m=+814.464890594" observedRunningTime="2025-11-28 13:34:18.162923865 +0000 UTC m=+814.970227199" watchObservedRunningTime="2025-11-28 13:34:18.164861093 +0000 UTC m=+814.972164427" Nov 28 13:34:18 crc kubenswrapper[4631]: I1128 13:34:18.333066 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-854b9f67c6-hfb26" podStartSLOduration=53.333042559 podStartE2EDuration="53.333042559s" podCreationTimestamp="2025-11-28 13:33:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 13:34:18.307149885 +0000 UTC m=+815.114453229" watchObservedRunningTime="2025-11-28 13:34:18.333042559 +0000 UTC m=+815.140345903" Nov 28 13:34:19 crc kubenswrapper[4631]: I1128 13:34:19.159714 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-7b64f4fb85-hdr72" Nov 28 13:34:19 crc kubenswrapper[4631]: I1128 13:34:19.634687 4631 patch_prober.go:28] interesting pod/machine-config-daemon-47sxw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 28 13:34:19 crc kubenswrapper[4631]: I1128 13:34:19.634775 4631 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" podUID="8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 28 13:34:19 crc kubenswrapper[4631]: I1128 13:34:19.634834 4631 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" Nov 28 13:34:19 crc kubenswrapper[4631]: I1128 13:34:19.635901 4631 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"914e9854c9ef96234991a24021b15fad0f64f1894afc1a19ce66cd4bc2c2af7d"} pod="openshift-machine-config-operator/machine-config-daemon-47sxw" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 28 13:34:19 crc kubenswrapper[4631]: I1128 13:34:19.635975 4631 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" podUID="8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1" containerName="machine-config-daemon" containerID="cri-o://914e9854c9ef96234991a24021b15fad0f64f1894afc1a19ce66cd4bc2c2af7d" gracePeriod=600 Nov 28 13:34:20 crc kubenswrapper[4631]: E1128 13:34:20.072369 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/heat-operator-controller-manager-5b77f656f-8pvq9" podUID="9cc92c5c-285e-4604-bc60-7637938a78c8" Nov 28 13:34:20 crc kubenswrapper[4631]: I1128 13:34:20.180843 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-955677c94-6bt2f" event={"ID":"06994d92-3dcb-461c-83ac-c7e65d4bd640","Type":"ContainerStarted","Data":"3ef2382c836514a3d62cfe084c2b817bd6f651239052f2309fce30c706ca88da"} Nov 28 13:34:20 crc kubenswrapper[4631]: I1128 13:34:20.181417 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-955677c94-6bt2f" Nov 28 13:34:20 crc kubenswrapper[4631]: I1128 13:34:20.186184 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-955677c94-6bt2f" Nov 28 13:34:20 crc kubenswrapper[4631]: I1128 13:34:20.193325 4631 generic.go:334] "Generic (PLEG): container finished" podID="8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1" containerID="914e9854c9ef96234991a24021b15fad0f64f1894afc1a19ce66cd4bc2c2af7d" exitCode=0 Nov 28 13:34:20 crc kubenswrapper[4631]: I1128 13:34:20.193415 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" event={"ID":"8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1","Type":"ContainerDied","Data":"914e9854c9ef96234991a24021b15fad0f64f1894afc1a19ce66cd4bc2c2af7d"} Nov 28 13:34:20 crc kubenswrapper[4631]: I1128 13:34:20.193457 4631 scope.go:117] "RemoveContainer" containerID="42d567fa8bd6c073cd89c5129f91adc5f5feefc7933a184e745232605d6012b2" Nov 28 13:34:20 crc kubenswrapper[4631]: I1128 13:34:20.222044 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5b77f656f-8pvq9" event={"ID":"9cc92c5c-285e-4604-bc60-7637938a78c8","Type":"ContainerStarted","Data":"adf92ce35526b2d906022326424fbb4b953b380e1637b41391425a1cde9a42a8"} Nov 28 13:34:20 crc kubenswrapper[4631]: I1128 13:34:20.228126 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-955677c94-6bt2f" podStartSLOduration=6.185509506 podStartE2EDuration="57.228104844s" podCreationTimestamp="2025-11-28 13:33:23 +0000 UTC" firstStartedPulling="2025-11-28 13:33:26.652938854 +0000 UTC m=+763.460242208" lastFinishedPulling="2025-11-28 13:34:17.695534202 +0000 UTC m=+814.502837546" observedRunningTime="2025-11-28 13:34:20.219553672 +0000 UTC m=+817.026857016" watchObservedRunningTime="2025-11-28 13:34:20.228104844 +0000 UTC m=+817.035408178" Nov 28 13:34:20 crc kubenswrapper[4631]: E1128 13:34:20.449435 4631 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0" Nov 28 13:34:20 crc kubenswrapper[4631]: E1128 13:34:20.449621 4631 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-ddkv7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod glance-operator-controller-manager-589cbd6b5b-npdtc_openstack-operators(59ee25ad-f169-47b9-9d5c-f7a7a253a591): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 28 13:34:20 crc kubenswrapper[4631]: E1128 13:34:20.452438 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"]" pod="openstack-operators/glance-operator-controller-manager-589cbd6b5b-npdtc" podUID="59ee25ad-f169-47b9-9d5c-f7a7a253a591" Nov 28 13:34:21 crc kubenswrapper[4631]: E1128 13:34:21.170592 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/placement-operator-controller-manager-57988cc5b5-wqpdp" podUID="05a6dcba-27b4-4b20-9275-f434b8943be9" Nov 28 13:34:21 crc kubenswrapper[4631]: I1128 13:34:21.241273 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-57988cc5b5-wqpdp" event={"ID":"05a6dcba-27b4-4b20-9275-f434b8943be9","Type":"ContainerStarted","Data":"e2853236b10b97d1c114e4677e92fd04e8744d6e536f2d30b357deb69dbf6dc8"} Nov 28 13:34:21 crc kubenswrapper[4631]: I1128 13:34:21.250950 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-67cb4dc6d4-db8kg" event={"ID":"ed17d71f-b72c-4f12-b5f7-d5cb3bb130dc","Type":"ContainerStarted","Data":"b4b01947555a4dd3f7ad96990530430471d0fef262dd6d93a4fced321f43643b"} Nov 28 13:34:21 crc kubenswrapper[4631]: I1128 13:34:21.251687 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-67cb4dc6d4-db8kg" Nov 28 13:34:21 crc kubenswrapper[4631]: I1128 13:34:21.255415 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-67cb4dc6d4-db8kg" Nov 28 13:34:21 crc kubenswrapper[4631]: I1128 13:34:21.256299 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" event={"ID":"8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1","Type":"ContainerStarted","Data":"a4b336d735e84003b86dbce21368f3f29fb0ded3e7a12a4562271fa2767a4ae8"} Nov 28 13:34:21 crc kubenswrapper[4631]: I1128 13:34:21.523858 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-67cb4dc6d4-db8kg" podStartSLOduration=6.896564679 podStartE2EDuration="58.523829941s" podCreationTimestamp="2025-11-28 13:33:23 +0000 UTC" firstStartedPulling="2025-11-28 13:33:26.249588031 +0000 UTC m=+763.056891375" lastFinishedPulling="2025-11-28 13:34:17.876853283 +0000 UTC m=+814.684156637" observedRunningTime="2025-11-28 13:34:21.499994389 +0000 UTC m=+818.307297733" watchObservedRunningTime="2025-11-28 13:34:21.523829941 +0000 UTC m=+818.331133285" Nov 28 13:34:22 crc kubenswrapper[4631]: E1128 13:34:22.129550 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/mariadb-operator-controller-manager-66f4dd4bc7-m9t96" podUID="cefc6050-9552-4dda-9864-39ac05626785" Nov 28 13:34:22 crc kubenswrapper[4631]: E1128 13:34:22.129908 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/keystone-operator-controller-manager-6b49bfffcd-5ks5f" podUID="f3e12436-8a74-4443-94a4-3b9717105b05" Nov 28 13:34:22 crc kubenswrapper[4631]: E1128 13:34:22.131027 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/neutron-operator-controller-manager-6fdcddb789-msc6n" podUID="2dab52af-5e82-4119-b4d5-56fb31abcb4d" Nov 28 13:34:22 crc kubenswrapper[4631]: I1128 13:34:22.283132 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hcvqs" event={"ID":"bf6aea40-3632-4c7c-90ea-5584dd081ef2","Type":"ContainerStarted","Data":"61f33b53d073e336d54579efe11721c6af8ba587cb0970340e04cc2357d84a82"} Nov 28 13:34:22 crc kubenswrapper[4631]: I1128 13:34:22.295457 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5cd6c7f4c8-jwd25" event={"ID":"fa36877c-8fc2-4074-a75d-135c2006807a","Type":"ContainerStarted","Data":"fbf8764ec1b8284a6cec33ecb2d714bf2ffda60829e01453d4d24663ec3a8389"} Nov 28 13:34:22 crc kubenswrapper[4631]: I1128 13:34:22.295676 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-5cd6c7f4c8-jwd25" Nov 28 13:34:22 crc kubenswrapper[4631]: I1128 13:34:22.301630 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-5cd6c7f4c8-jwd25" Nov 28 13:34:22 crc kubenswrapper[4631]: I1128 13:34:22.311665 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-5d494799bf-2mq7v" event={"ID":"aafff68d-b989-4edc-ab1c-b9d33edb8ee2","Type":"ContainerStarted","Data":"f2fcede62882b4e2d62396c66fd48a9662b1d65e6fb781341f848c25a3b5af04"} Nov 28 13:34:22 crc kubenswrapper[4631]: I1128 13:34:22.311746 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-5d494799bf-2mq7v" Nov 28 13:34:22 crc kubenswrapper[4631]: I1128 13:34:22.326175 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-6fdcddb789-msc6n" event={"ID":"2dab52af-5e82-4119-b4d5-56fb31abcb4d","Type":"ContainerStarted","Data":"50906de1d85f02e49949dcf19f6afabd113ddc1a11601687fd22ebaa552f80d8"} Nov 28 13:34:22 crc kubenswrapper[4631]: I1128 13:34:22.348760 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-6b49bfffcd-5ks5f" event={"ID":"f3e12436-8a74-4443-94a4-3b9717105b05","Type":"ContainerStarted","Data":"225890353ac71caf2493e0d09040c6bfee8ec5a05acfb6ec557404370f6c9f56"} Nov 28 13:34:22 crc kubenswrapper[4631]: I1128 13:34:22.371165 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-6b7f75547b-rmlgq" event={"ID":"8028d28d-2414-4ffe-9c5d-4854f77e980f","Type":"ContainerStarted","Data":"693feb3ebe9af15a231cc7606e6a69fe1e978a434a476e800cbb101acd47e8df"} Nov 28 13:34:22 crc kubenswrapper[4631]: I1128 13:34:22.371799 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-6b7f75547b-rmlgq" Nov 28 13:34:22 crc kubenswrapper[4631]: I1128 13:34:22.378278 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-6b7f75547b-rmlgq" Nov 28 13:34:22 crc kubenswrapper[4631]: I1128 13:34:22.379197 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5fcdb54b6b2f9vf" event={"ID":"ca18d816-8a63-4dcd-ac27-b1e7ad9ce4f6","Type":"ContainerStarted","Data":"62ce20e6766f267577a39774338fb9a85165bcd057c190ff7cbf3364873b396d"} Nov 28 13:34:22 crc kubenswrapper[4631]: I1128 13:34:22.382654 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-66f4dd4bc7-m9t96" event={"ID":"cefc6050-9552-4dda-9864-39ac05626785","Type":"ContainerStarted","Data":"70b69754b8c78cd22ec8f04cd1648a74b21c9166958475231557e54d399dcb50"} Nov 28 13:34:22 crc kubenswrapper[4631]: I1128 13:34:22.393448 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-5cd6c7f4c8-jwd25" podStartSLOduration=4.472999783 podStartE2EDuration="57.393424729s" podCreationTimestamp="2025-11-28 13:33:25 +0000 UTC" firstStartedPulling="2025-11-28 13:33:27.389169951 +0000 UTC m=+764.196473295" lastFinishedPulling="2025-11-28 13:34:20.309594897 +0000 UTC m=+817.116898241" observedRunningTime="2025-11-28 13:34:22.392389223 +0000 UTC m=+819.199692567" watchObservedRunningTime="2025-11-28 13:34:22.393424729 +0000 UTC m=+819.200728073" Nov 28 13:34:22 crc kubenswrapper[4631]: I1128 13:34:22.393843 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-r94wn" event={"ID":"f2aea9a5-85cf-464f-8eba-382b3db8d71d","Type":"ContainerStarted","Data":"cdfc3f2341ae5fdfab69b16972554f2e35f624507baba3f4c0924f572b232758"} Nov 28 13:34:22 crc kubenswrapper[4631]: I1128 13:34:22.395163 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-r94wn" Nov 28 13:34:22 crc kubenswrapper[4631]: I1128 13:34:22.400546 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-5d494799bf-2mq7v" podStartSLOduration=6.083130554 podStartE2EDuration="59.400528085s" podCreationTimestamp="2025-11-28 13:33:23 +0000 UTC" firstStartedPulling="2025-11-28 13:33:27.191033422 +0000 UTC m=+763.998336766" lastFinishedPulling="2025-11-28 13:34:20.508430953 +0000 UTC m=+817.315734297" observedRunningTime="2025-11-28 13:34:22.364031149 +0000 UTC m=+819.171334493" watchObservedRunningTime="2025-11-28 13:34:22.400528085 +0000 UTC m=+819.207831429" Nov 28 13:34:22 crc kubenswrapper[4631]: I1128 13:34:22.416874 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-r94wn" Nov 28 13:34:22 crc kubenswrapper[4631]: E1128 13:34:22.515231 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-nrpfx" podUID="10a95153-0898-4016-92a6-d56633555348" Nov 28 13:34:22 crc kubenswrapper[4631]: I1128 13:34:22.659196 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-6b7f75547b-rmlgq" podStartSLOduration=5.827242981 podStartE2EDuration="59.659174636s" podCreationTimestamp="2025-11-28 13:33:23 +0000 UTC" firstStartedPulling="2025-11-28 13:33:26.245649213 +0000 UTC m=+763.052952557" lastFinishedPulling="2025-11-28 13:34:20.077580868 +0000 UTC m=+816.884884212" observedRunningTime="2025-11-28 13:34:22.655086705 +0000 UTC m=+819.462390049" watchObservedRunningTime="2025-11-28 13:34:22.659174636 +0000 UTC m=+819.466477970" Nov 28 13:34:22 crc kubenswrapper[4631]: I1128 13:34:22.746119 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-r94wn" podStartSLOduration=5.84184033 podStartE2EDuration="58.746098184s" podCreationTimestamp="2025-11-28 13:33:24 +0000 UTC" firstStartedPulling="2025-11-28 13:33:27.398559625 +0000 UTC m=+764.205862969" lastFinishedPulling="2025-11-28 13:34:20.302817479 +0000 UTC m=+817.110120823" observedRunningTime="2025-11-28 13:34:22.722032296 +0000 UTC m=+819.529335650" watchObservedRunningTime="2025-11-28 13:34:22.746098184 +0000 UTC m=+819.553401528" Nov 28 13:34:22 crc kubenswrapper[4631]: I1128 13:34:22.836441 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-979dz"] Nov 28 13:34:22 crc kubenswrapper[4631]: I1128 13:34:22.838783 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-979dz" Nov 28 13:34:22 crc kubenswrapper[4631]: I1128 13:34:22.856894 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-979dz"] Nov 28 13:34:23 crc kubenswrapper[4631]: I1128 13:34:23.032476 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j8kvm\" (UniqueName: \"kubernetes.io/projected/fae1e74c-0fac-4e9c-bf7c-e91157b6ef2d-kube-api-access-j8kvm\") pod \"certified-operators-979dz\" (UID: \"fae1e74c-0fac-4e9c-bf7c-e91157b6ef2d\") " pod="openshift-marketplace/certified-operators-979dz" Nov 28 13:34:23 crc kubenswrapper[4631]: I1128 13:34:23.032584 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fae1e74c-0fac-4e9c-bf7c-e91157b6ef2d-utilities\") pod \"certified-operators-979dz\" (UID: \"fae1e74c-0fac-4e9c-bf7c-e91157b6ef2d\") " pod="openshift-marketplace/certified-operators-979dz" Nov 28 13:34:23 crc kubenswrapper[4631]: I1128 13:34:23.032670 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fae1e74c-0fac-4e9c-bf7c-e91157b6ef2d-catalog-content\") pod \"certified-operators-979dz\" (UID: \"fae1e74c-0fac-4e9c-bf7c-e91157b6ef2d\") " pod="openshift-marketplace/certified-operators-979dz" Nov 28 13:34:23 crc kubenswrapper[4631]: I1128 13:34:23.134193 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fae1e74c-0fac-4e9c-bf7c-e91157b6ef2d-utilities\") pod \"certified-operators-979dz\" (UID: \"fae1e74c-0fac-4e9c-bf7c-e91157b6ef2d\") " pod="openshift-marketplace/certified-operators-979dz" Nov 28 13:34:23 crc kubenswrapper[4631]: I1128 13:34:23.134271 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fae1e74c-0fac-4e9c-bf7c-e91157b6ef2d-catalog-content\") pod \"certified-operators-979dz\" (UID: \"fae1e74c-0fac-4e9c-bf7c-e91157b6ef2d\") " pod="openshift-marketplace/certified-operators-979dz" Nov 28 13:34:23 crc kubenswrapper[4631]: I1128 13:34:23.134303 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j8kvm\" (UniqueName: \"kubernetes.io/projected/fae1e74c-0fac-4e9c-bf7c-e91157b6ef2d-kube-api-access-j8kvm\") pod \"certified-operators-979dz\" (UID: \"fae1e74c-0fac-4e9c-bf7c-e91157b6ef2d\") " pod="openshift-marketplace/certified-operators-979dz" Nov 28 13:34:23 crc kubenswrapper[4631]: I1128 13:34:23.134830 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fae1e74c-0fac-4e9c-bf7c-e91157b6ef2d-utilities\") pod \"certified-operators-979dz\" (UID: \"fae1e74c-0fac-4e9c-bf7c-e91157b6ef2d\") " pod="openshift-marketplace/certified-operators-979dz" Nov 28 13:34:23 crc kubenswrapper[4631]: I1128 13:34:23.134895 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fae1e74c-0fac-4e9c-bf7c-e91157b6ef2d-catalog-content\") pod \"certified-operators-979dz\" (UID: \"fae1e74c-0fac-4e9c-bf7c-e91157b6ef2d\") " pod="openshift-marketplace/certified-operators-979dz" Nov 28 13:34:23 crc kubenswrapper[4631]: I1128 13:34:23.186323 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j8kvm\" (UniqueName: \"kubernetes.io/projected/fae1e74c-0fac-4e9c-bf7c-e91157b6ef2d-kube-api-access-j8kvm\") pod \"certified-operators-979dz\" (UID: \"fae1e74c-0fac-4e9c-bf7c-e91157b6ef2d\") " pod="openshift-marketplace/certified-operators-979dz" Nov 28 13:34:23 crc kubenswrapper[4631]: I1128 13:34:23.467613 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-979dz" Nov 28 13:34:26 crc kubenswrapper[4631]: I1128 13:34:26.467620 4631 generic.go:334] "Generic (PLEG): container finished" podID="bf6aea40-3632-4c7c-90ea-5584dd081ef2" containerID="61f33b53d073e336d54579efe11721c6af8ba587cb0970340e04cc2357d84a82" exitCode=0 Nov 28 13:34:26 crc kubenswrapper[4631]: I1128 13:34:26.468105 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hcvqs" event={"ID":"bf6aea40-3632-4c7c-90ea-5584dd081ef2","Type":"ContainerDied","Data":"61f33b53d073e336d54579efe11721c6af8ba587cb0970340e04cc2357d84a82"} Nov 28 13:34:26 crc kubenswrapper[4631]: E1128 13:34:26.602966 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-tpbhk" podUID="bb908ace-bb62-4f55-8131-90e48da7585b" Nov 28 13:34:26 crc kubenswrapper[4631]: E1128 13:34:26.637611 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/manila-operator-controller-manager-5d499bf58b-dck2d" podUID="dbbf77e7-3c97-480f-83b1-0eb8bb8f386d" Nov 28 13:34:26 crc kubenswrapper[4631]: E1128 13:34:26.683605 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/octavia-operator-controller-manager-64cdc6ff96-4qzz7" podUID="ecfd67ad-5136-4e6e-9e41-3d933d10ea2c" Nov 28 13:34:26 crc kubenswrapper[4631]: I1128 13:34:26.795528 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-979dz"] Nov 28 13:34:26 crc kubenswrapper[4631]: E1128 13:34:26.819894 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/watcher-operator-controller-manager-656dcb59d4-wh9v9" podUID="826f27ac-a938-4292-a357-ff54c4117986" Nov 28 13:34:27 crc kubenswrapper[4631]: E1128 13:34:27.136052 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/ovn-operator-controller-manager-56897c768d-wkpql" podUID="d09fac54-dd24-47df-83a1-23cae3102e7a" Nov 28 13:34:27 crc kubenswrapper[4631]: I1128 13:34:27.611981 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-589cbd6b5b-npdtc" event={"ID":"59ee25ad-f169-47b9-9d5c-f7a7a253a591","Type":"ContainerStarted","Data":"1ee20a6a172188b699692142b529ec0f74590b3b53b3814ac65ab6b8bba8bf75"} Nov 28 13:34:27 crc kubenswrapper[4631]: I1128 13:34:27.636594 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-6fdcddb789-msc6n" event={"ID":"2dab52af-5e82-4119-b4d5-56fb31abcb4d","Type":"ContainerStarted","Data":"0128e1b147130ac956975efd08ffe7f978d35f0e018d66bff6f67a2d1ac41134"} Nov 28 13:34:27 crc kubenswrapper[4631]: I1128 13:34:27.637136 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-6fdcddb789-msc6n" Nov 28 13:34:27 crc kubenswrapper[4631]: I1128 13:34:27.655010 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-57548d458d-h7npd" event={"ID":"0079ab2d-c5a8-45eb-b6e8-9574158de32f","Type":"ContainerStarted","Data":"7096fab1da055af46a7b1b3238811141a6c3854a000220e01badabd86e615dfd"} Nov 28 13:34:27 crc kubenswrapper[4631]: I1128 13:34:27.667532 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-656dcb59d4-wh9v9" event={"ID":"826f27ac-a938-4292-a357-ff54c4117986","Type":"ContainerStarted","Data":"aae25f8d55ac806a269768779b2d4851515e66a118fe594c591d544ff8d188d6"} Nov 28 13:34:27 crc kubenswrapper[4631]: I1128 13:34:27.679758 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-tpbhk" event={"ID":"bb908ace-bb62-4f55-8131-90e48da7585b","Type":"ContainerStarted","Data":"8b071b7f7dab305dd9ad6ec6915bdf0bb009649b3b9349f5eecfc27423786598"} Nov 28 13:34:27 crc kubenswrapper[4631]: I1128 13:34:27.692549 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5fcdb54b6b2f9vf" event={"ID":"ca18d816-8a63-4dcd-ac27-b1e7ad9ce4f6","Type":"ContainerStarted","Data":"95ba46919d836c72a880f5abf23b5cfa901c3a360283fec13f4df0eb93e7f6d6"} Nov 28 13:34:27 crc kubenswrapper[4631]: I1128 13:34:27.693309 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5fcdb54b6b2f9vf" Nov 28 13:34:27 crc kubenswrapper[4631]: I1128 13:34:27.701984 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-6b49bfffcd-5ks5f" event={"ID":"f3e12436-8a74-4443-94a4-3b9717105b05","Type":"ContainerStarted","Data":"f4efadfe09b6b091eaf5b6c6211616382778b2af8e820344cdb879e11e15fc09"} Nov 28 13:34:27 crc kubenswrapper[4631]: I1128 13:34:27.703249 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-6b49bfffcd-5ks5f" Nov 28 13:34:27 crc kubenswrapper[4631]: I1128 13:34:27.711670 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5fcdb54b6b2f9vf" Nov 28 13:34:27 crc kubenswrapper[4631]: I1128 13:34:27.721055 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-6fdcddb789-msc6n" podStartSLOduration=4.397590146 podStartE2EDuration="1m3.721036649s" podCreationTimestamp="2025-11-28 13:33:24 +0000 UTC" firstStartedPulling="2025-11-28 13:33:27.370112458 +0000 UTC m=+764.177415802" lastFinishedPulling="2025-11-28 13:34:26.693558971 +0000 UTC m=+823.500862305" observedRunningTime="2025-11-28 13:34:27.712034805 +0000 UTC m=+824.519338149" watchObservedRunningTime="2025-11-28 13:34:27.721036649 +0000 UTC m=+824.528339993" Nov 28 13:34:27 crc kubenswrapper[4631]: I1128 13:34:27.721404 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-66f4dd4bc7-m9t96" event={"ID":"cefc6050-9552-4dda-9864-39ac05626785","Type":"ContainerStarted","Data":"4dacca919a555ce95dcaca8fc6a1b843ff412e88be3dbeaa8174b045a144ce96"} Nov 28 13:34:27 crc kubenswrapper[4631]: I1128 13:34:27.722365 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-66f4dd4bc7-m9t96" Nov 28 13:34:27 crc kubenswrapper[4631]: I1128 13:34:27.731246 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-57988cc5b5-wqpdp" event={"ID":"05a6dcba-27b4-4b20-9275-f434b8943be9","Type":"ContainerStarted","Data":"27393f1b1825a61afe6c3d3d856984613c30b4aa97c2c599895ea29a4ee6cd59"} Nov 28 13:34:27 crc kubenswrapper[4631]: I1128 13:34:27.731737 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-57988cc5b5-wqpdp" Nov 28 13:34:27 crc kubenswrapper[4631]: I1128 13:34:27.741009 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-56897c768d-wkpql" event={"ID":"d09fac54-dd24-47df-83a1-23cae3102e7a","Type":"ContainerStarted","Data":"8bc266eb01838b3da5a187797ffdb9ba6cc1cb7f55dbd225f43a4258f0e0a8b6"} Nov 28 13:34:27 crc kubenswrapper[4631]: I1128 13:34:27.757921 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-64cdc6ff96-4qzz7" event={"ID":"ecfd67ad-5136-4e6e-9e41-3d933d10ea2c","Type":"ContainerStarted","Data":"cc96c741fb0a4c5ddbc55945787b94617b48957ac4c4d1071a012d0264bdfb5c"} Nov 28 13:34:27 crc kubenswrapper[4631]: I1128 13:34:27.771393 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zfd5w" event={"ID":"8206d20e-c1fb-46c2-b9ab-b13a38922d29","Type":"ContainerStarted","Data":"3f1891ed8a453bbaae662460bbfbdd9a1a431b08a3262e81cd5600286b6ad881"} Nov 28 13:34:27 crc kubenswrapper[4631]: I1128 13:34:27.781132 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-979dz" event={"ID":"fae1e74c-0fac-4e9c-bf7c-e91157b6ef2d","Type":"ContainerStarted","Data":"5e74429f2f439a94652ca770bb9d1a12008c99360a1d72272ee014d3197c3c81"} Nov 28 13:34:27 crc kubenswrapper[4631]: I1128 13:34:27.781200 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-979dz" event={"ID":"fae1e74c-0fac-4e9c-bf7c-e91157b6ef2d","Type":"ContainerStarted","Data":"67547c6fa195cea5bb34d439a64e3610e23dfb05bbdf57b4467ec77388c66d58"} Nov 28 13:34:27 crc kubenswrapper[4631]: I1128 13:34:27.790734 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-5d499bf58b-dck2d" event={"ID":"dbbf77e7-3c97-480f-83b1-0eb8bb8f386d","Type":"ContainerStarted","Data":"2a5a5005bbd5b2960ae596625600d3893021051e04525e1c27c792e811e2ac63"} Nov 28 13:34:27 crc kubenswrapper[4631]: I1128 13:34:27.794953 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-6b49bfffcd-5ks5f" podStartSLOduration=5.364475003 podStartE2EDuration="1m4.794928272s" podCreationTimestamp="2025-11-28 13:33:23 +0000 UTC" firstStartedPulling="2025-11-28 13:33:26.935651513 +0000 UTC m=+763.742954857" lastFinishedPulling="2025-11-28 13:34:26.366104782 +0000 UTC m=+823.173408126" observedRunningTime="2025-11-28 13:34:27.787768385 +0000 UTC m=+824.595071729" watchObservedRunningTime="2025-11-28 13:34:27.794928272 +0000 UTC m=+824.602231616" Nov 28 13:34:27 crc kubenswrapper[4631]: I1128 13:34:27.827665 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-d77b94747-2g446" event={"ID":"2d59108a-2b92-421b-beb8-157c2eca09ac","Type":"ContainerStarted","Data":"b64a70cb59053822d9d255d598301466ca2eea444b76870823adfc22c21dccb4"} Nov 28 13:34:27 crc kubenswrapper[4631]: I1128 13:34:27.828838 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-d77b94747-2g446" Nov 28 13:34:27 crc kubenswrapper[4631]: I1128 13:34:27.837401 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-d77b94747-2g446" Nov 28 13:34:27 crc kubenswrapper[4631]: I1128 13:34:27.848466 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5b77f656f-8pvq9" event={"ID":"9cc92c5c-285e-4604-bc60-7637938a78c8","Type":"ContainerStarted","Data":"e5efcb87deafbb539a2032998e70ed60c6684f04f16f013a4cdaf41ec869b4e4"} Nov 28 13:34:27 crc kubenswrapper[4631]: I1128 13:34:27.848873 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-5b77f656f-8pvq9" Nov 28 13:34:27 crc kubenswrapper[4631]: I1128 13:34:27.892773 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5fcdb54b6b2f9vf" podStartSLOduration=57.332008208 podStartE2EDuration="1m3.892749161s" podCreationTimestamp="2025-11-28 13:33:24 +0000 UTC" firstStartedPulling="2025-11-28 13:34:13.972722371 +0000 UTC m=+810.780025715" lastFinishedPulling="2025-11-28 13:34:20.533463324 +0000 UTC m=+817.340766668" observedRunningTime="2025-11-28 13:34:27.869989096 +0000 UTC m=+824.677292440" watchObservedRunningTime="2025-11-28 13:34:27.892749161 +0000 UTC m=+824.700052505" Nov 28 13:34:27 crc kubenswrapper[4631]: I1128 13:34:27.942731 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-d77b94747-2g446" podStartSLOduration=9.831687469 podStartE2EDuration="1m3.942709341s" podCreationTimestamp="2025-11-28 13:33:24 +0000 UTC" firstStartedPulling="2025-11-28 13:33:27.378928397 +0000 UTC m=+764.186231741" lastFinishedPulling="2025-11-28 13:34:21.489950269 +0000 UTC m=+818.297253613" observedRunningTime="2025-11-28 13:34:27.94062828 +0000 UTC m=+824.747931624" watchObservedRunningTime="2025-11-28 13:34:27.942709341 +0000 UTC m=+824.750012685" Nov 28 13:34:28 crc kubenswrapper[4631]: I1128 13:34:28.037226 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-5b77f656f-8pvq9" podStartSLOduration=5.089333724 podStartE2EDuration="1m5.037202638s" podCreationTimestamp="2025-11-28 13:33:23 +0000 UTC" firstStartedPulling="2025-11-28 13:33:26.414636868 +0000 UTC m=+763.221940212" lastFinishedPulling="2025-11-28 13:34:26.362505782 +0000 UTC m=+823.169809126" observedRunningTime="2025-11-28 13:34:28.033229699 +0000 UTC m=+824.840533063" watchObservedRunningTime="2025-11-28 13:34:28.037202638 +0000 UTC m=+824.844505972" Nov 28 13:34:28 crc kubenswrapper[4631]: I1128 13:34:28.094173 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-57988cc5b5-wqpdp" podStartSLOduration=4.963445233 podStartE2EDuration="1m4.094142981s" podCreationTimestamp="2025-11-28 13:33:24 +0000 UTC" firstStartedPulling="2025-11-28 13:33:27.234912202 +0000 UTC m=+764.042215536" lastFinishedPulling="2025-11-28 13:34:26.36560994 +0000 UTC m=+823.172913284" observedRunningTime="2025-11-28 13:34:28.08765861 +0000 UTC m=+824.894961954" watchObservedRunningTime="2025-11-28 13:34:28.094142981 +0000 UTC m=+824.901446325" Nov 28 13:34:28 crc kubenswrapper[4631]: I1128 13:34:28.177583 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-66f4dd4bc7-m9t96" podStartSLOduration=4.737430363 podStartE2EDuration="1m4.177559122s" podCreationTimestamp="2025-11-28 13:33:24 +0000 UTC" firstStartedPulling="2025-11-28 13:33:26.965450692 +0000 UTC m=+763.772754036" lastFinishedPulling="2025-11-28 13:34:26.405579441 +0000 UTC m=+823.212882795" observedRunningTime="2025-11-28 13:34:28.16942928 +0000 UTC m=+824.976732624" watchObservedRunningTime="2025-11-28 13:34:28.177559122 +0000 UTC m=+824.984862466" Nov 28 13:34:28 crc kubenswrapper[4631]: I1128 13:34:28.880919 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hcvqs" event={"ID":"bf6aea40-3632-4c7c-90ea-5584dd081ef2","Type":"ContainerStarted","Data":"531419fd624134668e4d7c293fc7f7ffd4c5f7db0a7eea6ec97bc0761ed5f95c"} Nov 28 13:34:28 crc kubenswrapper[4631]: I1128 13:34:28.891843 4631 generic.go:334] "Generic (PLEG): container finished" podID="fae1e74c-0fac-4e9c-bf7c-e91157b6ef2d" containerID="5e74429f2f439a94652ca770bb9d1a12008c99360a1d72272ee014d3197c3c81" exitCode=0 Nov 28 13:34:28 crc kubenswrapper[4631]: I1128 13:34:28.891985 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-979dz" event={"ID":"fae1e74c-0fac-4e9c-bf7c-e91157b6ef2d","Type":"ContainerDied","Data":"5e74429f2f439a94652ca770bb9d1a12008c99360a1d72272ee014d3197c3c81"} Nov 28 13:34:28 crc kubenswrapper[4631]: I1128 13:34:28.903056 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-589cbd6b5b-npdtc" event={"ID":"59ee25ad-f169-47b9-9d5c-f7a7a253a591","Type":"ContainerStarted","Data":"7a8fc4c684257e35f77803f9cd0e383f13fdceb17d80b9638008aba2b990bd27"} Nov 28 13:34:28 crc kubenswrapper[4631]: I1128 13:34:28.903995 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-589cbd6b5b-npdtc" Nov 28 13:34:28 crc kubenswrapper[4631]: I1128 13:34:28.913551 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-hcvqs" podStartSLOduration=14.611532807 podStartE2EDuration="21.913531032s" podCreationTimestamp="2025-11-28 13:34:07 +0000 UTC" firstStartedPulling="2025-11-28 13:34:19.882618387 +0000 UTC m=+816.689921731" lastFinishedPulling="2025-11-28 13:34:27.184616612 +0000 UTC m=+823.991919956" observedRunningTime="2025-11-28 13:34:28.910265661 +0000 UTC m=+825.717569025" watchObservedRunningTime="2025-11-28 13:34:28.913531032 +0000 UTC m=+825.720834376" Nov 28 13:34:28 crc kubenswrapper[4631]: I1128 13:34:28.916147 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-57548d458d-h7npd" event={"ID":"0079ab2d-c5a8-45eb-b6e8-9574158de32f","Type":"ContainerStarted","Data":"ecde1986859ed06db2bb32841544e71f9734dcdf6f5370e159a4e713c9511d29"} Nov 28 13:34:28 crc kubenswrapper[4631]: I1128 13:34:28.917034 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-57548d458d-h7npd" Nov 28 13:34:28 crc kubenswrapper[4631]: I1128 13:34:28.930995 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-56897c768d-wkpql" event={"ID":"d09fac54-dd24-47df-83a1-23cae3102e7a","Type":"ContainerStarted","Data":"5fdccb7aa4cf0cb1117e99fee1c691ab430e695f48ce1b4830ffdbb1c6333822"} Nov 28 13:34:28 crc kubenswrapper[4631]: I1128 13:34:28.931804 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-56897c768d-wkpql" Nov 28 13:34:28 crc kubenswrapper[4631]: I1128 13:34:28.938754 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-64cdc6ff96-4qzz7" Nov 28 13:34:28 crc kubenswrapper[4631]: I1128 13:34:28.960757 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-589cbd6b5b-npdtc" podStartSLOduration=6.064723597 podStartE2EDuration="1m5.960731614s" podCreationTimestamp="2025-11-28 13:33:23 +0000 UTC" firstStartedPulling="2025-11-28 13:33:26.749701416 +0000 UTC m=+763.557004760" lastFinishedPulling="2025-11-28 13:34:26.645709433 +0000 UTC m=+823.453012777" observedRunningTime="2025-11-28 13:34:28.95088641 +0000 UTC m=+825.758189744" watchObservedRunningTime="2025-11-28 13:34:28.960731614 +0000 UTC m=+825.768034958" Nov 28 13:34:28 crc kubenswrapper[4631]: I1128 13:34:28.963665 4631 generic.go:334] "Generic (PLEG): container finished" podID="8206d20e-c1fb-46c2-b9ab-b13a38922d29" containerID="3f1891ed8a453bbaae662460bbfbdd9a1a431b08a3262e81cd5600286b6ad881" exitCode=0 Nov 28 13:34:28 crc kubenswrapper[4631]: I1128 13:34:28.963793 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zfd5w" event={"ID":"8206d20e-c1fb-46c2-b9ab-b13a38922d29","Type":"ContainerDied","Data":"3f1891ed8a453bbaae662460bbfbdd9a1a431b08a3262e81cd5600286b6ad881"} Nov 28 13:34:28 crc kubenswrapper[4631]: I1128 13:34:28.971017 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-656dcb59d4-wh9v9" event={"ID":"826f27ac-a938-4292-a357-ff54c4117986","Type":"ContainerStarted","Data":"9788d40367ec5c28677c5724450f9a136117fea186bd9bba3d74cfe9862537f0"} Nov 28 13:34:28 crc kubenswrapper[4631]: I1128 13:34:28.971051 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-656dcb59d4-wh9v9" Nov 28 13:34:28 crc kubenswrapper[4631]: I1128 13:34:28.982933 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-57548d458d-h7npd" podStartSLOduration=53.630641056 podStartE2EDuration="1m5.982892724s" podCreationTimestamp="2025-11-28 13:33:23 +0000 UTC" firstStartedPulling="2025-11-28 13:34:13.972744153 +0000 UTC m=+810.780047497" lastFinishedPulling="2025-11-28 13:34:26.324995821 +0000 UTC m=+823.132299165" observedRunningTime="2025-11-28 13:34:28.974928496 +0000 UTC m=+825.782231840" watchObservedRunningTime="2025-11-28 13:34:28.982892724 +0000 UTC m=+825.790196078" Nov 28 13:34:29 crc kubenswrapper[4631]: I1128 13:34:29.012101 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-56897c768d-wkpql" podStartSLOduration=4.141423966 podStartE2EDuration="1m5.012076319s" podCreationTimestamp="2025-11-28 13:33:24 +0000 UTC" firstStartedPulling="2025-11-28 13:33:27.370104858 +0000 UTC m=+764.177408212" lastFinishedPulling="2025-11-28 13:34:28.240757221 +0000 UTC m=+825.048060565" observedRunningTime="2025-11-28 13:34:29.002740167 +0000 UTC m=+825.810043521" watchObservedRunningTime="2025-11-28 13:34:29.012076319 +0000 UTC m=+825.819379663" Nov 28 13:34:29 crc kubenswrapper[4631]: I1128 13:34:29.031676 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-64cdc6ff96-4qzz7" podStartSLOduration=3.876574251 podStartE2EDuration="1m5.031653425s" podCreationTimestamp="2025-11-28 13:33:24 +0000 UTC" firstStartedPulling="2025-11-28 13:33:27.240190072 +0000 UTC m=+764.047493416" lastFinishedPulling="2025-11-28 13:34:28.395269246 +0000 UTC m=+825.202572590" observedRunningTime="2025-11-28 13:34:29.029724377 +0000 UTC m=+825.837027731" watchObservedRunningTime="2025-11-28 13:34:29.031653425 +0000 UTC m=+825.838956769" Nov 28 13:34:29 crc kubenswrapper[4631]: I1128 13:34:29.070526 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-656dcb59d4-wh9v9" podStartSLOduration=3.866060281 podStartE2EDuration="1m5.07050615s" podCreationTimestamp="2025-11-28 13:33:24 +0000 UTC" firstStartedPulling="2025-11-28 13:33:27.203905682 +0000 UTC m=+764.011209026" lastFinishedPulling="2025-11-28 13:34:28.408351551 +0000 UTC m=+825.215654895" observedRunningTime="2025-11-28 13:34:29.067569216 +0000 UTC m=+825.874872560" watchObservedRunningTime="2025-11-28 13:34:29.07050615 +0000 UTC m=+825.877809494" Nov 28 13:34:29 crc kubenswrapper[4631]: I1128 13:34:29.980142 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-5d499bf58b-dck2d" event={"ID":"dbbf77e7-3c97-480f-83b1-0eb8bb8f386d","Type":"ContainerStarted","Data":"cbdb644659ec4a286f89cd3cf49139bb83c61154a8db8b628faf935c66f427a5"} Nov 28 13:34:29 crc kubenswrapper[4631]: I1128 13:34:29.980355 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-5d499bf58b-dck2d" Nov 28 13:34:29 crc kubenswrapper[4631]: I1128 13:34:29.984250 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-64cdc6ff96-4qzz7" event={"ID":"ecfd67ad-5136-4e6e-9e41-3d933d10ea2c","Type":"ContainerStarted","Data":"16d435988b44ba90e9539e11713dc473c8e312e9c6920450c035663ad0fc9fdc"} Nov 28 13:34:29 crc kubenswrapper[4631]: I1128 13:34:29.990010 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zfd5w" event={"ID":"8206d20e-c1fb-46c2-b9ab-b13a38922d29","Type":"ContainerStarted","Data":"56446a4e746416d657556e269c65833b9ea3fadb483431af90d6ad81ddfa0864"} Nov 28 13:34:29 crc kubenswrapper[4631]: I1128 13:34:29.992209 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-tpbhk" event={"ID":"bb908ace-bb62-4f55-8131-90e48da7585b","Type":"ContainerStarted","Data":"bb0b8ee82804a183e9f2b920901c227b0e093b60568154555bdced0b587b594e"} Nov 28 13:34:29 crc kubenswrapper[4631]: I1128 13:34:29.992243 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-tpbhk" Nov 28 13:34:30 crc kubenswrapper[4631]: I1128 13:34:30.008104 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-5d499bf58b-dck2d" podStartSLOduration=4.232389965 podStartE2EDuration="1m6.008084465s" podCreationTimestamp="2025-11-28 13:33:24 +0000 UTC" firstStartedPulling="2025-11-28 13:33:26.898758237 +0000 UTC m=+763.706061571" lastFinishedPulling="2025-11-28 13:34:28.674452727 +0000 UTC m=+825.481756071" observedRunningTime="2025-11-28 13:34:30.002873275 +0000 UTC m=+826.810176619" watchObservedRunningTime="2025-11-28 13:34:30.008084465 +0000 UTC m=+826.815387809" Nov 28 13:34:30 crc kubenswrapper[4631]: I1128 13:34:30.052208 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-zfd5w" podStartSLOduration=12.250158917 podStartE2EDuration="21.0521883s" podCreationTimestamp="2025-11-28 13:34:09 +0000 UTC" firstStartedPulling="2025-11-28 13:34:19.872528947 +0000 UTC m=+816.679832331" lastFinishedPulling="2025-11-28 13:34:28.67455837 +0000 UTC m=+825.481861714" observedRunningTime="2025-11-28 13:34:30.050162289 +0000 UTC m=+826.857465633" watchObservedRunningTime="2025-11-28 13:34:30.0521883 +0000 UTC m=+826.859491644" Nov 28 13:34:30 crc kubenswrapper[4631]: I1128 13:34:30.082228 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-tpbhk" podStartSLOduration=4.125782558 podStartE2EDuration="1m6.082204175s" podCreationTimestamp="2025-11-28 13:33:24 +0000 UTC" firstStartedPulling="2025-11-28 13:33:26.722798338 +0000 UTC m=+763.530101682" lastFinishedPulling="2025-11-28 13:34:28.679219945 +0000 UTC m=+825.486523299" observedRunningTime="2025-11-28 13:34:30.079844776 +0000 UTC m=+826.887148120" watchObservedRunningTime="2025-11-28 13:34:30.082204175 +0000 UTC m=+826.889507519" Nov 28 13:34:30 crc kubenswrapper[4631]: I1128 13:34:30.230736 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-zfd5w" Nov 28 13:34:30 crc kubenswrapper[4631]: I1128 13:34:30.230978 4631 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-zfd5w" Nov 28 13:34:31 crc kubenswrapper[4631]: I1128 13:34:31.280319 4631 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-marketplace-zfd5w" podUID="8206d20e-c1fb-46c2-b9ab-b13a38922d29" containerName="registry-server" probeResult="failure" output=< Nov 28 13:34:31 crc kubenswrapper[4631]: timeout: failed to connect service ":50051" within 1s Nov 28 13:34:31 crc kubenswrapper[4631]: > Nov 28 13:34:31 crc kubenswrapper[4631]: I1128 13:34:31.925750 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-854b9f67c6-hfb26" Nov 28 13:34:33 crc kubenswrapper[4631]: I1128 13:34:33.028834 4631 generic.go:334] "Generic (PLEG): container finished" podID="fae1e74c-0fac-4e9c-bf7c-e91157b6ef2d" containerID="883f4f46b8a6397e2ba95e686a17cd2247504413b480191bf3ecd245e071a84a" exitCode=0 Nov 28 13:34:33 crc kubenswrapper[4631]: I1128 13:34:33.028894 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-979dz" event={"ID":"fae1e74c-0fac-4e9c-bf7c-e91157b6ef2d","Type":"ContainerDied","Data":"883f4f46b8a6397e2ba95e686a17cd2247504413b480191bf3ecd245e071a84a"} Nov 28 13:34:34 crc kubenswrapper[4631]: I1128 13:34:34.451702 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-589cbd6b5b-npdtc" Nov 28 13:34:34 crc kubenswrapper[4631]: I1128 13:34:34.479421 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-5b77f656f-8pvq9" Nov 28 13:34:34 crc kubenswrapper[4631]: I1128 13:34:34.555497 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-5d494799bf-2mq7v" Nov 28 13:34:34 crc kubenswrapper[4631]: I1128 13:34:34.695106 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-tpbhk" Nov 28 13:34:34 crc kubenswrapper[4631]: I1128 13:34:34.734793 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-6b49bfffcd-5ks5f" Nov 28 13:34:34 crc kubenswrapper[4631]: I1128 13:34:34.833251 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-66f4dd4bc7-m9t96" Nov 28 13:34:34 crc kubenswrapper[4631]: I1128 13:34:34.874461 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-64cdc6ff96-4qzz7" Nov 28 13:34:34 crc kubenswrapper[4631]: I1128 13:34:34.892415 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-5d499bf58b-dck2d" Nov 28 13:34:35 crc kubenswrapper[4631]: I1128 13:34:35.012596 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-6fdcddb789-msc6n" Nov 28 13:34:35 crc kubenswrapper[4631]: I1128 13:34:35.016113 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-56897c768d-wkpql" Nov 28 13:34:35 crc kubenswrapper[4631]: I1128 13:34:35.046615 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-979dz" event={"ID":"fae1e74c-0fac-4e9c-bf7c-e91157b6ef2d","Type":"ContainerStarted","Data":"4c7bcc4d636f1c48a6bf4c2b1ebb7d39c60d282f43f08a9eac088a2cd35b530f"} Nov 28 13:34:35 crc kubenswrapper[4631]: I1128 13:34:35.052086 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-57988cc5b5-wqpdp" Nov 28 13:34:35 crc kubenswrapper[4631]: I1128 13:34:35.092642 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-979dz" podStartSLOduration=6.557635248 podStartE2EDuration="13.092622521s" podCreationTimestamp="2025-11-28 13:34:22 +0000 UTC" firstStartedPulling="2025-11-28 13:34:27.784140945 +0000 UTC m=+824.591444289" lastFinishedPulling="2025-11-28 13:34:34.319128218 +0000 UTC m=+831.126431562" observedRunningTime="2025-11-28 13:34:35.089606385 +0000 UTC m=+831.896909749" watchObservedRunningTime="2025-11-28 13:34:35.092622521 +0000 UTC m=+831.899925865" Nov 28 13:34:35 crc kubenswrapper[4631]: I1128 13:34:35.578158 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-656dcb59d4-wh9v9" Nov 28 13:34:37 crc kubenswrapper[4631]: I1128 13:34:37.861058 4631 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-hcvqs" Nov 28 13:34:37 crc kubenswrapper[4631]: I1128 13:34:37.862861 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-hcvqs" Nov 28 13:34:38 crc kubenswrapper[4631]: I1128 13:34:38.914868 4631 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-hcvqs" podUID="bf6aea40-3632-4c7c-90ea-5584dd081ef2" containerName="registry-server" probeResult="failure" output=< Nov 28 13:34:38 crc kubenswrapper[4631]: timeout: failed to connect service ":50051" within 1s Nov 28 13:34:38 crc kubenswrapper[4631]: > Nov 28 13:34:39 crc kubenswrapper[4631]: I1128 13:34:39.085986 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-nrpfx" event={"ID":"10a95153-0898-4016-92a6-d56633555348","Type":"ContainerStarted","Data":"d444385368013a75196cff87b5c0b82b4d0a2e2336fabaeb27a435540d824a51"} Nov 28 13:34:39 crc kubenswrapper[4631]: I1128 13:34:39.104590 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-nrpfx" podStartSLOduration=3.50287563 podStartE2EDuration="1m14.104563218s" podCreationTimestamp="2025-11-28 13:33:25 +0000 UTC" firstStartedPulling="2025-11-28 13:33:27.422414017 +0000 UTC m=+764.229717361" lastFinishedPulling="2025-11-28 13:34:38.024101605 +0000 UTC m=+834.831404949" observedRunningTime="2025-11-28 13:34:39.104167759 +0000 UTC m=+835.911471113" watchObservedRunningTime="2025-11-28 13:34:39.104563218 +0000 UTC m=+835.911866582" Nov 28 13:34:40 crc kubenswrapper[4631]: I1128 13:34:40.240271 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-57548d458d-h7npd" Nov 28 13:34:40 crc kubenswrapper[4631]: I1128 13:34:40.299891 4631 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-zfd5w" Nov 28 13:34:40 crc kubenswrapper[4631]: I1128 13:34:40.359693 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-zfd5w" Nov 28 13:34:40 crc kubenswrapper[4631]: I1128 13:34:40.544559 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-zfd5w"] Nov 28 13:34:42 crc kubenswrapper[4631]: I1128 13:34:42.108570 4631 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-zfd5w" podUID="8206d20e-c1fb-46c2-b9ab-b13a38922d29" containerName="registry-server" containerID="cri-o://56446a4e746416d657556e269c65833b9ea3fadb483431af90d6ad81ddfa0864" gracePeriod=2 Nov 28 13:34:42 crc kubenswrapper[4631]: I1128 13:34:42.578094 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zfd5w" Nov 28 13:34:42 crc kubenswrapper[4631]: I1128 13:34:42.718858 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h87rf\" (UniqueName: \"kubernetes.io/projected/8206d20e-c1fb-46c2-b9ab-b13a38922d29-kube-api-access-h87rf\") pod \"8206d20e-c1fb-46c2-b9ab-b13a38922d29\" (UID: \"8206d20e-c1fb-46c2-b9ab-b13a38922d29\") " Nov 28 13:34:42 crc kubenswrapper[4631]: I1128 13:34:42.719007 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8206d20e-c1fb-46c2-b9ab-b13a38922d29-utilities\") pod \"8206d20e-c1fb-46c2-b9ab-b13a38922d29\" (UID: \"8206d20e-c1fb-46c2-b9ab-b13a38922d29\") " Nov 28 13:34:42 crc kubenswrapper[4631]: I1128 13:34:42.719058 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8206d20e-c1fb-46c2-b9ab-b13a38922d29-catalog-content\") pod \"8206d20e-c1fb-46c2-b9ab-b13a38922d29\" (UID: \"8206d20e-c1fb-46c2-b9ab-b13a38922d29\") " Nov 28 13:34:42 crc kubenswrapper[4631]: I1128 13:34:42.720107 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8206d20e-c1fb-46c2-b9ab-b13a38922d29-utilities" (OuterVolumeSpecName: "utilities") pod "8206d20e-c1fb-46c2-b9ab-b13a38922d29" (UID: "8206d20e-c1fb-46c2-b9ab-b13a38922d29"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 13:34:42 crc kubenswrapper[4631]: I1128 13:34:42.726782 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8206d20e-c1fb-46c2-b9ab-b13a38922d29-kube-api-access-h87rf" (OuterVolumeSpecName: "kube-api-access-h87rf") pod "8206d20e-c1fb-46c2-b9ab-b13a38922d29" (UID: "8206d20e-c1fb-46c2-b9ab-b13a38922d29"). InnerVolumeSpecName "kube-api-access-h87rf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:34:42 crc kubenswrapper[4631]: I1128 13:34:42.737877 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8206d20e-c1fb-46c2-b9ab-b13a38922d29-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8206d20e-c1fb-46c2-b9ab-b13a38922d29" (UID: "8206d20e-c1fb-46c2-b9ab-b13a38922d29"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 13:34:42 crc kubenswrapper[4631]: I1128 13:34:42.821068 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h87rf\" (UniqueName: \"kubernetes.io/projected/8206d20e-c1fb-46c2-b9ab-b13a38922d29-kube-api-access-h87rf\") on node \"crc\" DevicePath \"\"" Nov 28 13:34:42 crc kubenswrapper[4631]: I1128 13:34:42.821107 4631 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8206d20e-c1fb-46c2-b9ab-b13a38922d29-utilities\") on node \"crc\" DevicePath \"\"" Nov 28 13:34:42 crc kubenswrapper[4631]: I1128 13:34:42.821118 4631 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8206d20e-c1fb-46c2-b9ab-b13a38922d29-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 28 13:34:43 crc kubenswrapper[4631]: I1128 13:34:43.120083 4631 generic.go:334] "Generic (PLEG): container finished" podID="8206d20e-c1fb-46c2-b9ab-b13a38922d29" containerID="56446a4e746416d657556e269c65833b9ea3fadb483431af90d6ad81ddfa0864" exitCode=0 Nov 28 13:34:43 crc kubenswrapper[4631]: I1128 13:34:43.120537 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zfd5w" event={"ID":"8206d20e-c1fb-46c2-b9ab-b13a38922d29","Type":"ContainerDied","Data":"56446a4e746416d657556e269c65833b9ea3fadb483431af90d6ad81ddfa0864"} Nov 28 13:34:43 crc kubenswrapper[4631]: I1128 13:34:43.120617 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zfd5w" Nov 28 13:34:43 crc kubenswrapper[4631]: I1128 13:34:43.120644 4631 scope.go:117] "RemoveContainer" containerID="56446a4e746416d657556e269c65833b9ea3fadb483431af90d6ad81ddfa0864" Nov 28 13:34:43 crc kubenswrapper[4631]: I1128 13:34:43.120625 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zfd5w" event={"ID":"8206d20e-c1fb-46c2-b9ab-b13a38922d29","Type":"ContainerDied","Data":"6c49094ec9c124d09271e55d9afae45554dcc2a31a1a915cd4c670db01a6b489"} Nov 28 13:34:43 crc kubenswrapper[4631]: I1128 13:34:43.150996 4631 scope.go:117] "RemoveContainer" containerID="3f1891ed8a453bbaae662460bbfbdd9a1a431b08a3262e81cd5600286b6ad881" Nov 28 13:34:43 crc kubenswrapper[4631]: I1128 13:34:43.177366 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-zfd5w"] Nov 28 13:34:43 crc kubenswrapper[4631]: I1128 13:34:43.179552 4631 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-zfd5w"] Nov 28 13:34:43 crc kubenswrapper[4631]: I1128 13:34:43.188255 4631 scope.go:117] "RemoveContainer" containerID="9200a264305d5c2092cb9c6e312b5db7863ca4b3d226a9427048aad7767fcb88" Nov 28 13:34:43 crc kubenswrapper[4631]: I1128 13:34:43.206513 4631 scope.go:117] "RemoveContainer" containerID="56446a4e746416d657556e269c65833b9ea3fadb483431af90d6ad81ddfa0864" Nov 28 13:34:43 crc kubenswrapper[4631]: E1128 13:34:43.207050 4631 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"56446a4e746416d657556e269c65833b9ea3fadb483431af90d6ad81ddfa0864\": container with ID starting with 56446a4e746416d657556e269c65833b9ea3fadb483431af90d6ad81ddfa0864 not found: ID does not exist" containerID="56446a4e746416d657556e269c65833b9ea3fadb483431af90d6ad81ddfa0864" Nov 28 13:34:43 crc kubenswrapper[4631]: I1128 13:34:43.207119 4631 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"56446a4e746416d657556e269c65833b9ea3fadb483431af90d6ad81ddfa0864"} err="failed to get container status \"56446a4e746416d657556e269c65833b9ea3fadb483431af90d6ad81ddfa0864\": rpc error: code = NotFound desc = could not find container \"56446a4e746416d657556e269c65833b9ea3fadb483431af90d6ad81ddfa0864\": container with ID starting with 56446a4e746416d657556e269c65833b9ea3fadb483431af90d6ad81ddfa0864 not found: ID does not exist" Nov 28 13:34:43 crc kubenswrapper[4631]: I1128 13:34:43.207166 4631 scope.go:117] "RemoveContainer" containerID="3f1891ed8a453bbaae662460bbfbdd9a1a431b08a3262e81cd5600286b6ad881" Nov 28 13:34:43 crc kubenswrapper[4631]: E1128 13:34:43.207610 4631 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3f1891ed8a453bbaae662460bbfbdd9a1a431b08a3262e81cd5600286b6ad881\": container with ID starting with 3f1891ed8a453bbaae662460bbfbdd9a1a431b08a3262e81cd5600286b6ad881 not found: ID does not exist" containerID="3f1891ed8a453bbaae662460bbfbdd9a1a431b08a3262e81cd5600286b6ad881" Nov 28 13:34:43 crc kubenswrapper[4631]: I1128 13:34:43.207648 4631 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3f1891ed8a453bbaae662460bbfbdd9a1a431b08a3262e81cd5600286b6ad881"} err="failed to get container status \"3f1891ed8a453bbaae662460bbfbdd9a1a431b08a3262e81cd5600286b6ad881\": rpc error: code = NotFound desc = could not find container \"3f1891ed8a453bbaae662460bbfbdd9a1a431b08a3262e81cd5600286b6ad881\": container with ID starting with 3f1891ed8a453bbaae662460bbfbdd9a1a431b08a3262e81cd5600286b6ad881 not found: ID does not exist" Nov 28 13:34:43 crc kubenswrapper[4631]: I1128 13:34:43.207671 4631 scope.go:117] "RemoveContainer" containerID="9200a264305d5c2092cb9c6e312b5db7863ca4b3d226a9427048aad7767fcb88" Nov 28 13:34:43 crc kubenswrapper[4631]: E1128 13:34:43.207939 4631 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9200a264305d5c2092cb9c6e312b5db7863ca4b3d226a9427048aad7767fcb88\": container with ID starting with 9200a264305d5c2092cb9c6e312b5db7863ca4b3d226a9427048aad7767fcb88 not found: ID does not exist" containerID="9200a264305d5c2092cb9c6e312b5db7863ca4b3d226a9427048aad7767fcb88" Nov 28 13:34:43 crc kubenswrapper[4631]: I1128 13:34:43.207975 4631 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9200a264305d5c2092cb9c6e312b5db7863ca4b3d226a9427048aad7767fcb88"} err="failed to get container status \"9200a264305d5c2092cb9c6e312b5db7863ca4b3d226a9427048aad7767fcb88\": rpc error: code = NotFound desc = could not find container \"9200a264305d5c2092cb9c6e312b5db7863ca4b3d226a9427048aad7767fcb88\": container with ID starting with 9200a264305d5c2092cb9c6e312b5db7863ca4b3d226a9427048aad7767fcb88 not found: ID does not exist" Nov 28 13:34:43 crc kubenswrapper[4631]: I1128 13:34:43.468144 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-979dz" Nov 28 13:34:43 crc kubenswrapper[4631]: I1128 13:34:43.468210 4631 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-979dz" Nov 28 13:34:43 crc kubenswrapper[4631]: I1128 13:34:43.528257 4631 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8206d20e-c1fb-46c2-b9ab-b13a38922d29" path="/var/lib/kubelet/pods/8206d20e-c1fb-46c2-b9ab-b13a38922d29/volumes" Nov 28 13:34:43 crc kubenswrapper[4631]: I1128 13:34:43.529212 4631 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-979dz" Nov 28 13:34:44 crc kubenswrapper[4631]: I1128 13:34:44.184501 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-979dz" Nov 28 13:34:45 crc kubenswrapper[4631]: I1128 13:34:45.946682 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-979dz"] Nov 28 13:34:46 crc kubenswrapper[4631]: I1128 13:34:46.145170 4631 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-979dz" podUID="fae1e74c-0fac-4e9c-bf7c-e91157b6ef2d" containerName="registry-server" containerID="cri-o://4c7bcc4d636f1c48a6bf4c2b1ebb7d39c60d282f43f08a9eac088a2cd35b530f" gracePeriod=2 Nov 28 13:34:46 crc kubenswrapper[4631]: I1128 13:34:46.590042 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-979dz" Nov 28 13:34:46 crc kubenswrapper[4631]: I1128 13:34:46.681018 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fae1e74c-0fac-4e9c-bf7c-e91157b6ef2d-utilities\") pod \"fae1e74c-0fac-4e9c-bf7c-e91157b6ef2d\" (UID: \"fae1e74c-0fac-4e9c-bf7c-e91157b6ef2d\") " Nov 28 13:34:46 crc kubenswrapper[4631]: I1128 13:34:46.681231 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j8kvm\" (UniqueName: \"kubernetes.io/projected/fae1e74c-0fac-4e9c-bf7c-e91157b6ef2d-kube-api-access-j8kvm\") pod \"fae1e74c-0fac-4e9c-bf7c-e91157b6ef2d\" (UID: \"fae1e74c-0fac-4e9c-bf7c-e91157b6ef2d\") " Nov 28 13:34:46 crc kubenswrapper[4631]: I1128 13:34:46.681320 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fae1e74c-0fac-4e9c-bf7c-e91157b6ef2d-catalog-content\") pod \"fae1e74c-0fac-4e9c-bf7c-e91157b6ef2d\" (UID: \"fae1e74c-0fac-4e9c-bf7c-e91157b6ef2d\") " Nov 28 13:34:46 crc kubenswrapper[4631]: I1128 13:34:46.681881 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fae1e74c-0fac-4e9c-bf7c-e91157b6ef2d-utilities" (OuterVolumeSpecName: "utilities") pod "fae1e74c-0fac-4e9c-bf7c-e91157b6ef2d" (UID: "fae1e74c-0fac-4e9c-bf7c-e91157b6ef2d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 13:34:46 crc kubenswrapper[4631]: I1128 13:34:46.693223 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fae1e74c-0fac-4e9c-bf7c-e91157b6ef2d-kube-api-access-j8kvm" (OuterVolumeSpecName: "kube-api-access-j8kvm") pod "fae1e74c-0fac-4e9c-bf7c-e91157b6ef2d" (UID: "fae1e74c-0fac-4e9c-bf7c-e91157b6ef2d"). InnerVolumeSpecName "kube-api-access-j8kvm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:34:46 crc kubenswrapper[4631]: I1128 13:34:46.735223 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fae1e74c-0fac-4e9c-bf7c-e91157b6ef2d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "fae1e74c-0fac-4e9c-bf7c-e91157b6ef2d" (UID: "fae1e74c-0fac-4e9c-bf7c-e91157b6ef2d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 13:34:46 crc kubenswrapper[4631]: I1128 13:34:46.783510 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j8kvm\" (UniqueName: \"kubernetes.io/projected/fae1e74c-0fac-4e9c-bf7c-e91157b6ef2d-kube-api-access-j8kvm\") on node \"crc\" DevicePath \"\"" Nov 28 13:34:46 crc kubenswrapper[4631]: I1128 13:34:46.783561 4631 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fae1e74c-0fac-4e9c-bf7c-e91157b6ef2d-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 28 13:34:46 crc kubenswrapper[4631]: I1128 13:34:46.783577 4631 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fae1e74c-0fac-4e9c-bf7c-e91157b6ef2d-utilities\") on node \"crc\" DevicePath \"\"" Nov 28 13:34:47 crc kubenswrapper[4631]: I1128 13:34:47.173547 4631 generic.go:334] "Generic (PLEG): container finished" podID="fae1e74c-0fac-4e9c-bf7c-e91157b6ef2d" containerID="4c7bcc4d636f1c48a6bf4c2b1ebb7d39c60d282f43f08a9eac088a2cd35b530f" exitCode=0 Nov 28 13:34:47 crc kubenswrapper[4631]: I1128 13:34:47.174062 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-979dz" event={"ID":"fae1e74c-0fac-4e9c-bf7c-e91157b6ef2d","Type":"ContainerDied","Data":"4c7bcc4d636f1c48a6bf4c2b1ebb7d39c60d282f43f08a9eac088a2cd35b530f"} Nov 28 13:34:47 crc kubenswrapper[4631]: I1128 13:34:47.174116 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-979dz" event={"ID":"fae1e74c-0fac-4e9c-bf7c-e91157b6ef2d","Type":"ContainerDied","Data":"67547c6fa195cea5bb34d439a64e3610e23dfb05bbdf57b4467ec77388c66d58"} Nov 28 13:34:47 crc kubenswrapper[4631]: I1128 13:34:47.174167 4631 scope.go:117] "RemoveContainer" containerID="4c7bcc4d636f1c48a6bf4c2b1ebb7d39c60d282f43f08a9eac088a2cd35b530f" Nov 28 13:34:47 crc kubenswrapper[4631]: I1128 13:34:47.174500 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-979dz" Nov 28 13:34:47 crc kubenswrapper[4631]: I1128 13:34:47.221717 4631 scope.go:117] "RemoveContainer" containerID="883f4f46b8a6397e2ba95e686a17cd2247504413b480191bf3ecd245e071a84a" Nov 28 13:34:47 crc kubenswrapper[4631]: I1128 13:34:47.256332 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-979dz"] Nov 28 13:34:47 crc kubenswrapper[4631]: I1128 13:34:47.256383 4631 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-979dz"] Nov 28 13:34:47 crc kubenswrapper[4631]: I1128 13:34:47.261510 4631 scope.go:117] "RemoveContainer" containerID="5e74429f2f439a94652ca770bb9d1a12008c99360a1d72272ee014d3197c3c81" Nov 28 13:34:47 crc kubenswrapper[4631]: I1128 13:34:47.293531 4631 scope.go:117] "RemoveContainer" containerID="4c7bcc4d636f1c48a6bf4c2b1ebb7d39c60d282f43f08a9eac088a2cd35b530f" Nov 28 13:34:47 crc kubenswrapper[4631]: E1128 13:34:47.294128 4631 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4c7bcc4d636f1c48a6bf4c2b1ebb7d39c60d282f43f08a9eac088a2cd35b530f\": container with ID starting with 4c7bcc4d636f1c48a6bf4c2b1ebb7d39c60d282f43f08a9eac088a2cd35b530f not found: ID does not exist" containerID="4c7bcc4d636f1c48a6bf4c2b1ebb7d39c60d282f43f08a9eac088a2cd35b530f" Nov 28 13:34:47 crc kubenswrapper[4631]: I1128 13:34:47.294180 4631 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4c7bcc4d636f1c48a6bf4c2b1ebb7d39c60d282f43f08a9eac088a2cd35b530f"} err="failed to get container status \"4c7bcc4d636f1c48a6bf4c2b1ebb7d39c60d282f43f08a9eac088a2cd35b530f\": rpc error: code = NotFound desc = could not find container \"4c7bcc4d636f1c48a6bf4c2b1ebb7d39c60d282f43f08a9eac088a2cd35b530f\": container with ID starting with 4c7bcc4d636f1c48a6bf4c2b1ebb7d39c60d282f43f08a9eac088a2cd35b530f not found: ID does not exist" Nov 28 13:34:47 crc kubenswrapper[4631]: I1128 13:34:47.294243 4631 scope.go:117] "RemoveContainer" containerID="883f4f46b8a6397e2ba95e686a17cd2247504413b480191bf3ecd245e071a84a" Nov 28 13:34:47 crc kubenswrapper[4631]: E1128 13:34:47.294563 4631 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"883f4f46b8a6397e2ba95e686a17cd2247504413b480191bf3ecd245e071a84a\": container with ID starting with 883f4f46b8a6397e2ba95e686a17cd2247504413b480191bf3ecd245e071a84a not found: ID does not exist" containerID="883f4f46b8a6397e2ba95e686a17cd2247504413b480191bf3ecd245e071a84a" Nov 28 13:34:47 crc kubenswrapper[4631]: I1128 13:34:47.294686 4631 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"883f4f46b8a6397e2ba95e686a17cd2247504413b480191bf3ecd245e071a84a"} err="failed to get container status \"883f4f46b8a6397e2ba95e686a17cd2247504413b480191bf3ecd245e071a84a\": rpc error: code = NotFound desc = could not find container \"883f4f46b8a6397e2ba95e686a17cd2247504413b480191bf3ecd245e071a84a\": container with ID starting with 883f4f46b8a6397e2ba95e686a17cd2247504413b480191bf3ecd245e071a84a not found: ID does not exist" Nov 28 13:34:47 crc kubenswrapper[4631]: I1128 13:34:47.294701 4631 scope.go:117] "RemoveContainer" containerID="5e74429f2f439a94652ca770bb9d1a12008c99360a1d72272ee014d3197c3c81" Nov 28 13:34:47 crc kubenswrapper[4631]: E1128 13:34:47.294918 4631 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5e74429f2f439a94652ca770bb9d1a12008c99360a1d72272ee014d3197c3c81\": container with ID starting with 5e74429f2f439a94652ca770bb9d1a12008c99360a1d72272ee014d3197c3c81 not found: ID does not exist" containerID="5e74429f2f439a94652ca770bb9d1a12008c99360a1d72272ee014d3197c3c81" Nov 28 13:34:47 crc kubenswrapper[4631]: I1128 13:34:47.294937 4631 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5e74429f2f439a94652ca770bb9d1a12008c99360a1d72272ee014d3197c3c81"} err="failed to get container status \"5e74429f2f439a94652ca770bb9d1a12008c99360a1d72272ee014d3197c3c81\": rpc error: code = NotFound desc = could not find container \"5e74429f2f439a94652ca770bb9d1a12008c99360a1d72272ee014d3197c3c81\": container with ID starting with 5e74429f2f439a94652ca770bb9d1a12008c99360a1d72272ee014d3197c3c81 not found: ID does not exist" Nov 28 13:34:47 crc kubenswrapper[4631]: I1128 13:34:47.523246 4631 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fae1e74c-0fac-4e9c-bf7c-e91157b6ef2d" path="/var/lib/kubelet/pods/fae1e74c-0fac-4e9c-bf7c-e91157b6ef2d/volumes" Nov 28 13:34:47 crc kubenswrapper[4631]: I1128 13:34:47.942809 4631 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-hcvqs" Nov 28 13:34:48 crc kubenswrapper[4631]: I1128 13:34:48.018116 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-hcvqs" Nov 28 13:34:50 crc kubenswrapper[4631]: I1128 13:34:50.156433 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-hcvqs"] Nov 28 13:34:50 crc kubenswrapper[4631]: I1128 13:34:50.157269 4631 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-hcvqs" podUID="bf6aea40-3632-4c7c-90ea-5584dd081ef2" containerName="registry-server" containerID="cri-o://531419fd624134668e4d7c293fc7f7ffd4c5f7db0a7eea6ec97bc0761ed5f95c" gracePeriod=2 Nov 28 13:34:50 crc kubenswrapper[4631]: I1128 13:34:50.571578 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hcvqs" Nov 28 13:34:50 crc kubenswrapper[4631]: I1128 13:34:50.655356 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bf6aea40-3632-4c7c-90ea-5584dd081ef2-catalog-content\") pod \"bf6aea40-3632-4c7c-90ea-5584dd081ef2\" (UID: \"bf6aea40-3632-4c7c-90ea-5584dd081ef2\") " Nov 28 13:34:50 crc kubenswrapper[4631]: I1128 13:34:50.655454 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bf6aea40-3632-4c7c-90ea-5584dd081ef2-utilities\") pod \"bf6aea40-3632-4c7c-90ea-5584dd081ef2\" (UID: \"bf6aea40-3632-4c7c-90ea-5584dd081ef2\") " Nov 28 13:34:50 crc kubenswrapper[4631]: I1128 13:34:50.655593 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-76t8n\" (UniqueName: \"kubernetes.io/projected/bf6aea40-3632-4c7c-90ea-5584dd081ef2-kube-api-access-76t8n\") pod \"bf6aea40-3632-4c7c-90ea-5584dd081ef2\" (UID: \"bf6aea40-3632-4c7c-90ea-5584dd081ef2\") " Nov 28 13:34:50 crc kubenswrapper[4631]: I1128 13:34:50.657015 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bf6aea40-3632-4c7c-90ea-5584dd081ef2-utilities" (OuterVolumeSpecName: "utilities") pod "bf6aea40-3632-4c7c-90ea-5584dd081ef2" (UID: "bf6aea40-3632-4c7c-90ea-5584dd081ef2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 13:34:50 crc kubenswrapper[4631]: I1128 13:34:50.664651 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf6aea40-3632-4c7c-90ea-5584dd081ef2-kube-api-access-76t8n" (OuterVolumeSpecName: "kube-api-access-76t8n") pod "bf6aea40-3632-4c7c-90ea-5584dd081ef2" (UID: "bf6aea40-3632-4c7c-90ea-5584dd081ef2"). InnerVolumeSpecName "kube-api-access-76t8n". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:34:50 crc kubenswrapper[4631]: I1128 13:34:50.766581 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bf6aea40-3632-4c7c-90ea-5584dd081ef2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "bf6aea40-3632-4c7c-90ea-5584dd081ef2" (UID: "bf6aea40-3632-4c7c-90ea-5584dd081ef2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 13:34:50 crc kubenswrapper[4631]: I1128 13:34:50.768703 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bf6aea40-3632-4c7c-90ea-5584dd081ef2-catalog-content\") pod \"bf6aea40-3632-4c7c-90ea-5584dd081ef2\" (UID: \"bf6aea40-3632-4c7c-90ea-5584dd081ef2\") " Nov 28 13:34:50 crc kubenswrapper[4631]: W1128 13:34:50.768941 4631 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/bf6aea40-3632-4c7c-90ea-5584dd081ef2/volumes/kubernetes.io~empty-dir/catalog-content Nov 28 13:34:50 crc kubenswrapper[4631]: I1128 13:34:50.769000 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bf6aea40-3632-4c7c-90ea-5584dd081ef2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "bf6aea40-3632-4c7c-90ea-5584dd081ef2" (UID: "bf6aea40-3632-4c7c-90ea-5584dd081ef2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 13:34:50 crc kubenswrapper[4631]: I1128 13:34:50.769811 4631 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bf6aea40-3632-4c7c-90ea-5584dd081ef2-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 28 13:34:50 crc kubenswrapper[4631]: I1128 13:34:50.769843 4631 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bf6aea40-3632-4c7c-90ea-5584dd081ef2-utilities\") on node \"crc\" DevicePath \"\"" Nov 28 13:34:50 crc kubenswrapper[4631]: I1128 13:34:50.769855 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-76t8n\" (UniqueName: \"kubernetes.io/projected/bf6aea40-3632-4c7c-90ea-5584dd081ef2-kube-api-access-76t8n\") on node \"crc\" DevicePath \"\"" Nov 28 13:34:51 crc kubenswrapper[4631]: I1128 13:34:51.208260 4631 generic.go:334] "Generic (PLEG): container finished" podID="bf6aea40-3632-4c7c-90ea-5584dd081ef2" containerID="531419fd624134668e4d7c293fc7f7ffd4c5f7db0a7eea6ec97bc0761ed5f95c" exitCode=0 Nov 28 13:34:51 crc kubenswrapper[4631]: I1128 13:34:51.208339 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hcvqs" event={"ID":"bf6aea40-3632-4c7c-90ea-5584dd081ef2","Type":"ContainerDied","Data":"531419fd624134668e4d7c293fc7f7ffd4c5f7db0a7eea6ec97bc0761ed5f95c"} Nov 28 13:34:51 crc kubenswrapper[4631]: I1128 13:34:51.208386 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hcvqs" event={"ID":"bf6aea40-3632-4c7c-90ea-5584dd081ef2","Type":"ContainerDied","Data":"6e5c15dab80a930be599aa068a59bc4e2407286fc1851689ae98fd0c777da5df"} Nov 28 13:34:51 crc kubenswrapper[4631]: I1128 13:34:51.208415 4631 scope.go:117] "RemoveContainer" containerID="531419fd624134668e4d7c293fc7f7ffd4c5f7db0a7eea6ec97bc0761ed5f95c" Nov 28 13:34:51 crc kubenswrapper[4631]: I1128 13:34:51.209848 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hcvqs" Nov 28 13:34:51 crc kubenswrapper[4631]: I1128 13:34:51.228477 4631 scope.go:117] "RemoveContainer" containerID="61f33b53d073e336d54579efe11721c6af8ba587cb0970340e04cc2357d84a82" Nov 28 13:34:51 crc kubenswrapper[4631]: I1128 13:34:51.239166 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-hcvqs"] Nov 28 13:34:51 crc kubenswrapper[4631]: I1128 13:34:51.259070 4631 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-hcvqs"] Nov 28 13:34:51 crc kubenswrapper[4631]: I1128 13:34:51.265310 4631 scope.go:117] "RemoveContainer" containerID="128147b0b653091336f64aea110f6a9a5219757366275397a467ddcdb05f1ccf" Nov 28 13:34:51 crc kubenswrapper[4631]: I1128 13:34:51.298138 4631 scope.go:117] "RemoveContainer" containerID="531419fd624134668e4d7c293fc7f7ffd4c5f7db0a7eea6ec97bc0761ed5f95c" Nov 28 13:34:51 crc kubenswrapper[4631]: E1128 13:34:51.298870 4631 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"531419fd624134668e4d7c293fc7f7ffd4c5f7db0a7eea6ec97bc0761ed5f95c\": container with ID starting with 531419fd624134668e4d7c293fc7f7ffd4c5f7db0a7eea6ec97bc0761ed5f95c not found: ID does not exist" containerID="531419fd624134668e4d7c293fc7f7ffd4c5f7db0a7eea6ec97bc0761ed5f95c" Nov 28 13:34:51 crc kubenswrapper[4631]: I1128 13:34:51.298914 4631 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"531419fd624134668e4d7c293fc7f7ffd4c5f7db0a7eea6ec97bc0761ed5f95c"} err="failed to get container status \"531419fd624134668e4d7c293fc7f7ffd4c5f7db0a7eea6ec97bc0761ed5f95c\": rpc error: code = NotFound desc = could not find container \"531419fd624134668e4d7c293fc7f7ffd4c5f7db0a7eea6ec97bc0761ed5f95c\": container with ID starting with 531419fd624134668e4d7c293fc7f7ffd4c5f7db0a7eea6ec97bc0761ed5f95c not found: ID does not exist" Nov 28 13:34:51 crc kubenswrapper[4631]: I1128 13:34:51.298952 4631 scope.go:117] "RemoveContainer" containerID="61f33b53d073e336d54579efe11721c6af8ba587cb0970340e04cc2357d84a82" Nov 28 13:34:51 crc kubenswrapper[4631]: E1128 13:34:51.299842 4631 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"61f33b53d073e336d54579efe11721c6af8ba587cb0970340e04cc2357d84a82\": container with ID starting with 61f33b53d073e336d54579efe11721c6af8ba587cb0970340e04cc2357d84a82 not found: ID does not exist" containerID="61f33b53d073e336d54579efe11721c6af8ba587cb0970340e04cc2357d84a82" Nov 28 13:34:51 crc kubenswrapper[4631]: I1128 13:34:51.299962 4631 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"61f33b53d073e336d54579efe11721c6af8ba587cb0970340e04cc2357d84a82"} err="failed to get container status \"61f33b53d073e336d54579efe11721c6af8ba587cb0970340e04cc2357d84a82\": rpc error: code = NotFound desc = could not find container \"61f33b53d073e336d54579efe11721c6af8ba587cb0970340e04cc2357d84a82\": container with ID starting with 61f33b53d073e336d54579efe11721c6af8ba587cb0970340e04cc2357d84a82 not found: ID does not exist" Nov 28 13:34:51 crc kubenswrapper[4631]: I1128 13:34:51.300000 4631 scope.go:117] "RemoveContainer" containerID="128147b0b653091336f64aea110f6a9a5219757366275397a467ddcdb05f1ccf" Nov 28 13:34:51 crc kubenswrapper[4631]: E1128 13:34:51.300402 4631 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"128147b0b653091336f64aea110f6a9a5219757366275397a467ddcdb05f1ccf\": container with ID starting with 128147b0b653091336f64aea110f6a9a5219757366275397a467ddcdb05f1ccf not found: ID does not exist" containerID="128147b0b653091336f64aea110f6a9a5219757366275397a467ddcdb05f1ccf" Nov 28 13:34:51 crc kubenswrapper[4631]: I1128 13:34:51.300430 4631 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"128147b0b653091336f64aea110f6a9a5219757366275397a467ddcdb05f1ccf"} err="failed to get container status \"128147b0b653091336f64aea110f6a9a5219757366275397a467ddcdb05f1ccf\": rpc error: code = NotFound desc = could not find container \"128147b0b653091336f64aea110f6a9a5219757366275397a467ddcdb05f1ccf\": container with ID starting with 128147b0b653091336f64aea110f6a9a5219757366275397a467ddcdb05f1ccf not found: ID does not exist" Nov 28 13:34:51 crc kubenswrapper[4631]: I1128 13:34:51.523201 4631 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf6aea40-3632-4c7c-90ea-5584dd081ef2" path="/var/lib/kubelet/pods/bf6aea40-3632-4c7c-90ea-5584dd081ef2/volumes" Nov 28 13:34:54 crc kubenswrapper[4631]: I1128 13:34:54.360072 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-5kbsx"] Nov 28 13:34:54 crc kubenswrapper[4631]: E1128 13:34:54.361556 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf6aea40-3632-4c7c-90ea-5584dd081ef2" containerName="extract-content" Nov 28 13:34:54 crc kubenswrapper[4631]: I1128 13:34:54.361643 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf6aea40-3632-4c7c-90ea-5584dd081ef2" containerName="extract-content" Nov 28 13:34:54 crc kubenswrapper[4631]: E1128 13:34:54.361724 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf6aea40-3632-4c7c-90ea-5584dd081ef2" containerName="registry-server" Nov 28 13:34:54 crc kubenswrapper[4631]: I1128 13:34:54.361775 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf6aea40-3632-4c7c-90ea-5584dd081ef2" containerName="registry-server" Nov 28 13:34:54 crc kubenswrapper[4631]: E1128 13:34:54.361831 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fae1e74c-0fac-4e9c-bf7c-e91157b6ef2d" containerName="extract-content" Nov 28 13:34:54 crc kubenswrapper[4631]: I1128 13:34:54.361888 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="fae1e74c-0fac-4e9c-bf7c-e91157b6ef2d" containerName="extract-content" Nov 28 13:34:54 crc kubenswrapper[4631]: E1128 13:34:54.361955 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fae1e74c-0fac-4e9c-bf7c-e91157b6ef2d" containerName="extract-utilities" Nov 28 13:34:54 crc kubenswrapper[4631]: I1128 13:34:54.362013 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="fae1e74c-0fac-4e9c-bf7c-e91157b6ef2d" containerName="extract-utilities" Nov 28 13:34:54 crc kubenswrapper[4631]: E1128 13:34:54.362073 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fae1e74c-0fac-4e9c-bf7c-e91157b6ef2d" containerName="registry-server" Nov 28 13:34:54 crc kubenswrapper[4631]: I1128 13:34:54.362122 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="fae1e74c-0fac-4e9c-bf7c-e91157b6ef2d" containerName="registry-server" Nov 28 13:34:54 crc kubenswrapper[4631]: E1128 13:34:54.362181 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8206d20e-c1fb-46c2-b9ab-b13a38922d29" containerName="extract-content" Nov 28 13:34:54 crc kubenswrapper[4631]: I1128 13:34:54.362231 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="8206d20e-c1fb-46c2-b9ab-b13a38922d29" containerName="extract-content" Nov 28 13:34:54 crc kubenswrapper[4631]: E1128 13:34:54.362318 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8206d20e-c1fb-46c2-b9ab-b13a38922d29" containerName="extract-utilities" Nov 28 13:34:54 crc kubenswrapper[4631]: I1128 13:34:54.362387 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="8206d20e-c1fb-46c2-b9ab-b13a38922d29" containerName="extract-utilities" Nov 28 13:34:54 crc kubenswrapper[4631]: E1128 13:34:54.362455 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8206d20e-c1fb-46c2-b9ab-b13a38922d29" containerName="registry-server" Nov 28 13:34:54 crc kubenswrapper[4631]: I1128 13:34:54.362506 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="8206d20e-c1fb-46c2-b9ab-b13a38922d29" containerName="registry-server" Nov 28 13:34:54 crc kubenswrapper[4631]: E1128 13:34:54.362569 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf6aea40-3632-4c7c-90ea-5584dd081ef2" containerName="extract-utilities" Nov 28 13:34:54 crc kubenswrapper[4631]: I1128 13:34:54.362633 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf6aea40-3632-4c7c-90ea-5584dd081ef2" containerName="extract-utilities" Nov 28 13:34:54 crc kubenswrapper[4631]: I1128 13:34:54.362882 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="8206d20e-c1fb-46c2-b9ab-b13a38922d29" containerName="registry-server" Nov 28 13:34:54 crc kubenswrapper[4631]: I1128 13:34:54.362970 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="bf6aea40-3632-4c7c-90ea-5584dd081ef2" containerName="registry-server" Nov 28 13:34:54 crc kubenswrapper[4631]: I1128 13:34:54.363057 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="fae1e74c-0fac-4e9c-bf7c-e91157b6ef2d" containerName="registry-server" Nov 28 13:34:54 crc kubenswrapper[4631]: I1128 13:34:54.364507 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5kbsx" Nov 28 13:34:54 crc kubenswrapper[4631]: I1128 13:34:54.371902 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-5kbsx"] Nov 28 13:34:54 crc kubenswrapper[4631]: I1128 13:34:54.430397 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/03ba4ed9-11ea-4c20-baaf-8975710c6708-catalog-content\") pod \"community-operators-5kbsx\" (UID: \"03ba4ed9-11ea-4c20-baaf-8975710c6708\") " pod="openshift-marketplace/community-operators-5kbsx" Nov 28 13:34:54 crc kubenswrapper[4631]: I1128 13:34:54.430584 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/03ba4ed9-11ea-4c20-baaf-8975710c6708-utilities\") pod \"community-operators-5kbsx\" (UID: \"03ba4ed9-11ea-4c20-baaf-8975710c6708\") " pod="openshift-marketplace/community-operators-5kbsx" Nov 28 13:34:54 crc kubenswrapper[4631]: I1128 13:34:54.430639 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5lgkc\" (UniqueName: \"kubernetes.io/projected/03ba4ed9-11ea-4c20-baaf-8975710c6708-kube-api-access-5lgkc\") pod \"community-operators-5kbsx\" (UID: \"03ba4ed9-11ea-4c20-baaf-8975710c6708\") " pod="openshift-marketplace/community-operators-5kbsx" Nov 28 13:34:54 crc kubenswrapper[4631]: I1128 13:34:54.532497 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/03ba4ed9-11ea-4c20-baaf-8975710c6708-catalog-content\") pod \"community-operators-5kbsx\" (UID: \"03ba4ed9-11ea-4c20-baaf-8975710c6708\") " pod="openshift-marketplace/community-operators-5kbsx" Nov 28 13:34:54 crc kubenswrapper[4631]: I1128 13:34:54.532597 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/03ba4ed9-11ea-4c20-baaf-8975710c6708-utilities\") pod \"community-operators-5kbsx\" (UID: \"03ba4ed9-11ea-4c20-baaf-8975710c6708\") " pod="openshift-marketplace/community-operators-5kbsx" Nov 28 13:34:54 crc kubenswrapper[4631]: I1128 13:34:54.532634 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5lgkc\" (UniqueName: \"kubernetes.io/projected/03ba4ed9-11ea-4c20-baaf-8975710c6708-kube-api-access-5lgkc\") pod \"community-operators-5kbsx\" (UID: \"03ba4ed9-11ea-4c20-baaf-8975710c6708\") " pod="openshift-marketplace/community-operators-5kbsx" Nov 28 13:34:54 crc kubenswrapper[4631]: I1128 13:34:54.533589 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/03ba4ed9-11ea-4c20-baaf-8975710c6708-catalog-content\") pod \"community-operators-5kbsx\" (UID: \"03ba4ed9-11ea-4c20-baaf-8975710c6708\") " pod="openshift-marketplace/community-operators-5kbsx" Nov 28 13:34:54 crc kubenswrapper[4631]: I1128 13:34:54.533824 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/03ba4ed9-11ea-4c20-baaf-8975710c6708-utilities\") pod \"community-operators-5kbsx\" (UID: \"03ba4ed9-11ea-4c20-baaf-8975710c6708\") " pod="openshift-marketplace/community-operators-5kbsx" Nov 28 13:34:54 crc kubenswrapper[4631]: I1128 13:34:54.557451 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5lgkc\" (UniqueName: \"kubernetes.io/projected/03ba4ed9-11ea-4c20-baaf-8975710c6708-kube-api-access-5lgkc\") pod \"community-operators-5kbsx\" (UID: \"03ba4ed9-11ea-4c20-baaf-8975710c6708\") " pod="openshift-marketplace/community-operators-5kbsx" Nov 28 13:34:54 crc kubenswrapper[4631]: I1128 13:34:54.684764 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5kbsx" Nov 28 13:34:55 crc kubenswrapper[4631]: I1128 13:34:55.254407 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-5kbsx"] Nov 28 13:34:56 crc kubenswrapper[4631]: I1128 13:34:56.246425 4631 generic.go:334] "Generic (PLEG): container finished" podID="03ba4ed9-11ea-4c20-baaf-8975710c6708" containerID="17ab23e3b640d1fff2f89f44b9ccfaeaa88a21edb104d9e450e9cba4860a98c1" exitCode=0 Nov 28 13:34:56 crc kubenswrapper[4631]: I1128 13:34:56.246544 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5kbsx" event={"ID":"03ba4ed9-11ea-4c20-baaf-8975710c6708","Type":"ContainerDied","Data":"17ab23e3b640d1fff2f89f44b9ccfaeaa88a21edb104d9e450e9cba4860a98c1"} Nov 28 13:34:56 crc kubenswrapper[4631]: I1128 13:34:56.247187 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5kbsx" event={"ID":"03ba4ed9-11ea-4c20-baaf-8975710c6708","Type":"ContainerStarted","Data":"1acd129545f28ef171bd341f19aa9260adee0a846dd7b798e4b6c5b9c8988ac1"} Nov 28 13:34:57 crc kubenswrapper[4631]: I1128 13:34:57.400329 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-mnfbw"] Nov 28 13:34:57 crc kubenswrapper[4631]: I1128 13:34:57.402387 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-mnfbw" Nov 28 13:34:57 crc kubenswrapper[4631]: I1128 13:34:57.405858 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Nov 28 13:34:57 crc kubenswrapper[4631]: I1128 13:34:57.406173 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Nov 28 13:34:57 crc kubenswrapper[4631]: I1128 13:34:57.406376 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Nov 28 13:34:57 crc kubenswrapper[4631]: I1128 13:34:57.406509 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-h6xfp" Nov 28 13:34:57 crc kubenswrapper[4631]: I1128 13:34:57.483593 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h49sd\" (UniqueName: \"kubernetes.io/projected/51fb0dc5-be7d-4ffb-a0e1-3050f570b1c3-kube-api-access-h49sd\") pod \"dnsmasq-dns-675f4bcbfc-mnfbw\" (UID: \"51fb0dc5-be7d-4ffb-a0e1-3050f570b1c3\") " pod="openstack/dnsmasq-dns-675f4bcbfc-mnfbw" Nov 28 13:34:57 crc kubenswrapper[4631]: I1128 13:34:57.483680 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/51fb0dc5-be7d-4ffb-a0e1-3050f570b1c3-config\") pod \"dnsmasq-dns-675f4bcbfc-mnfbw\" (UID: \"51fb0dc5-be7d-4ffb-a0e1-3050f570b1c3\") " pod="openstack/dnsmasq-dns-675f4bcbfc-mnfbw" Nov 28 13:34:57 crc kubenswrapper[4631]: I1128 13:34:57.494133 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-zmbf9"] Nov 28 13:34:57 crc kubenswrapper[4631]: I1128 13:34:57.495865 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-zmbf9" Nov 28 13:34:57 crc kubenswrapper[4631]: I1128 13:34:57.499158 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Nov 28 13:34:57 crc kubenswrapper[4631]: I1128 13:34:57.526331 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-mnfbw"] Nov 28 13:34:57 crc kubenswrapper[4631]: I1128 13:34:57.539469 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-zmbf9"] Nov 28 13:34:57 crc kubenswrapper[4631]: I1128 13:34:57.585470 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/51fb0dc5-be7d-4ffb-a0e1-3050f570b1c3-config\") pod \"dnsmasq-dns-675f4bcbfc-mnfbw\" (UID: \"51fb0dc5-be7d-4ffb-a0e1-3050f570b1c3\") " pod="openstack/dnsmasq-dns-675f4bcbfc-mnfbw" Nov 28 13:34:57 crc kubenswrapper[4631]: I1128 13:34:57.585601 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f884a85b-d7a2-4ca8-aa32-9c4e8578c064-config\") pod \"dnsmasq-dns-78dd6ddcc-zmbf9\" (UID: \"f884a85b-d7a2-4ca8-aa32-9c4e8578c064\") " pod="openstack/dnsmasq-dns-78dd6ddcc-zmbf9" Nov 28 13:34:57 crc kubenswrapper[4631]: I1128 13:34:57.585624 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ln2hw\" (UniqueName: \"kubernetes.io/projected/f884a85b-d7a2-4ca8-aa32-9c4e8578c064-kube-api-access-ln2hw\") pod \"dnsmasq-dns-78dd6ddcc-zmbf9\" (UID: \"f884a85b-d7a2-4ca8-aa32-9c4e8578c064\") " pod="openstack/dnsmasq-dns-78dd6ddcc-zmbf9" Nov 28 13:34:57 crc kubenswrapper[4631]: I1128 13:34:57.585655 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h49sd\" (UniqueName: \"kubernetes.io/projected/51fb0dc5-be7d-4ffb-a0e1-3050f570b1c3-kube-api-access-h49sd\") pod \"dnsmasq-dns-675f4bcbfc-mnfbw\" (UID: \"51fb0dc5-be7d-4ffb-a0e1-3050f570b1c3\") " pod="openstack/dnsmasq-dns-675f4bcbfc-mnfbw" Nov 28 13:34:57 crc kubenswrapper[4631]: I1128 13:34:57.585707 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f884a85b-d7a2-4ca8-aa32-9c4e8578c064-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-zmbf9\" (UID: \"f884a85b-d7a2-4ca8-aa32-9c4e8578c064\") " pod="openstack/dnsmasq-dns-78dd6ddcc-zmbf9" Nov 28 13:34:57 crc kubenswrapper[4631]: I1128 13:34:57.587346 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/51fb0dc5-be7d-4ffb-a0e1-3050f570b1c3-config\") pod \"dnsmasq-dns-675f4bcbfc-mnfbw\" (UID: \"51fb0dc5-be7d-4ffb-a0e1-3050f570b1c3\") " pod="openstack/dnsmasq-dns-675f4bcbfc-mnfbw" Nov 28 13:34:57 crc kubenswrapper[4631]: I1128 13:34:57.620518 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h49sd\" (UniqueName: \"kubernetes.io/projected/51fb0dc5-be7d-4ffb-a0e1-3050f570b1c3-kube-api-access-h49sd\") pod \"dnsmasq-dns-675f4bcbfc-mnfbw\" (UID: \"51fb0dc5-be7d-4ffb-a0e1-3050f570b1c3\") " pod="openstack/dnsmasq-dns-675f4bcbfc-mnfbw" Nov 28 13:34:57 crc kubenswrapper[4631]: I1128 13:34:57.687271 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f884a85b-d7a2-4ca8-aa32-9c4e8578c064-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-zmbf9\" (UID: \"f884a85b-d7a2-4ca8-aa32-9c4e8578c064\") " pod="openstack/dnsmasq-dns-78dd6ddcc-zmbf9" Nov 28 13:34:57 crc kubenswrapper[4631]: I1128 13:34:57.687400 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f884a85b-d7a2-4ca8-aa32-9c4e8578c064-config\") pod \"dnsmasq-dns-78dd6ddcc-zmbf9\" (UID: \"f884a85b-d7a2-4ca8-aa32-9c4e8578c064\") " pod="openstack/dnsmasq-dns-78dd6ddcc-zmbf9" Nov 28 13:34:57 crc kubenswrapper[4631]: I1128 13:34:57.687430 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ln2hw\" (UniqueName: \"kubernetes.io/projected/f884a85b-d7a2-4ca8-aa32-9c4e8578c064-kube-api-access-ln2hw\") pod \"dnsmasq-dns-78dd6ddcc-zmbf9\" (UID: \"f884a85b-d7a2-4ca8-aa32-9c4e8578c064\") " pod="openstack/dnsmasq-dns-78dd6ddcc-zmbf9" Nov 28 13:34:57 crc kubenswrapper[4631]: I1128 13:34:57.688436 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f884a85b-d7a2-4ca8-aa32-9c4e8578c064-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-zmbf9\" (UID: \"f884a85b-d7a2-4ca8-aa32-9c4e8578c064\") " pod="openstack/dnsmasq-dns-78dd6ddcc-zmbf9" Nov 28 13:34:57 crc kubenswrapper[4631]: I1128 13:34:57.688540 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f884a85b-d7a2-4ca8-aa32-9c4e8578c064-config\") pod \"dnsmasq-dns-78dd6ddcc-zmbf9\" (UID: \"f884a85b-d7a2-4ca8-aa32-9c4e8578c064\") " pod="openstack/dnsmasq-dns-78dd6ddcc-zmbf9" Nov 28 13:34:57 crc kubenswrapper[4631]: I1128 13:34:57.710363 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ln2hw\" (UniqueName: \"kubernetes.io/projected/f884a85b-d7a2-4ca8-aa32-9c4e8578c064-kube-api-access-ln2hw\") pod \"dnsmasq-dns-78dd6ddcc-zmbf9\" (UID: \"f884a85b-d7a2-4ca8-aa32-9c4e8578c064\") " pod="openstack/dnsmasq-dns-78dd6ddcc-zmbf9" Nov 28 13:34:57 crc kubenswrapper[4631]: I1128 13:34:57.719161 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-mnfbw" Nov 28 13:34:57 crc kubenswrapper[4631]: I1128 13:34:57.817734 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-zmbf9" Nov 28 13:34:58 crc kubenswrapper[4631]: I1128 13:34:58.297782 4631 generic.go:334] "Generic (PLEG): container finished" podID="03ba4ed9-11ea-4c20-baaf-8975710c6708" containerID="8a5833bbb45a201fe6bdee78e6f456aa459ba86952faeeceda5f1bdd2fce2a52" exitCode=0 Nov 28 13:34:58 crc kubenswrapper[4631]: I1128 13:34:58.298738 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5kbsx" event={"ID":"03ba4ed9-11ea-4c20-baaf-8975710c6708","Type":"ContainerDied","Data":"8a5833bbb45a201fe6bdee78e6f456aa459ba86952faeeceda5f1bdd2fce2a52"} Nov 28 13:34:58 crc kubenswrapper[4631]: I1128 13:34:58.650878 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-mnfbw"] Nov 28 13:34:58 crc kubenswrapper[4631]: W1128 13:34:58.657242 4631 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod51fb0dc5_be7d_4ffb_a0e1_3050f570b1c3.slice/crio-b1034ae3c0de521dfb8b9743d870163a09f3e2170f858fbb3354ecf377eb0b01 WatchSource:0}: Error finding container b1034ae3c0de521dfb8b9743d870163a09f3e2170f858fbb3354ecf377eb0b01: Status 404 returned error can't find the container with id b1034ae3c0de521dfb8b9743d870163a09f3e2170f858fbb3354ecf377eb0b01 Nov 28 13:34:58 crc kubenswrapper[4631]: I1128 13:34:58.759471 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-zmbf9"] Nov 28 13:34:59 crc kubenswrapper[4631]: I1128 13:34:59.314118 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-zmbf9" event={"ID":"f884a85b-d7a2-4ca8-aa32-9c4e8578c064","Type":"ContainerStarted","Data":"08468d82eb0f02639f9a54cc2d72b8dcc50ea479306820113b35cb321f72e5a1"} Nov 28 13:34:59 crc kubenswrapper[4631]: I1128 13:34:59.315186 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-mnfbw" event={"ID":"51fb0dc5-be7d-4ffb-a0e1-3050f570b1c3","Type":"ContainerStarted","Data":"b1034ae3c0de521dfb8b9743d870163a09f3e2170f858fbb3354ecf377eb0b01"} Nov 28 13:35:00 crc kubenswrapper[4631]: I1128 13:35:00.348456 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5kbsx" event={"ID":"03ba4ed9-11ea-4c20-baaf-8975710c6708","Type":"ContainerStarted","Data":"a83d391e2674fdc7da18b14130e6f1287d3f94f515d8a9241a80c096555d9ace"} Nov 28 13:35:00 crc kubenswrapper[4631]: I1128 13:35:00.396033 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-5kbsx" podStartSLOduration=3.38487462 podStartE2EDuration="6.396001491s" podCreationTimestamp="2025-11-28 13:34:54 +0000 UTC" firstStartedPulling="2025-11-28 13:34:56.249871112 +0000 UTC m=+853.057174456" lastFinishedPulling="2025-11-28 13:34:59.260997983 +0000 UTC m=+856.068301327" observedRunningTime="2025-11-28 13:35:00.390868022 +0000 UTC m=+857.198171366" watchObservedRunningTime="2025-11-28 13:35:00.396001491 +0000 UTC m=+857.203304855" Nov 28 13:35:00 crc kubenswrapper[4631]: I1128 13:35:00.437950 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-mnfbw"] Nov 28 13:35:00 crc kubenswrapper[4631]: I1128 13:35:00.501241 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-jl8jp"] Nov 28 13:35:00 crc kubenswrapper[4631]: I1128 13:35:00.509121 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-jl8jp" Nov 28 13:35:00 crc kubenswrapper[4631]: I1128 13:35:00.519237 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-jl8jp"] Nov 28 13:35:00 crc kubenswrapper[4631]: I1128 13:35:00.549736 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d605e389-3651-4b5c-a54d-a2013cbb460d-config\") pod \"dnsmasq-dns-666b6646f7-jl8jp\" (UID: \"d605e389-3651-4b5c-a54d-a2013cbb460d\") " pod="openstack/dnsmasq-dns-666b6646f7-jl8jp" Nov 28 13:35:00 crc kubenswrapper[4631]: I1128 13:35:00.549857 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d605e389-3651-4b5c-a54d-a2013cbb460d-dns-svc\") pod \"dnsmasq-dns-666b6646f7-jl8jp\" (UID: \"d605e389-3651-4b5c-a54d-a2013cbb460d\") " pod="openstack/dnsmasq-dns-666b6646f7-jl8jp" Nov 28 13:35:00 crc kubenswrapper[4631]: I1128 13:35:00.549905 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-696s2\" (UniqueName: \"kubernetes.io/projected/d605e389-3651-4b5c-a54d-a2013cbb460d-kube-api-access-696s2\") pod \"dnsmasq-dns-666b6646f7-jl8jp\" (UID: \"d605e389-3651-4b5c-a54d-a2013cbb460d\") " pod="openstack/dnsmasq-dns-666b6646f7-jl8jp" Nov 28 13:35:00 crc kubenswrapper[4631]: I1128 13:35:00.652043 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d605e389-3651-4b5c-a54d-a2013cbb460d-dns-svc\") pod \"dnsmasq-dns-666b6646f7-jl8jp\" (UID: \"d605e389-3651-4b5c-a54d-a2013cbb460d\") " pod="openstack/dnsmasq-dns-666b6646f7-jl8jp" Nov 28 13:35:00 crc kubenswrapper[4631]: I1128 13:35:00.652122 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-696s2\" (UniqueName: \"kubernetes.io/projected/d605e389-3651-4b5c-a54d-a2013cbb460d-kube-api-access-696s2\") pod \"dnsmasq-dns-666b6646f7-jl8jp\" (UID: \"d605e389-3651-4b5c-a54d-a2013cbb460d\") " pod="openstack/dnsmasq-dns-666b6646f7-jl8jp" Nov 28 13:35:00 crc kubenswrapper[4631]: I1128 13:35:00.652229 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d605e389-3651-4b5c-a54d-a2013cbb460d-config\") pod \"dnsmasq-dns-666b6646f7-jl8jp\" (UID: \"d605e389-3651-4b5c-a54d-a2013cbb460d\") " pod="openstack/dnsmasq-dns-666b6646f7-jl8jp" Nov 28 13:35:00 crc kubenswrapper[4631]: I1128 13:35:00.652979 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d605e389-3651-4b5c-a54d-a2013cbb460d-dns-svc\") pod \"dnsmasq-dns-666b6646f7-jl8jp\" (UID: \"d605e389-3651-4b5c-a54d-a2013cbb460d\") " pod="openstack/dnsmasq-dns-666b6646f7-jl8jp" Nov 28 13:35:00 crc kubenswrapper[4631]: I1128 13:35:00.653364 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d605e389-3651-4b5c-a54d-a2013cbb460d-config\") pod \"dnsmasq-dns-666b6646f7-jl8jp\" (UID: \"d605e389-3651-4b5c-a54d-a2013cbb460d\") " pod="openstack/dnsmasq-dns-666b6646f7-jl8jp" Nov 28 13:35:00 crc kubenswrapper[4631]: I1128 13:35:00.693472 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-696s2\" (UniqueName: \"kubernetes.io/projected/d605e389-3651-4b5c-a54d-a2013cbb460d-kube-api-access-696s2\") pod \"dnsmasq-dns-666b6646f7-jl8jp\" (UID: \"d605e389-3651-4b5c-a54d-a2013cbb460d\") " pod="openstack/dnsmasq-dns-666b6646f7-jl8jp" Nov 28 13:35:00 crc kubenswrapper[4631]: I1128 13:35:00.833868 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-jl8jp" Nov 28 13:35:01 crc kubenswrapper[4631]: I1128 13:35:01.158102 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-zmbf9"] Nov 28 13:35:01 crc kubenswrapper[4631]: I1128 13:35:01.263803 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-clh2s"] Nov 28 13:35:01 crc kubenswrapper[4631]: I1128 13:35:01.265463 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-clh2s" Nov 28 13:35:01 crc kubenswrapper[4631]: I1128 13:35:01.338310 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-clh2s"] Nov 28 13:35:01 crc kubenswrapper[4631]: I1128 13:35:01.364667 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-59vt5\" (UniqueName: \"kubernetes.io/projected/a9f46a16-a4a9-4ea0-af16-6c530dcd6386-kube-api-access-59vt5\") pod \"dnsmasq-dns-57d769cc4f-clh2s\" (UID: \"a9f46a16-a4a9-4ea0-af16-6c530dcd6386\") " pod="openstack/dnsmasq-dns-57d769cc4f-clh2s" Nov 28 13:35:01 crc kubenswrapper[4631]: I1128 13:35:01.364748 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a9f46a16-a4a9-4ea0-af16-6c530dcd6386-config\") pod \"dnsmasq-dns-57d769cc4f-clh2s\" (UID: \"a9f46a16-a4a9-4ea0-af16-6c530dcd6386\") " pod="openstack/dnsmasq-dns-57d769cc4f-clh2s" Nov 28 13:35:01 crc kubenswrapper[4631]: I1128 13:35:01.364774 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a9f46a16-a4a9-4ea0-af16-6c530dcd6386-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-clh2s\" (UID: \"a9f46a16-a4a9-4ea0-af16-6c530dcd6386\") " pod="openstack/dnsmasq-dns-57d769cc4f-clh2s" Nov 28 13:35:01 crc kubenswrapper[4631]: I1128 13:35:01.467312 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a9f46a16-a4a9-4ea0-af16-6c530dcd6386-config\") pod \"dnsmasq-dns-57d769cc4f-clh2s\" (UID: \"a9f46a16-a4a9-4ea0-af16-6c530dcd6386\") " pod="openstack/dnsmasq-dns-57d769cc4f-clh2s" Nov 28 13:35:01 crc kubenswrapper[4631]: I1128 13:35:01.467805 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-59vt5\" (UniqueName: \"kubernetes.io/projected/a9f46a16-a4a9-4ea0-af16-6c530dcd6386-kube-api-access-59vt5\") pod \"dnsmasq-dns-57d769cc4f-clh2s\" (UID: \"a9f46a16-a4a9-4ea0-af16-6c530dcd6386\") " pod="openstack/dnsmasq-dns-57d769cc4f-clh2s" Nov 28 13:35:01 crc kubenswrapper[4631]: I1128 13:35:01.467842 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a9f46a16-a4a9-4ea0-af16-6c530dcd6386-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-clh2s\" (UID: \"a9f46a16-a4a9-4ea0-af16-6c530dcd6386\") " pod="openstack/dnsmasq-dns-57d769cc4f-clh2s" Nov 28 13:35:01 crc kubenswrapper[4631]: I1128 13:35:01.468859 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a9f46a16-a4a9-4ea0-af16-6c530dcd6386-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-clh2s\" (UID: \"a9f46a16-a4a9-4ea0-af16-6c530dcd6386\") " pod="openstack/dnsmasq-dns-57d769cc4f-clh2s" Nov 28 13:35:01 crc kubenswrapper[4631]: I1128 13:35:01.469445 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a9f46a16-a4a9-4ea0-af16-6c530dcd6386-config\") pod \"dnsmasq-dns-57d769cc4f-clh2s\" (UID: \"a9f46a16-a4a9-4ea0-af16-6c530dcd6386\") " pod="openstack/dnsmasq-dns-57d769cc4f-clh2s" Nov 28 13:35:01 crc kubenswrapper[4631]: I1128 13:35:01.532330 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-59vt5\" (UniqueName: \"kubernetes.io/projected/a9f46a16-a4a9-4ea0-af16-6c530dcd6386-kube-api-access-59vt5\") pod \"dnsmasq-dns-57d769cc4f-clh2s\" (UID: \"a9f46a16-a4a9-4ea0-af16-6c530dcd6386\") " pod="openstack/dnsmasq-dns-57d769cc4f-clh2s" Nov 28 13:35:01 crc kubenswrapper[4631]: I1128 13:35:01.590952 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-clh2s" Nov 28 13:35:01 crc kubenswrapper[4631]: I1128 13:35:01.721084 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Nov 28 13:35:01 crc kubenswrapper[4631]: I1128 13:35:01.742024 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 28 13:35:01 crc kubenswrapper[4631]: I1128 13:35:01.765762 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Nov 28 13:35:01 crc kubenswrapper[4631]: I1128 13:35:01.765941 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Nov 28 13:35:01 crc kubenswrapper[4631]: I1128 13:35:01.765946 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Nov 28 13:35:01 crc kubenswrapper[4631]: I1128 13:35:01.766117 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-p7rc4" Nov 28 13:35:01 crc kubenswrapper[4631]: I1128 13:35:01.766355 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Nov 28 13:35:01 crc kubenswrapper[4631]: I1128 13:35:01.766536 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Nov 28 13:35:01 crc kubenswrapper[4631]: I1128 13:35:01.795971 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Nov 28 13:35:01 crc kubenswrapper[4631]: I1128 13:35:01.905379 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 28 13:35:01 crc kubenswrapper[4631]: I1128 13:35:01.906944 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/3a91e9fa-34b7-4688-9171-56a62043759d-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"3a91e9fa-34b7-4688-9171-56a62043759d\") " pod="openstack/rabbitmq-server-0" Nov 28 13:35:01 crc kubenswrapper[4631]: I1128 13:35:01.907003 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/3a91e9fa-34b7-4688-9171-56a62043759d-server-conf\") pod \"rabbitmq-server-0\" (UID: \"3a91e9fa-34b7-4688-9171-56a62043759d\") " pod="openstack/rabbitmq-server-0" Nov 28 13:35:01 crc kubenswrapper[4631]: I1128 13:35:01.907040 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/3a91e9fa-34b7-4688-9171-56a62043759d-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"3a91e9fa-34b7-4688-9171-56a62043759d\") " pod="openstack/rabbitmq-server-0" Nov 28 13:35:01 crc kubenswrapper[4631]: I1128 13:35:01.907090 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/3a91e9fa-34b7-4688-9171-56a62043759d-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"3a91e9fa-34b7-4688-9171-56a62043759d\") " pod="openstack/rabbitmq-server-0" Nov 28 13:35:01 crc kubenswrapper[4631]: I1128 13:35:01.907119 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/3a91e9fa-34b7-4688-9171-56a62043759d-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"3a91e9fa-34b7-4688-9171-56a62043759d\") " pod="openstack/rabbitmq-server-0" Nov 28 13:35:01 crc kubenswrapper[4631]: I1128 13:35:01.907142 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4x8mv\" (UniqueName: \"kubernetes.io/projected/3a91e9fa-34b7-4688-9171-56a62043759d-kube-api-access-4x8mv\") pod \"rabbitmq-server-0\" (UID: \"3a91e9fa-34b7-4688-9171-56a62043759d\") " pod="openstack/rabbitmq-server-0" Nov 28 13:35:01 crc kubenswrapper[4631]: I1128 13:35:01.907163 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/3a91e9fa-34b7-4688-9171-56a62043759d-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"3a91e9fa-34b7-4688-9171-56a62043759d\") " pod="openstack/rabbitmq-server-0" Nov 28 13:35:01 crc kubenswrapper[4631]: I1128 13:35:01.907186 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-server-0\" (UID: \"3a91e9fa-34b7-4688-9171-56a62043759d\") " pod="openstack/rabbitmq-server-0" Nov 28 13:35:01 crc kubenswrapper[4631]: I1128 13:35:01.907210 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/3a91e9fa-34b7-4688-9171-56a62043759d-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"3a91e9fa-34b7-4688-9171-56a62043759d\") " pod="openstack/rabbitmq-server-0" Nov 28 13:35:01 crc kubenswrapper[4631]: I1128 13:35:01.907241 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3a91e9fa-34b7-4688-9171-56a62043759d-config-data\") pod \"rabbitmq-server-0\" (UID: \"3a91e9fa-34b7-4688-9171-56a62043759d\") " pod="openstack/rabbitmq-server-0" Nov 28 13:35:01 crc kubenswrapper[4631]: I1128 13:35:01.907275 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/3a91e9fa-34b7-4688-9171-56a62043759d-pod-info\") pod \"rabbitmq-server-0\" (UID: \"3a91e9fa-34b7-4688-9171-56a62043759d\") " pod="openstack/rabbitmq-server-0" Nov 28 13:35:02 crc kubenswrapper[4631]: I1128 13:35:02.009745 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/3a91e9fa-34b7-4688-9171-56a62043759d-server-conf\") pod \"rabbitmq-server-0\" (UID: \"3a91e9fa-34b7-4688-9171-56a62043759d\") " pod="openstack/rabbitmq-server-0" Nov 28 13:35:02 crc kubenswrapper[4631]: I1128 13:35:02.009805 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/3a91e9fa-34b7-4688-9171-56a62043759d-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"3a91e9fa-34b7-4688-9171-56a62043759d\") " pod="openstack/rabbitmq-server-0" Nov 28 13:35:02 crc kubenswrapper[4631]: I1128 13:35:02.009856 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/3a91e9fa-34b7-4688-9171-56a62043759d-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"3a91e9fa-34b7-4688-9171-56a62043759d\") " pod="openstack/rabbitmq-server-0" Nov 28 13:35:02 crc kubenswrapper[4631]: I1128 13:35:02.009884 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/3a91e9fa-34b7-4688-9171-56a62043759d-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"3a91e9fa-34b7-4688-9171-56a62043759d\") " pod="openstack/rabbitmq-server-0" Nov 28 13:35:02 crc kubenswrapper[4631]: I1128 13:35:02.009905 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/3a91e9fa-34b7-4688-9171-56a62043759d-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"3a91e9fa-34b7-4688-9171-56a62043759d\") " pod="openstack/rabbitmq-server-0" Nov 28 13:35:02 crc kubenswrapper[4631]: I1128 13:35:02.009921 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4x8mv\" (UniqueName: \"kubernetes.io/projected/3a91e9fa-34b7-4688-9171-56a62043759d-kube-api-access-4x8mv\") pod \"rabbitmq-server-0\" (UID: \"3a91e9fa-34b7-4688-9171-56a62043759d\") " pod="openstack/rabbitmq-server-0" Nov 28 13:35:02 crc kubenswrapper[4631]: I1128 13:35:02.009942 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-server-0\" (UID: \"3a91e9fa-34b7-4688-9171-56a62043759d\") " pod="openstack/rabbitmq-server-0" Nov 28 13:35:02 crc kubenswrapper[4631]: I1128 13:35:02.009963 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/3a91e9fa-34b7-4688-9171-56a62043759d-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"3a91e9fa-34b7-4688-9171-56a62043759d\") " pod="openstack/rabbitmq-server-0" Nov 28 13:35:02 crc kubenswrapper[4631]: I1128 13:35:02.009990 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3a91e9fa-34b7-4688-9171-56a62043759d-config-data\") pod \"rabbitmq-server-0\" (UID: \"3a91e9fa-34b7-4688-9171-56a62043759d\") " pod="openstack/rabbitmq-server-0" Nov 28 13:35:02 crc kubenswrapper[4631]: I1128 13:35:02.010023 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/3a91e9fa-34b7-4688-9171-56a62043759d-pod-info\") pod \"rabbitmq-server-0\" (UID: \"3a91e9fa-34b7-4688-9171-56a62043759d\") " pod="openstack/rabbitmq-server-0" Nov 28 13:35:02 crc kubenswrapper[4631]: I1128 13:35:02.010047 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/3a91e9fa-34b7-4688-9171-56a62043759d-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"3a91e9fa-34b7-4688-9171-56a62043759d\") " pod="openstack/rabbitmq-server-0" Nov 28 13:35:02 crc kubenswrapper[4631]: I1128 13:35:02.011647 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/3a91e9fa-34b7-4688-9171-56a62043759d-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"3a91e9fa-34b7-4688-9171-56a62043759d\") " pod="openstack/rabbitmq-server-0" Nov 28 13:35:02 crc kubenswrapper[4631]: I1128 13:35:02.011813 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/3a91e9fa-34b7-4688-9171-56a62043759d-server-conf\") pod \"rabbitmq-server-0\" (UID: \"3a91e9fa-34b7-4688-9171-56a62043759d\") " pod="openstack/rabbitmq-server-0" Nov 28 13:35:02 crc kubenswrapper[4631]: I1128 13:35:02.012741 4631 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-server-0\" (UID: \"3a91e9fa-34b7-4688-9171-56a62043759d\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/rabbitmq-server-0" Nov 28 13:35:02 crc kubenswrapper[4631]: I1128 13:35:02.012910 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3a91e9fa-34b7-4688-9171-56a62043759d-config-data\") pod \"rabbitmq-server-0\" (UID: \"3a91e9fa-34b7-4688-9171-56a62043759d\") " pod="openstack/rabbitmq-server-0" Nov 28 13:35:02 crc kubenswrapper[4631]: I1128 13:35:02.027593 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/3a91e9fa-34b7-4688-9171-56a62043759d-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"3a91e9fa-34b7-4688-9171-56a62043759d\") " pod="openstack/rabbitmq-server-0" Nov 28 13:35:02 crc kubenswrapper[4631]: I1128 13:35:02.030030 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/3a91e9fa-34b7-4688-9171-56a62043759d-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"3a91e9fa-34b7-4688-9171-56a62043759d\") " pod="openstack/rabbitmq-server-0" Nov 28 13:35:02 crc kubenswrapper[4631]: I1128 13:35:02.033718 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/3a91e9fa-34b7-4688-9171-56a62043759d-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"3a91e9fa-34b7-4688-9171-56a62043759d\") " pod="openstack/rabbitmq-server-0" Nov 28 13:35:02 crc kubenswrapper[4631]: I1128 13:35:02.033977 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/3a91e9fa-34b7-4688-9171-56a62043759d-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"3a91e9fa-34b7-4688-9171-56a62043759d\") " pod="openstack/rabbitmq-server-0" Nov 28 13:35:02 crc kubenswrapper[4631]: I1128 13:35:02.038933 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/3a91e9fa-34b7-4688-9171-56a62043759d-pod-info\") pod \"rabbitmq-server-0\" (UID: \"3a91e9fa-34b7-4688-9171-56a62043759d\") " pod="openstack/rabbitmq-server-0" Nov 28 13:35:02 crc kubenswrapper[4631]: I1128 13:35:02.044806 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/3a91e9fa-34b7-4688-9171-56a62043759d-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"3a91e9fa-34b7-4688-9171-56a62043759d\") " pod="openstack/rabbitmq-server-0" Nov 28 13:35:02 crc kubenswrapper[4631]: I1128 13:35:02.045162 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4x8mv\" (UniqueName: \"kubernetes.io/projected/3a91e9fa-34b7-4688-9171-56a62043759d-kube-api-access-4x8mv\") pod \"rabbitmq-server-0\" (UID: \"3a91e9fa-34b7-4688-9171-56a62043759d\") " pod="openstack/rabbitmq-server-0" Nov 28 13:35:02 crc kubenswrapper[4631]: I1128 13:35:02.141336 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-server-0\" (UID: \"3a91e9fa-34b7-4688-9171-56a62043759d\") " pod="openstack/rabbitmq-server-0" Nov 28 13:35:02 crc kubenswrapper[4631]: I1128 13:35:02.205927 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-jl8jp"] Nov 28 13:35:02 crc kubenswrapper[4631]: I1128 13:35:02.395602 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 28 13:35:02 crc kubenswrapper[4631]: I1128 13:35:02.422463 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-jl8jp" event={"ID":"d605e389-3651-4b5c-a54d-a2013cbb460d","Type":"ContainerStarted","Data":"5d750248066cb1fae2c16050241dc5845c87a70e14401d14bd23d5018a05c1c2"} Nov 28 13:35:02 crc kubenswrapper[4631]: I1128 13:35:02.482387 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 28 13:35:02 crc kubenswrapper[4631]: I1128 13:35:02.485906 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 28 13:35:02 crc kubenswrapper[4631]: I1128 13:35:02.495208 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Nov 28 13:35:02 crc kubenswrapper[4631]: I1128 13:35:02.499901 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Nov 28 13:35:02 crc kubenswrapper[4631]: I1128 13:35:02.499942 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Nov 28 13:35:02 crc kubenswrapper[4631]: I1128 13:35:02.500231 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Nov 28 13:35:02 crc kubenswrapper[4631]: I1128 13:35:02.500336 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Nov 28 13:35:02 crc kubenswrapper[4631]: I1128 13:35:02.500428 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Nov 28 13:35:02 crc kubenswrapper[4631]: I1128 13:35:02.504183 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-vf4xz" Nov 28 13:35:02 crc kubenswrapper[4631]: I1128 13:35:02.507829 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 28 13:35:02 crc kubenswrapper[4631]: I1128 13:35:02.634932 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ed4141f2-5eff-4252-8dc7-ad60a1580189-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"ed4141f2-5eff-4252-8dc7-ad60a1580189\") " pod="openstack/rabbitmq-cell1-server-0" Nov 28 13:35:02 crc kubenswrapper[4631]: I1128 13:35:02.635018 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/ed4141f2-5eff-4252-8dc7-ad60a1580189-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"ed4141f2-5eff-4252-8dc7-ad60a1580189\") " pod="openstack/rabbitmq-cell1-server-0" Nov 28 13:35:02 crc kubenswrapper[4631]: I1128 13:35:02.635073 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/ed4141f2-5eff-4252-8dc7-ad60a1580189-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"ed4141f2-5eff-4252-8dc7-ad60a1580189\") " pod="openstack/rabbitmq-cell1-server-0" Nov 28 13:35:02 crc kubenswrapper[4631]: I1128 13:35:02.635171 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/ed4141f2-5eff-4252-8dc7-ad60a1580189-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"ed4141f2-5eff-4252-8dc7-ad60a1580189\") " pod="openstack/rabbitmq-cell1-server-0" Nov 28 13:35:02 crc kubenswrapper[4631]: I1128 13:35:02.635189 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/ed4141f2-5eff-4252-8dc7-ad60a1580189-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"ed4141f2-5eff-4252-8dc7-ad60a1580189\") " pod="openstack/rabbitmq-cell1-server-0" Nov 28 13:35:02 crc kubenswrapper[4631]: I1128 13:35:02.635312 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"ed4141f2-5eff-4252-8dc7-ad60a1580189\") " pod="openstack/rabbitmq-cell1-server-0" Nov 28 13:35:02 crc kubenswrapper[4631]: I1128 13:35:02.635333 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/ed4141f2-5eff-4252-8dc7-ad60a1580189-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"ed4141f2-5eff-4252-8dc7-ad60a1580189\") " pod="openstack/rabbitmq-cell1-server-0" Nov 28 13:35:02 crc kubenswrapper[4631]: I1128 13:35:02.635377 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4vt5k\" (UniqueName: \"kubernetes.io/projected/ed4141f2-5eff-4252-8dc7-ad60a1580189-kube-api-access-4vt5k\") pod \"rabbitmq-cell1-server-0\" (UID: \"ed4141f2-5eff-4252-8dc7-ad60a1580189\") " pod="openstack/rabbitmq-cell1-server-0" Nov 28 13:35:02 crc kubenswrapper[4631]: I1128 13:35:02.635416 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/ed4141f2-5eff-4252-8dc7-ad60a1580189-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"ed4141f2-5eff-4252-8dc7-ad60a1580189\") " pod="openstack/rabbitmq-cell1-server-0" Nov 28 13:35:02 crc kubenswrapper[4631]: I1128 13:35:02.635465 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/ed4141f2-5eff-4252-8dc7-ad60a1580189-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"ed4141f2-5eff-4252-8dc7-ad60a1580189\") " pod="openstack/rabbitmq-cell1-server-0" Nov 28 13:35:02 crc kubenswrapper[4631]: I1128 13:35:02.635484 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/ed4141f2-5eff-4252-8dc7-ad60a1580189-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"ed4141f2-5eff-4252-8dc7-ad60a1580189\") " pod="openstack/rabbitmq-cell1-server-0" Nov 28 13:35:02 crc kubenswrapper[4631]: I1128 13:35:02.711321 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-clh2s"] Nov 28 13:35:02 crc kubenswrapper[4631]: I1128 13:35:02.737108 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/ed4141f2-5eff-4252-8dc7-ad60a1580189-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"ed4141f2-5eff-4252-8dc7-ad60a1580189\") " pod="openstack/rabbitmq-cell1-server-0" Nov 28 13:35:02 crc kubenswrapper[4631]: I1128 13:35:02.737141 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/ed4141f2-5eff-4252-8dc7-ad60a1580189-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"ed4141f2-5eff-4252-8dc7-ad60a1580189\") " pod="openstack/rabbitmq-cell1-server-0" Nov 28 13:35:02 crc kubenswrapper[4631]: I1128 13:35:02.737165 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"ed4141f2-5eff-4252-8dc7-ad60a1580189\") " pod="openstack/rabbitmq-cell1-server-0" Nov 28 13:35:02 crc kubenswrapper[4631]: I1128 13:35:02.737183 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/ed4141f2-5eff-4252-8dc7-ad60a1580189-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"ed4141f2-5eff-4252-8dc7-ad60a1580189\") " pod="openstack/rabbitmq-cell1-server-0" Nov 28 13:35:02 crc kubenswrapper[4631]: I1128 13:35:02.737203 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4vt5k\" (UniqueName: \"kubernetes.io/projected/ed4141f2-5eff-4252-8dc7-ad60a1580189-kube-api-access-4vt5k\") pod \"rabbitmq-cell1-server-0\" (UID: \"ed4141f2-5eff-4252-8dc7-ad60a1580189\") " pod="openstack/rabbitmq-cell1-server-0" Nov 28 13:35:02 crc kubenswrapper[4631]: I1128 13:35:02.737230 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/ed4141f2-5eff-4252-8dc7-ad60a1580189-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"ed4141f2-5eff-4252-8dc7-ad60a1580189\") " pod="openstack/rabbitmq-cell1-server-0" Nov 28 13:35:02 crc kubenswrapper[4631]: I1128 13:35:02.737268 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/ed4141f2-5eff-4252-8dc7-ad60a1580189-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"ed4141f2-5eff-4252-8dc7-ad60a1580189\") " pod="openstack/rabbitmq-cell1-server-0" Nov 28 13:35:02 crc kubenswrapper[4631]: I1128 13:35:02.737300 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/ed4141f2-5eff-4252-8dc7-ad60a1580189-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"ed4141f2-5eff-4252-8dc7-ad60a1580189\") " pod="openstack/rabbitmq-cell1-server-0" Nov 28 13:35:02 crc kubenswrapper[4631]: I1128 13:35:02.737327 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ed4141f2-5eff-4252-8dc7-ad60a1580189-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"ed4141f2-5eff-4252-8dc7-ad60a1580189\") " pod="openstack/rabbitmq-cell1-server-0" Nov 28 13:35:02 crc kubenswrapper[4631]: I1128 13:35:02.738846 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/ed4141f2-5eff-4252-8dc7-ad60a1580189-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"ed4141f2-5eff-4252-8dc7-ad60a1580189\") " pod="openstack/rabbitmq-cell1-server-0" Nov 28 13:35:02 crc kubenswrapper[4631]: I1128 13:35:02.739255 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/ed4141f2-5eff-4252-8dc7-ad60a1580189-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"ed4141f2-5eff-4252-8dc7-ad60a1580189\") " pod="openstack/rabbitmq-cell1-server-0" Nov 28 13:35:02 crc kubenswrapper[4631]: I1128 13:35:02.740347 4631 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"ed4141f2-5eff-4252-8dc7-ad60a1580189\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/rabbitmq-cell1-server-0" Nov 28 13:35:02 crc kubenswrapper[4631]: I1128 13:35:02.740360 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/ed4141f2-5eff-4252-8dc7-ad60a1580189-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"ed4141f2-5eff-4252-8dc7-ad60a1580189\") " pod="openstack/rabbitmq-cell1-server-0" Nov 28 13:35:02 crc kubenswrapper[4631]: I1128 13:35:02.741177 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ed4141f2-5eff-4252-8dc7-ad60a1580189-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"ed4141f2-5eff-4252-8dc7-ad60a1580189\") " pod="openstack/rabbitmq-cell1-server-0" Nov 28 13:35:02 crc kubenswrapper[4631]: I1128 13:35:02.741548 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/ed4141f2-5eff-4252-8dc7-ad60a1580189-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"ed4141f2-5eff-4252-8dc7-ad60a1580189\") " pod="openstack/rabbitmq-cell1-server-0" Nov 28 13:35:02 crc kubenswrapper[4631]: I1128 13:35:02.741705 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/ed4141f2-5eff-4252-8dc7-ad60a1580189-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"ed4141f2-5eff-4252-8dc7-ad60a1580189\") " pod="openstack/rabbitmq-cell1-server-0" Nov 28 13:35:02 crc kubenswrapper[4631]: I1128 13:35:02.743836 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/ed4141f2-5eff-4252-8dc7-ad60a1580189-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"ed4141f2-5eff-4252-8dc7-ad60a1580189\") " pod="openstack/rabbitmq-cell1-server-0" Nov 28 13:35:02 crc kubenswrapper[4631]: I1128 13:35:02.765968 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/ed4141f2-5eff-4252-8dc7-ad60a1580189-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"ed4141f2-5eff-4252-8dc7-ad60a1580189\") " pod="openstack/rabbitmq-cell1-server-0" Nov 28 13:35:02 crc kubenswrapper[4631]: I1128 13:35:02.774935 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/ed4141f2-5eff-4252-8dc7-ad60a1580189-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"ed4141f2-5eff-4252-8dc7-ad60a1580189\") " pod="openstack/rabbitmq-cell1-server-0" Nov 28 13:35:02 crc kubenswrapper[4631]: I1128 13:35:02.786484 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/ed4141f2-5eff-4252-8dc7-ad60a1580189-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"ed4141f2-5eff-4252-8dc7-ad60a1580189\") " pod="openstack/rabbitmq-cell1-server-0" Nov 28 13:35:02 crc kubenswrapper[4631]: I1128 13:35:02.802261 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/ed4141f2-5eff-4252-8dc7-ad60a1580189-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"ed4141f2-5eff-4252-8dc7-ad60a1580189\") " pod="openstack/rabbitmq-cell1-server-0" Nov 28 13:35:02 crc kubenswrapper[4631]: I1128 13:35:02.815904 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4vt5k\" (UniqueName: \"kubernetes.io/projected/ed4141f2-5eff-4252-8dc7-ad60a1580189-kube-api-access-4vt5k\") pod \"rabbitmq-cell1-server-0\" (UID: \"ed4141f2-5eff-4252-8dc7-ad60a1580189\") " pod="openstack/rabbitmq-cell1-server-0" Nov 28 13:35:02 crc kubenswrapper[4631]: I1128 13:35:02.830498 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"ed4141f2-5eff-4252-8dc7-ad60a1580189\") " pod="openstack/rabbitmq-cell1-server-0" Nov 28 13:35:02 crc kubenswrapper[4631]: I1128 13:35:02.850570 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 28 13:35:03 crc kubenswrapper[4631]: I1128 13:35:03.293331 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 28 13:35:03 crc kubenswrapper[4631]: I1128 13:35:03.510566 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"3a91e9fa-34b7-4688-9171-56a62043759d","Type":"ContainerStarted","Data":"eca7fbb416ef9f4ae40f525f94aa899b89d54b720c33ba9321840fe268dd2633"} Nov 28 13:35:03 crc kubenswrapper[4631]: I1128 13:35:03.592144 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-clh2s" event={"ID":"a9f46a16-a4a9-4ea0-af16-6c530dcd6386","Type":"ContainerStarted","Data":"3be66d5894663175baf25766072009a3f44f58601ca4e7a07b87de050d976c94"} Nov 28 13:35:03 crc kubenswrapper[4631]: I1128 13:35:03.671913 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Nov 28 13:35:03 crc kubenswrapper[4631]: I1128 13:35:03.674758 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Nov 28 13:35:03 crc kubenswrapper[4631]: I1128 13:35:03.690224 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Nov 28 13:35:03 crc kubenswrapper[4631]: I1128 13:35:03.696267 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Nov 28 13:35:03 crc kubenswrapper[4631]: I1128 13:35:03.696713 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-47zs4" Nov 28 13:35:03 crc kubenswrapper[4631]: I1128 13:35:03.697569 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Nov 28 13:35:03 crc kubenswrapper[4631]: I1128 13:35:03.697942 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Nov 28 13:35:03 crc kubenswrapper[4631]: I1128 13:35:03.713793 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Nov 28 13:35:03 crc kubenswrapper[4631]: I1128 13:35:03.791009 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/1a711f62-44fc-4046-a3aa-b5ef77b45e62-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"1a711f62-44fc-4046-a3aa-b5ef77b45e62\") " pod="openstack/openstack-galera-0" Nov 28 13:35:03 crc kubenswrapper[4631]: I1128 13:35:03.791072 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1a711f62-44fc-4046-a3aa-b5ef77b45e62-operator-scripts\") pod \"openstack-galera-0\" (UID: \"1a711f62-44fc-4046-a3aa-b5ef77b45e62\") " pod="openstack/openstack-galera-0" Nov 28 13:35:03 crc kubenswrapper[4631]: I1128 13:35:03.791094 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a711f62-44fc-4046-a3aa-b5ef77b45e62-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"1a711f62-44fc-4046-a3aa-b5ef77b45e62\") " pod="openstack/openstack-galera-0" Nov 28 13:35:03 crc kubenswrapper[4631]: I1128 13:35:03.791145 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-galera-0\" (UID: \"1a711f62-44fc-4046-a3aa-b5ef77b45e62\") " pod="openstack/openstack-galera-0" Nov 28 13:35:03 crc kubenswrapper[4631]: I1128 13:35:03.791166 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/1a711f62-44fc-4046-a3aa-b5ef77b45e62-kolla-config\") pod \"openstack-galera-0\" (UID: \"1a711f62-44fc-4046-a3aa-b5ef77b45e62\") " pod="openstack/openstack-galera-0" Nov 28 13:35:03 crc kubenswrapper[4631]: I1128 13:35:03.791192 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/1a711f62-44fc-4046-a3aa-b5ef77b45e62-config-data-default\") pod \"openstack-galera-0\" (UID: \"1a711f62-44fc-4046-a3aa-b5ef77b45e62\") " pod="openstack/openstack-galera-0" Nov 28 13:35:03 crc kubenswrapper[4631]: I1128 13:35:03.791224 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kpprt\" (UniqueName: \"kubernetes.io/projected/1a711f62-44fc-4046-a3aa-b5ef77b45e62-kube-api-access-kpprt\") pod \"openstack-galera-0\" (UID: \"1a711f62-44fc-4046-a3aa-b5ef77b45e62\") " pod="openstack/openstack-galera-0" Nov 28 13:35:03 crc kubenswrapper[4631]: I1128 13:35:03.791263 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/1a711f62-44fc-4046-a3aa-b5ef77b45e62-config-data-generated\") pod \"openstack-galera-0\" (UID: \"1a711f62-44fc-4046-a3aa-b5ef77b45e62\") " pod="openstack/openstack-galera-0" Nov 28 13:35:03 crc kubenswrapper[4631]: I1128 13:35:03.908820 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/1a711f62-44fc-4046-a3aa-b5ef77b45e62-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"1a711f62-44fc-4046-a3aa-b5ef77b45e62\") " pod="openstack/openstack-galera-0" Nov 28 13:35:03 crc kubenswrapper[4631]: I1128 13:35:03.912397 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1a711f62-44fc-4046-a3aa-b5ef77b45e62-operator-scripts\") pod \"openstack-galera-0\" (UID: \"1a711f62-44fc-4046-a3aa-b5ef77b45e62\") " pod="openstack/openstack-galera-0" Nov 28 13:35:03 crc kubenswrapper[4631]: I1128 13:35:03.912565 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a711f62-44fc-4046-a3aa-b5ef77b45e62-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"1a711f62-44fc-4046-a3aa-b5ef77b45e62\") " pod="openstack/openstack-galera-0" Nov 28 13:35:03 crc kubenswrapper[4631]: I1128 13:35:03.912808 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-galera-0\" (UID: \"1a711f62-44fc-4046-a3aa-b5ef77b45e62\") " pod="openstack/openstack-galera-0" Nov 28 13:35:03 crc kubenswrapper[4631]: I1128 13:35:03.919389 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/1a711f62-44fc-4046-a3aa-b5ef77b45e62-kolla-config\") pod \"openstack-galera-0\" (UID: \"1a711f62-44fc-4046-a3aa-b5ef77b45e62\") " pod="openstack/openstack-galera-0" Nov 28 13:35:03 crc kubenswrapper[4631]: I1128 13:35:03.925445 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/1a711f62-44fc-4046-a3aa-b5ef77b45e62-config-data-default\") pod \"openstack-galera-0\" (UID: \"1a711f62-44fc-4046-a3aa-b5ef77b45e62\") " pod="openstack/openstack-galera-0" Nov 28 13:35:03 crc kubenswrapper[4631]: I1128 13:35:03.925752 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kpprt\" (UniqueName: \"kubernetes.io/projected/1a711f62-44fc-4046-a3aa-b5ef77b45e62-kube-api-access-kpprt\") pod \"openstack-galera-0\" (UID: \"1a711f62-44fc-4046-a3aa-b5ef77b45e62\") " pod="openstack/openstack-galera-0" Nov 28 13:35:03 crc kubenswrapper[4631]: I1128 13:35:03.925889 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/1a711f62-44fc-4046-a3aa-b5ef77b45e62-config-data-generated\") pod \"openstack-galera-0\" (UID: \"1a711f62-44fc-4046-a3aa-b5ef77b45e62\") " pod="openstack/openstack-galera-0" Nov 28 13:35:03 crc kubenswrapper[4631]: I1128 13:35:03.922379 4631 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-galera-0\" (UID: \"1a711f62-44fc-4046-a3aa-b5ef77b45e62\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/openstack-galera-0" Nov 28 13:35:03 crc kubenswrapper[4631]: I1128 13:35:03.922857 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/1a711f62-44fc-4046-a3aa-b5ef77b45e62-kolla-config\") pod \"openstack-galera-0\" (UID: \"1a711f62-44fc-4046-a3aa-b5ef77b45e62\") " pod="openstack/openstack-galera-0" Nov 28 13:35:03 crc kubenswrapper[4631]: I1128 13:35:03.918248 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/1a711f62-44fc-4046-a3aa-b5ef77b45e62-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"1a711f62-44fc-4046-a3aa-b5ef77b45e62\") " pod="openstack/openstack-galera-0" Nov 28 13:35:03 crc kubenswrapper[4631]: I1128 13:35:03.929251 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/1a711f62-44fc-4046-a3aa-b5ef77b45e62-config-data-generated\") pod \"openstack-galera-0\" (UID: \"1a711f62-44fc-4046-a3aa-b5ef77b45e62\") " pod="openstack/openstack-galera-0" Nov 28 13:35:03 crc kubenswrapper[4631]: I1128 13:35:03.929682 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/1a711f62-44fc-4046-a3aa-b5ef77b45e62-config-data-default\") pod \"openstack-galera-0\" (UID: \"1a711f62-44fc-4046-a3aa-b5ef77b45e62\") " pod="openstack/openstack-galera-0" Nov 28 13:35:03 crc kubenswrapper[4631]: I1128 13:35:03.924477 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1a711f62-44fc-4046-a3aa-b5ef77b45e62-operator-scripts\") pod \"openstack-galera-0\" (UID: \"1a711f62-44fc-4046-a3aa-b5ef77b45e62\") " pod="openstack/openstack-galera-0" Nov 28 13:35:03 crc kubenswrapper[4631]: I1128 13:35:03.949879 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a711f62-44fc-4046-a3aa-b5ef77b45e62-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"1a711f62-44fc-4046-a3aa-b5ef77b45e62\") " pod="openstack/openstack-galera-0" Nov 28 13:35:03 crc kubenswrapper[4631]: I1128 13:35:03.999617 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 28 13:35:04 crc kubenswrapper[4631]: I1128 13:35:04.007068 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kpprt\" (UniqueName: \"kubernetes.io/projected/1a711f62-44fc-4046-a3aa-b5ef77b45e62-kube-api-access-kpprt\") pod \"openstack-galera-0\" (UID: \"1a711f62-44fc-4046-a3aa-b5ef77b45e62\") " pod="openstack/openstack-galera-0" Nov 28 13:35:04 crc kubenswrapper[4631]: I1128 13:35:04.013963 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-galera-0\" (UID: \"1a711f62-44fc-4046-a3aa-b5ef77b45e62\") " pod="openstack/openstack-galera-0" Nov 28 13:35:04 crc kubenswrapper[4631]: I1128 13:35:04.067100 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Nov 28 13:35:04 crc kubenswrapper[4631]: I1128 13:35:04.601268 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"ed4141f2-5eff-4252-8dc7-ad60a1580189","Type":"ContainerStarted","Data":"f3f6ed0a79454bb6e8514a61efc6444fdc3d68ebeae1172304b680407d122dff"} Nov 28 13:35:04 crc kubenswrapper[4631]: I1128 13:35:04.685041 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-5kbsx" Nov 28 13:35:04 crc kubenswrapper[4631]: I1128 13:35:04.686305 4631 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-5kbsx" Nov 28 13:35:04 crc kubenswrapper[4631]: I1128 13:35:04.698122 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Nov 28 13:35:04 crc kubenswrapper[4631]: W1128 13:35:04.760658 4631 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1a711f62_44fc_4046_a3aa_b5ef77b45e62.slice/crio-5de16137842de7cd0d25c4686388638295a47e7d271a33759a47cedbf6c716eb WatchSource:0}: Error finding container 5de16137842de7cd0d25c4686388638295a47e7d271a33759a47cedbf6c716eb: Status 404 returned error can't find the container with id 5de16137842de7cd0d25c4686388638295a47e7d271a33759a47cedbf6c716eb Nov 28 13:35:04 crc kubenswrapper[4631]: I1128 13:35:04.870002 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Nov 28 13:35:04 crc kubenswrapper[4631]: I1128 13:35:04.875060 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Nov 28 13:35:04 crc kubenswrapper[4631]: I1128 13:35:04.899252 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Nov 28 13:35:04 crc kubenswrapper[4631]: I1128 13:35:04.899493 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Nov 28 13:35:04 crc kubenswrapper[4631]: I1128 13:35:04.899659 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-jhwlh" Nov 28 13:35:04 crc kubenswrapper[4631]: I1128 13:35:04.899698 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Nov 28 13:35:04 crc kubenswrapper[4631]: I1128 13:35:04.930946 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Nov 28 13:35:05 crc kubenswrapper[4631]: I1128 13:35:05.065304 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4bc20a10-95d5-4003-86d9-c6848bc1b188-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"4bc20a10-95d5-4003-86d9-c6848bc1b188\") " pod="openstack/openstack-cell1-galera-0" Nov 28 13:35:05 crc kubenswrapper[4631]: I1128 13:35:05.065398 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/4bc20a10-95d5-4003-86d9-c6848bc1b188-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"4bc20a10-95d5-4003-86d9-c6848bc1b188\") " pod="openstack/openstack-cell1-galera-0" Nov 28 13:35:05 crc kubenswrapper[4631]: I1128 13:35:05.065553 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4bc20a10-95d5-4003-86d9-c6848bc1b188-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"4bc20a10-95d5-4003-86d9-c6848bc1b188\") " pod="openstack/openstack-cell1-galera-0" Nov 28 13:35:05 crc kubenswrapper[4631]: I1128 13:35:05.065595 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/4bc20a10-95d5-4003-86d9-c6848bc1b188-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"4bc20a10-95d5-4003-86d9-c6848bc1b188\") " pod="openstack/openstack-cell1-galera-0" Nov 28 13:35:05 crc kubenswrapper[4631]: I1128 13:35:05.065690 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/4bc20a10-95d5-4003-86d9-c6848bc1b188-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"4bc20a10-95d5-4003-86d9-c6848bc1b188\") " pod="openstack/openstack-cell1-galera-0" Nov 28 13:35:05 crc kubenswrapper[4631]: I1128 13:35:05.065746 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"openstack-cell1-galera-0\" (UID: \"4bc20a10-95d5-4003-86d9-c6848bc1b188\") " pod="openstack/openstack-cell1-galera-0" Nov 28 13:35:05 crc kubenswrapper[4631]: I1128 13:35:05.065823 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/4bc20a10-95d5-4003-86d9-c6848bc1b188-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"4bc20a10-95d5-4003-86d9-c6848bc1b188\") " pod="openstack/openstack-cell1-galera-0" Nov 28 13:35:05 crc kubenswrapper[4631]: I1128 13:35:05.065878 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fdm8t\" (UniqueName: \"kubernetes.io/projected/4bc20a10-95d5-4003-86d9-c6848bc1b188-kube-api-access-fdm8t\") pod \"openstack-cell1-galera-0\" (UID: \"4bc20a10-95d5-4003-86d9-c6848bc1b188\") " pod="openstack/openstack-cell1-galera-0" Nov 28 13:35:05 crc kubenswrapper[4631]: I1128 13:35:05.174338 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4bc20a10-95d5-4003-86d9-c6848bc1b188-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"4bc20a10-95d5-4003-86d9-c6848bc1b188\") " pod="openstack/openstack-cell1-galera-0" Nov 28 13:35:05 crc kubenswrapper[4631]: I1128 13:35:05.174394 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/4bc20a10-95d5-4003-86d9-c6848bc1b188-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"4bc20a10-95d5-4003-86d9-c6848bc1b188\") " pod="openstack/openstack-cell1-galera-0" Nov 28 13:35:05 crc kubenswrapper[4631]: I1128 13:35:05.174451 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/4bc20a10-95d5-4003-86d9-c6848bc1b188-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"4bc20a10-95d5-4003-86d9-c6848bc1b188\") " pod="openstack/openstack-cell1-galera-0" Nov 28 13:35:05 crc kubenswrapper[4631]: I1128 13:35:05.174485 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"openstack-cell1-galera-0\" (UID: \"4bc20a10-95d5-4003-86d9-c6848bc1b188\") " pod="openstack/openstack-cell1-galera-0" Nov 28 13:35:05 crc kubenswrapper[4631]: I1128 13:35:05.174523 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/4bc20a10-95d5-4003-86d9-c6848bc1b188-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"4bc20a10-95d5-4003-86d9-c6848bc1b188\") " pod="openstack/openstack-cell1-galera-0" Nov 28 13:35:05 crc kubenswrapper[4631]: I1128 13:35:05.174554 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fdm8t\" (UniqueName: \"kubernetes.io/projected/4bc20a10-95d5-4003-86d9-c6848bc1b188-kube-api-access-fdm8t\") pod \"openstack-cell1-galera-0\" (UID: \"4bc20a10-95d5-4003-86d9-c6848bc1b188\") " pod="openstack/openstack-cell1-galera-0" Nov 28 13:35:05 crc kubenswrapper[4631]: I1128 13:35:05.174592 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4bc20a10-95d5-4003-86d9-c6848bc1b188-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"4bc20a10-95d5-4003-86d9-c6848bc1b188\") " pod="openstack/openstack-cell1-galera-0" Nov 28 13:35:05 crc kubenswrapper[4631]: I1128 13:35:05.174613 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/4bc20a10-95d5-4003-86d9-c6848bc1b188-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"4bc20a10-95d5-4003-86d9-c6848bc1b188\") " pod="openstack/openstack-cell1-galera-0" Nov 28 13:35:05 crc kubenswrapper[4631]: I1128 13:35:05.176993 4631 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"openstack-cell1-galera-0\" (UID: \"4bc20a10-95d5-4003-86d9-c6848bc1b188\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/openstack-cell1-galera-0" Nov 28 13:35:05 crc kubenswrapper[4631]: I1128 13:35:05.178989 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/4bc20a10-95d5-4003-86d9-c6848bc1b188-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"4bc20a10-95d5-4003-86d9-c6848bc1b188\") " pod="openstack/openstack-cell1-galera-0" Nov 28 13:35:05 crc kubenswrapper[4631]: I1128 13:35:05.179138 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/4bc20a10-95d5-4003-86d9-c6848bc1b188-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"4bc20a10-95d5-4003-86d9-c6848bc1b188\") " pod="openstack/openstack-cell1-galera-0" Nov 28 13:35:05 crc kubenswrapper[4631]: I1128 13:35:05.179380 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/4bc20a10-95d5-4003-86d9-c6848bc1b188-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"4bc20a10-95d5-4003-86d9-c6848bc1b188\") " pod="openstack/openstack-cell1-galera-0" Nov 28 13:35:05 crc kubenswrapper[4631]: I1128 13:35:05.180740 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4bc20a10-95d5-4003-86d9-c6848bc1b188-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"4bc20a10-95d5-4003-86d9-c6848bc1b188\") " pod="openstack/openstack-cell1-galera-0" Nov 28 13:35:05 crc kubenswrapper[4631]: I1128 13:35:05.197861 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/4bc20a10-95d5-4003-86d9-c6848bc1b188-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"4bc20a10-95d5-4003-86d9-c6848bc1b188\") " pod="openstack/openstack-cell1-galera-0" Nov 28 13:35:05 crc kubenswrapper[4631]: I1128 13:35:05.205657 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4bc20a10-95d5-4003-86d9-c6848bc1b188-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"4bc20a10-95d5-4003-86d9-c6848bc1b188\") " pod="openstack/openstack-cell1-galera-0" Nov 28 13:35:05 crc kubenswrapper[4631]: I1128 13:35:05.277793 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fdm8t\" (UniqueName: \"kubernetes.io/projected/4bc20a10-95d5-4003-86d9-c6848bc1b188-kube-api-access-fdm8t\") pod \"openstack-cell1-galera-0\" (UID: \"4bc20a10-95d5-4003-86d9-c6848bc1b188\") " pod="openstack/openstack-cell1-galera-0" Nov 28 13:35:05 crc kubenswrapper[4631]: I1128 13:35:05.289407 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"openstack-cell1-galera-0\" (UID: \"4bc20a10-95d5-4003-86d9-c6848bc1b188\") " pod="openstack/openstack-cell1-galera-0" Nov 28 13:35:05 crc kubenswrapper[4631]: I1128 13:35:05.502362 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Nov 28 13:35:05 crc kubenswrapper[4631]: I1128 13:35:05.609055 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Nov 28 13:35:05 crc kubenswrapper[4631]: I1128 13:35:05.616818 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Nov 28 13:35:05 crc kubenswrapper[4631]: I1128 13:35:05.633771 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-gwncw" Nov 28 13:35:05 crc kubenswrapper[4631]: I1128 13:35:05.634700 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Nov 28 13:35:05 crc kubenswrapper[4631]: I1128 13:35:05.649809 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Nov 28 13:35:05 crc kubenswrapper[4631]: I1128 13:35:05.667469 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Nov 28 13:35:05 crc kubenswrapper[4631]: I1128 13:35:05.686570 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/d35228a7-a0fe-496d-9d62-5f1d5a44c72f-kolla-config\") pod \"memcached-0\" (UID: \"d35228a7-a0fe-496d-9d62-5f1d5a44c72f\") " pod="openstack/memcached-0" Nov 28 13:35:05 crc kubenswrapper[4631]: I1128 13:35:05.686626 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8ppj8\" (UniqueName: \"kubernetes.io/projected/d35228a7-a0fe-496d-9d62-5f1d5a44c72f-kube-api-access-8ppj8\") pod \"memcached-0\" (UID: \"d35228a7-a0fe-496d-9d62-5f1d5a44c72f\") " pod="openstack/memcached-0" Nov 28 13:35:05 crc kubenswrapper[4631]: I1128 13:35:05.686683 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/d35228a7-a0fe-496d-9d62-5f1d5a44c72f-memcached-tls-certs\") pod \"memcached-0\" (UID: \"d35228a7-a0fe-496d-9d62-5f1d5a44c72f\") " pod="openstack/memcached-0" Nov 28 13:35:05 crc kubenswrapper[4631]: I1128 13:35:05.686753 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d35228a7-a0fe-496d-9d62-5f1d5a44c72f-config-data\") pod \"memcached-0\" (UID: \"d35228a7-a0fe-496d-9d62-5f1d5a44c72f\") " pod="openstack/memcached-0" Nov 28 13:35:05 crc kubenswrapper[4631]: I1128 13:35:05.686776 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d35228a7-a0fe-496d-9d62-5f1d5a44c72f-combined-ca-bundle\") pod \"memcached-0\" (UID: \"d35228a7-a0fe-496d-9d62-5f1d5a44c72f\") " pod="openstack/memcached-0" Nov 28 13:35:05 crc kubenswrapper[4631]: I1128 13:35:05.747665 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"1a711f62-44fc-4046-a3aa-b5ef77b45e62","Type":"ContainerStarted","Data":"5de16137842de7cd0d25c4686388638295a47e7d271a33759a47cedbf6c716eb"} Nov 28 13:35:05 crc kubenswrapper[4631]: I1128 13:35:05.789528 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d35228a7-a0fe-496d-9d62-5f1d5a44c72f-combined-ca-bundle\") pod \"memcached-0\" (UID: \"d35228a7-a0fe-496d-9d62-5f1d5a44c72f\") " pod="openstack/memcached-0" Nov 28 13:35:05 crc kubenswrapper[4631]: I1128 13:35:05.789599 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d35228a7-a0fe-496d-9d62-5f1d5a44c72f-config-data\") pod \"memcached-0\" (UID: \"d35228a7-a0fe-496d-9d62-5f1d5a44c72f\") " pod="openstack/memcached-0" Nov 28 13:35:05 crc kubenswrapper[4631]: I1128 13:35:05.789647 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/d35228a7-a0fe-496d-9d62-5f1d5a44c72f-kolla-config\") pod \"memcached-0\" (UID: \"d35228a7-a0fe-496d-9d62-5f1d5a44c72f\") " pod="openstack/memcached-0" Nov 28 13:35:05 crc kubenswrapper[4631]: I1128 13:35:05.789678 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8ppj8\" (UniqueName: \"kubernetes.io/projected/d35228a7-a0fe-496d-9d62-5f1d5a44c72f-kube-api-access-8ppj8\") pod \"memcached-0\" (UID: \"d35228a7-a0fe-496d-9d62-5f1d5a44c72f\") " pod="openstack/memcached-0" Nov 28 13:35:05 crc kubenswrapper[4631]: I1128 13:35:05.789743 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/d35228a7-a0fe-496d-9d62-5f1d5a44c72f-memcached-tls-certs\") pod \"memcached-0\" (UID: \"d35228a7-a0fe-496d-9d62-5f1d5a44c72f\") " pod="openstack/memcached-0" Nov 28 13:35:05 crc kubenswrapper[4631]: I1128 13:35:05.791031 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/d35228a7-a0fe-496d-9d62-5f1d5a44c72f-kolla-config\") pod \"memcached-0\" (UID: \"d35228a7-a0fe-496d-9d62-5f1d5a44c72f\") " pod="openstack/memcached-0" Nov 28 13:35:05 crc kubenswrapper[4631]: I1128 13:35:05.791616 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d35228a7-a0fe-496d-9d62-5f1d5a44c72f-config-data\") pod \"memcached-0\" (UID: \"d35228a7-a0fe-496d-9d62-5f1d5a44c72f\") " pod="openstack/memcached-0" Nov 28 13:35:05 crc kubenswrapper[4631]: I1128 13:35:05.819267 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/d35228a7-a0fe-496d-9d62-5f1d5a44c72f-memcached-tls-certs\") pod \"memcached-0\" (UID: \"d35228a7-a0fe-496d-9d62-5f1d5a44c72f\") " pod="openstack/memcached-0" Nov 28 13:35:05 crc kubenswrapper[4631]: I1128 13:35:05.820028 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d35228a7-a0fe-496d-9d62-5f1d5a44c72f-combined-ca-bundle\") pod \"memcached-0\" (UID: \"d35228a7-a0fe-496d-9d62-5f1d5a44c72f\") " pod="openstack/memcached-0" Nov 28 13:35:05 crc kubenswrapper[4631]: I1128 13:35:05.902847 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8ppj8\" (UniqueName: \"kubernetes.io/projected/d35228a7-a0fe-496d-9d62-5f1d5a44c72f-kube-api-access-8ppj8\") pod \"memcached-0\" (UID: \"d35228a7-a0fe-496d-9d62-5f1d5a44c72f\") " pod="openstack/memcached-0" Nov 28 13:35:05 crc kubenswrapper[4631]: I1128 13:35:05.948875 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Nov 28 13:35:05 crc kubenswrapper[4631]: I1128 13:35:05.949537 4631 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-5kbsx" podUID="03ba4ed9-11ea-4c20-baaf-8975710c6708" containerName="registry-server" probeResult="failure" output=< Nov 28 13:35:05 crc kubenswrapper[4631]: timeout: failed to connect service ":50051" within 1s Nov 28 13:35:05 crc kubenswrapper[4631]: > Nov 28 13:35:06 crc kubenswrapper[4631]: I1128 13:35:06.938039 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Nov 28 13:35:07 crc kubenswrapper[4631]: I1128 13:35:07.158238 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Nov 28 13:35:07 crc kubenswrapper[4631]: I1128 13:35:07.959429 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"4bc20a10-95d5-4003-86d9-c6848bc1b188","Type":"ContainerStarted","Data":"9bc61412f3e0bb50a5da0439bdde433d14f15a7aea5f7876b8e5d8ab9e3e40bb"} Nov 28 13:35:07 crc kubenswrapper[4631]: I1128 13:35:07.976180 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"d35228a7-a0fe-496d-9d62-5f1d5a44c72f","Type":"ContainerStarted","Data":"74d85868c5361bdaf80809c8f33be45fd591c59f3d2faae51fb26d0d0d54ab55"} Nov 28 13:35:08 crc kubenswrapper[4631]: I1128 13:35:08.203968 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Nov 28 13:35:08 crc kubenswrapper[4631]: I1128 13:35:08.209256 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 28 13:35:08 crc kubenswrapper[4631]: I1128 13:35:08.230617 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-gz5w4" Nov 28 13:35:08 crc kubenswrapper[4631]: I1128 13:35:08.251137 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 28 13:35:08 crc kubenswrapper[4631]: I1128 13:35:08.373006 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xj59v\" (UniqueName: \"kubernetes.io/projected/83d77da3-fbb7-434f-abaf-5be9b599a9bf-kube-api-access-xj59v\") pod \"kube-state-metrics-0\" (UID: \"83d77da3-fbb7-434f-abaf-5be9b599a9bf\") " pod="openstack/kube-state-metrics-0" Nov 28 13:35:08 crc kubenswrapper[4631]: I1128 13:35:08.483277 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xj59v\" (UniqueName: \"kubernetes.io/projected/83d77da3-fbb7-434f-abaf-5be9b599a9bf-kube-api-access-xj59v\") pod \"kube-state-metrics-0\" (UID: \"83d77da3-fbb7-434f-abaf-5be9b599a9bf\") " pod="openstack/kube-state-metrics-0" Nov 28 13:35:08 crc kubenswrapper[4631]: I1128 13:35:08.541080 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xj59v\" (UniqueName: \"kubernetes.io/projected/83d77da3-fbb7-434f-abaf-5be9b599a9bf-kube-api-access-xj59v\") pod \"kube-state-metrics-0\" (UID: \"83d77da3-fbb7-434f-abaf-5be9b599a9bf\") " pod="openstack/kube-state-metrics-0" Nov 28 13:35:08 crc kubenswrapper[4631]: I1128 13:35:08.575800 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 28 13:35:09 crc kubenswrapper[4631]: I1128 13:35:09.726106 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 28 13:35:09 crc kubenswrapper[4631]: W1128 13:35:09.765499 4631 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod83d77da3_fbb7_434f_abaf_5be9b599a9bf.slice/crio-0feb5123c599b3af93212c76b682c0ee877de3076bacb746e6bad20ec87f297f WatchSource:0}: Error finding container 0feb5123c599b3af93212c76b682c0ee877de3076bacb746e6bad20ec87f297f: Status 404 returned error can't find the container with id 0feb5123c599b3af93212c76b682c0ee877de3076bacb746e6bad20ec87f297f Nov 28 13:35:10 crc kubenswrapper[4631]: I1128 13:35:10.074173 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"83d77da3-fbb7-434f-abaf-5be9b599a9bf","Type":"ContainerStarted","Data":"0feb5123c599b3af93212c76b682c0ee877de3076bacb746e6bad20ec87f297f"} Nov 28 13:35:11 crc kubenswrapper[4631]: I1128 13:35:11.876665 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Nov 28 13:35:11 crc kubenswrapper[4631]: I1128 13:35:11.881789 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Nov 28 13:35:11 crc kubenswrapper[4631]: I1128 13:35:11.892369 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Nov 28 13:35:11 crc kubenswrapper[4631]: I1128 13:35:11.893153 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Nov 28 13:35:11 crc kubenswrapper[4631]: I1128 13:35:11.893659 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Nov 28 13:35:11 crc kubenswrapper[4631]: I1128 13:35:11.896811 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-zbzrf" Nov 28 13:35:11 crc kubenswrapper[4631]: I1128 13:35:11.897042 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Nov 28 13:35:11 crc kubenswrapper[4631]: I1128 13:35:11.982167 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/ca11061d-b0a1-464d-9e1d-5572fc439580-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"ca11061d-b0a1-464d-9e1d-5572fc439580\") " pod="openstack/ovsdbserver-nb-0" Nov 28 13:35:11 crc kubenswrapper[4631]: I1128 13:35:11.982254 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca11061d-b0a1-464d-9e1d-5572fc439580-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"ca11061d-b0a1-464d-9e1d-5572fc439580\") " pod="openstack/ovsdbserver-nb-0" Nov 28 13:35:11 crc kubenswrapper[4631]: I1128 13:35:11.982280 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ca11061d-b0a1-464d-9e1d-5572fc439580-config\") pod \"ovsdbserver-nb-0\" (UID: \"ca11061d-b0a1-464d-9e1d-5572fc439580\") " pod="openstack/ovsdbserver-nb-0" Nov 28 13:35:11 crc kubenswrapper[4631]: I1128 13:35:11.982321 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mqt7w\" (UniqueName: \"kubernetes.io/projected/ca11061d-b0a1-464d-9e1d-5572fc439580-kube-api-access-mqt7w\") pod \"ovsdbserver-nb-0\" (UID: \"ca11061d-b0a1-464d-9e1d-5572fc439580\") " pod="openstack/ovsdbserver-nb-0" Nov 28 13:35:11 crc kubenswrapper[4631]: I1128 13:35:11.982350 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"ovsdbserver-nb-0\" (UID: \"ca11061d-b0a1-464d-9e1d-5572fc439580\") " pod="openstack/ovsdbserver-nb-0" Nov 28 13:35:11 crc kubenswrapper[4631]: I1128 13:35:11.982402 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/ca11061d-b0a1-464d-9e1d-5572fc439580-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"ca11061d-b0a1-464d-9e1d-5572fc439580\") " pod="openstack/ovsdbserver-nb-0" Nov 28 13:35:11 crc kubenswrapper[4631]: I1128 13:35:11.982417 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/ca11061d-b0a1-464d-9e1d-5572fc439580-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"ca11061d-b0a1-464d-9e1d-5572fc439580\") " pod="openstack/ovsdbserver-nb-0" Nov 28 13:35:11 crc kubenswrapper[4631]: I1128 13:35:11.982468 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ca11061d-b0a1-464d-9e1d-5572fc439580-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"ca11061d-b0a1-464d-9e1d-5572fc439580\") " pod="openstack/ovsdbserver-nb-0" Nov 28 13:35:12 crc kubenswrapper[4631]: I1128 13:35:12.090105 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ca11061d-b0a1-464d-9e1d-5572fc439580-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"ca11061d-b0a1-464d-9e1d-5572fc439580\") " pod="openstack/ovsdbserver-nb-0" Nov 28 13:35:12 crc kubenswrapper[4631]: I1128 13:35:12.090202 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/ca11061d-b0a1-464d-9e1d-5572fc439580-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"ca11061d-b0a1-464d-9e1d-5572fc439580\") " pod="openstack/ovsdbserver-nb-0" Nov 28 13:35:12 crc kubenswrapper[4631]: I1128 13:35:12.090297 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca11061d-b0a1-464d-9e1d-5572fc439580-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"ca11061d-b0a1-464d-9e1d-5572fc439580\") " pod="openstack/ovsdbserver-nb-0" Nov 28 13:35:12 crc kubenswrapper[4631]: I1128 13:35:12.090329 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ca11061d-b0a1-464d-9e1d-5572fc439580-config\") pod \"ovsdbserver-nb-0\" (UID: \"ca11061d-b0a1-464d-9e1d-5572fc439580\") " pod="openstack/ovsdbserver-nb-0" Nov 28 13:35:12 crc kubenswrapper[4631]: I1128 13:35:12.090380 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mqt7w\" (UniqueName: \"kubernetes.io/projected/ca11061d-b0a1-464d-9e1d-5572fc439580-kube-api-access-mqt7w\") pod \"ovsdbserver-nb-0\" (UID: \"ca11061d-b0a1-464d-9e1d-5572fc439580\") " pod="openstack/ovsdbserver-nb-0" Nov 28 13:35:12 crc kubenswrapper[4631]: I1128 13:35:12.090434 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"ovsdbserver-nb-0\" (UID: \"ca11061d-b0a1-464d-9e1d-5572fc439580\") " pod="openstack/ovsdbserver-nb-0" Nov 28 13:35:12 crc kubenswrapper[4631]: I1128 13:35:12.090552 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/ca11061d-b0a1-464d-9e1d-5572fc439580-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"ca11061d-b0a1-464d-9e1d-5572fc439580\") " pod="openstack/ovsdbserver-nb-0" Nov 28 13:35:12 crc kubenswrapper[4631]: I1128 13:35:12.090572 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/ca11061d-b0a1-464d-9e1d-5572fc439580-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"ca11061d-b0a1-464d-9e1d-5572fc439580\") " pod="openstack/ovsdbserver-nb-0" Nov 28 13:35:12 crc kubenswrapper[4631]: I1128 13:35:12.095454 4631 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"ovsdbserver-nb-0\" (UID: \"ca11061d-b0a1-464d-9e1d-5572fc439580\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/ovsdbserver-nb-0" Nov 28 13:35:12 crc kubenswrapper[4631]: I1128 13:35:12.188097 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ca11061d-b0a1-464d-9e1d-5572fc439580-config\") pod \"ovsdbserver-nb-0\" (UID: \"ca11061d-b0a1-464d-9e1d-5572fc439580\") " pod="openstack/ovsdbserver-nb-0" Nov 28 13:35:12 crc kubenswrapper[4631]: I1128 13:35:12.188862 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/ca11061d-b0a1-464d-9e1d-5572fc439580-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"ca11061d-b0a1-464d-9e1d-5572fc439580\") " pod="openstack/ovsdbserver-nb-0" Nov 28 13:35:12 crc kubenswrapper[4631]: I1128 13:35:12.191221 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ca11061d-b0a1-464d-9e1d-5572fc439580-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"ca11061d-b0a1-464d-9e1d-5572fc439580\") " pod="openstack/ovsdbserver-nb-0" Nov 28 13:35:12 crc kubenswrapper[4631]: I1128 13:35:12.229245 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca11061d-b0a1-464d-9e1d-5572fc439580-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"ca11061d-b0a1-464d-9e1d-5572fc439580\") " pod="openstack/ovsdbserver-nb-0" Nov 28 13:35:12 crc kubenswrapper[4631]: I1128 13:35:12.229553 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/ca11061d-b0a1-464d-9e1d-5572fc439580-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"ca11061d-b0a1-464d-9e1d-5572fc439580\") " pod="openstack/ovsdbserver-nb-0" Nov 28 13:35:12 crc kubenswrapper[4631]: I1128 13:35:12.230150 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mqt7w\" (UniqueName: \"kubernetes.io/projected/ca11061d-b0a1-464d-9e1d-5572fc439580-kube-api-access-mqt7w\") pod \"ovsdbserver-nb-0\" (UID: \"ca11061d-b0a1-464d-9e1d-5572fc439580\") " pod="openstack/ovsdbserver-nb-0" Nov 28 13:35:12 crc kubenswrapper[4631]: I1128 13:35:12.258687 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/ca11061d-b0a1-464d-9e1d-5572fc439580-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"ca11061d-b0a1-464d-9e1d-5572fc439580\") " pod="openstack/ovsdbserver-nb-0" Nov 28 13:35:12 crc kubenswrapper[4631]: I1128 13:35:12.273268 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Nov 28 13:35:12 crc kubenswrapper[4631]: I1128 13:35:12.285320 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"ovsdbserver-nb-0\" (UID: \"ca11061d-b0a1-464d-9e1d-5572fc439580\") " pod="openstack/ovsdbserver-nb-0" Nov 28 13:35:12 crc kubenswrapper[4631]: I1128 13:35:12.490173 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Nov 28 13:35:13 crc kubenswrapper[4631]: I1128 13:35:13.583555 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-7jgdf"] Nov 28 13:35:13 crc kubenswrapper[4631]: I1128 13:35:13.584904 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-7jgdf"] Nov 28 13:35:13 crc kubenswrapper[4631]: I1128 13:35:13.585008 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-7jgdf" Nov 28 13:35:13 crc kubenswrapper[4631]: I1128 13:35:13.593038 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Nov 28 13:35:13 crc kubenswrapper[4631]: I1128 13:35:13.593218 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovncontroller-ovndbs" Nov 28 13:35:13 crc kubenswrapper[4631]: I1128 13:35:13.600308 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-wswhm" Nov 28 13:35:13 crc kubenswrapper[4631]: I1128 13:35:13.640823 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/7fc5dcd5-fde5-4a72-8d4a-cda51785e77f-var-log-ovn\") pod \"ovn-controller-7jgdf\" (UID: \"7fc5dcd5-fde5-4a72-8d4a-cda51785e77f\") " pod="openstack/ovn-controller-7jgdf" Nov 28 13:35:13 crc kubenswrapper[4631]: I1128 13:35:13.640942 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/7fc5dcd5-fde5-4a72-8d4a-cda51785e77f-var-run\") pod \"ovn-controller-7jgdf\" (UID: \"7fc5dcd5-fde5-4a72-8d4a-cda51785e77f\") " pod="openstack/ovn-controller-7jgdf" Nov 28 13:35:13 crc kubenswrapper[4631]: I1128 13:35:13.641052 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7fc5dcd5-fde5-4a72-8d4a-cda51785e77f-combined-ca-bundle\") pod \"ovn-controller-7jgdf\" (UID: \"7fc5dcd5-fde5-4a72-8d4a-cda51785e77f\") " pod="openstack/ovn-controller-7jgdf" Nov 28 13:35:13 crc kubenswrapper[4631]: I1128 13:35:13.641122 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/7fc5dcd5-fde5-4a72-8d4a-cda51785e77f-var-run-ovn\") pod \"ovn-controller-7jgdf\" (UID: \"7fc5dcd5-fde5-4a72-8d4a-cda51785e77f\") " pod="openstack/ovn-controller-7jgdf" Nov 28 13:35:13 crc kubenswrapper[4631]: I1128 13:35:13.641142 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7fc5dcd5-fde5-4a72-8d4a-cda51785e77f-scripts\") pod \"ovn-controller-7jgdf\" (UID: \"7fc5dcd5-fde5-4a72-8d4a-cda51785e77f\") " pod="openstack/ovn-controller-7jgdf" Nov 28 13:35:13 crc kubenswrapper[4631]: I1128 13:35:13.641205 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/7fc5dcd5-fde5-4a72-8d4a-cda51785e77f-ovn-controller-tls-certs\") pod \"ovn-controller-7jgdf\" (UID: \"7fc5dcd5-fde5-4a72-8d4a-cda51785e77f\") " pod="openstack/ovn-controller-7jgdf" Nov 28 13:35:13 crc kubenswrapper[4631]: I1128 13:35:13.641227 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-56tdl\" (UniqueName: \"kubernetes.io/projected/7fc5dcd5-fde5-4a72-8d4a-cda51785e77f-kube-api-access-56tdl\") pod \"ovn-controller-7jgdf\" (UID: \"7fc5dcd5-fde5-4a72-8d4a-cda51785e77f\") " pod="openstack/ovn-controller-7jgdf" Nov 28 13:35:13 crc kubenswrapper[4631]: I1128 13:35:13.657484 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-lskqw"] Nov 28 13:35:13 crc kubenswrapper[4631]: I1128 13:35:13.732048 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-lskqw"] Nov 28 13:35:13 crc kubenswrapper[4631]: I1128 13:35:13.732248 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-lskqw" Nov 28 13:35:13 crc kubenswrapper[4631]: I1128 13:35:13.775687 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/7fc5dcd5-fde5-4a72-8d4a-cda51785e77f-var-log-ovn\") pod \"ovn-controller-7jgdf\" (UID: \"7fc5dcd5-fde5-4a72-8d4a-cda51785e77f\") " pod="openstack/ovn-controller-7jgdf" Nov 28 13:35:13 crc kubenswrapper[4631]: I1128 13:35:13.775784 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g4qql\" (UniqueName: \"kubernetes.io/projected/7fc0ee4e-435d-4297-94eb-9e8c6fa90004-kube-api-access-g4qql\") pod \"ovn-controller-ovs-lskqw\" (UID: \"7fc0ee4e-435d-4297-94eb-9e8c6fa90004\") " pod="openstack/ovn-controller-ovs-lskqw" Nov 28 13:35:13 crc kubenswrapper[4631]: I1128 13:35:13.775826 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/7fc5dcd5-fde5-4a72-8d4a-cda51785e77f-var-run\") pod \"ovn-controller-7jgdf\" (UID: \"7fc5dcd5-fde5-4a72-8d4a-cda51785e77f\") " pod="openstack/ovn-controller-7jgdf" Nov 28 13:35:13 crc kubenswrapper[4631]: I1128 13:35:13.775854 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/7fc0ee4e-435d-4297-94eb-9e8c6fa90004-var-log\") pod \"ovn-controller-ovs-lskqw\" (UID: \"7fc0ee4e-435d-4297-94eb-9e8c6fa90004\") " pod="openstack/ovn-controller-ovs-lskqw" Nov 28 13:35:13 crc kubenswrapper[4631]: I1128 13:35:13.776010 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7fc5dcd5-fde5-4a72-8d4a-cda51785e77f-combined-ca-bundle\") pod \"ovn-controller-7jgdf\" (UID: \"7fc5dcd5-fde5-4a72-8d4a-cda51785e77f\") " pod="openstack/ovn-controller-7jgdf" Nov 28 13:35:13 crc kubenswrapper[4631]: I1128 13:35:13.776046 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/7fc5dcd5-fde5-4a72-8d4a-cda51785e77f-var-run-ovn\") pod \"ovn-controller-7jgdf\" (UID: \"7fc5dcd5-fde5-4a72-8d4a-cda51785e77f\") " pod="openstack/ovn-controller-7jgdf" Nov 28 13:35:13 crc kubenswrapper[4631]: I1128 13:35:13.776072 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7fc5dcd5-fde5-4a72-8d4a-cda51785e77f-scripts\") pod \"ovn-controller-7jgdf\" (UID: \"7fc5dcd5-fde5-4a72-8d4a-cda51785e77f\") " pod="openstack/ovn-controller-7jgdf" Nov 28 13:35:13 crc kubenswrapper[4631]: I1128 13:35:13.776098 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/7fc0ee4e-435d-4297-94eb-9e8c6fa90004-etc-ovs\") pod \"ovn-controller-ovs-lskqw\" (UID: \"7fc0ee4e-435d-4297-94eb-9e8c6fa90004\") " pod="openstack/ovn-controller-ovs-lskqw" Nov 28 13:35:13 crc kubenswrapper[4631]: I1128 13:35:13.776126 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/7fc0ee4e-435d-4297-94eb-9e8c6fa90004-var-run\") pod \"ovn-controller-ovs-lskqw\" (UID: \"7fc0ee4e-435d-4297-94eb-9e8c6fa90004\") " pod="openstack/ovn-controller-ovs-lskqw" Nov 28 13:35:13 crc kubenswrapper[4631]: I1128 13:35:13.776181 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/7fc0ee4e-435d-4297-94eb-9e8c6fa90004-var-lib\") pod \"ovn-controller-ovs-lskqw\" (UID: \"7fc0ee4e-435d-4297-94eb-9e8c6fa90004\") " pod="openstack/ovn-controller-ovs-lskqw" Nov 28 13:35:13 crc kubenswrapper[4631]: I1128 13:35:13.776219 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-56tdl\" (UniqueName: \"kubernetes.io/projected/7fc5dcd5-fde5-4a72-8d4a-cda51785e77f-kube-api-access-56tdl\") pod \"ovn-controller-7jgdf\" (UID: \"7fc5dcd5-fde5-4a72-8d4a-cda51785e77f\") " pod="openstack/ovn-controller-7jgdf" Nov 28 13:35:13 crc kubenswrapper[4631]: I1128 13:35:13.776237 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/7fc5dcd5-fde5-4a72-8d4a-cda51785e77f-ovn-controller-tls-certs\") pod \"ovn-controller-7jgdf\" (UID: \"7fc5dcd5-fde5-4a72-8d4a-cda51785e77f\") " pod="openstack/ovn-controller-7jgdf" Nov 28 13:35:13 crc kubenswrapper[4631]: I1128 13:35:13.776345 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7fc0ee4e-435d-4297-94eb-9e8c6fa90004-scripts\") pod \"ovn-controller-ovs-lskqw\" (UID: \"7fc0ee4e-435d-4297-94eb-9e8c6fa90004\") " pod="openstack/ovn-controller-ovs-lskqw" Nov 28 13:35:13 crc kubenswrapper[4631]: I1128 13:35:13.778714 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/7fc5dcd5-fde5-4a72-8d4a-cda51785e77f-var-log-ovn\") pod \"ovn-controller-7jgdf\" (UID: \"7fc5dcd5-fde5-4a72-8d4a-cda51785e77f\") " pod="openstack/ovn-controller-7jgdf" Nov 28 13:35:13 crc kubenswrapper[4631]: I1128 13:35:13.780050 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/7fc5dcd5-fde5-4a72-8d4a-cda51785e77f-var-run\") pod \"ovn-controller-7jgdf\" (UID: \"7fc5dcd5-fde5-4a72-8d4a-cda51785e77f\") " pod="openstack/ovn-controller-7jgdf" Nov 28 13:35:13 crc kubenswrapper[4631]: I1128 13:35:13.781098 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/7fc5dcd5-fde5-4a72-8d4a-cda51785e77f-var-run-ovn\") pod \"ovn-controller-7jgdf\" (UID: \"7fc5dcd5-fde5-4a72-8d4a-cda51785e77f\") " pod="openstack/ovn-controller-7jgdf" Nov 28 13:35:13 crc kubenswrapper[4631]: I1128 13:35:13.791004 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7fc5dcd5-fde5-4a72-8d4a-cda51785e77f-combined-ca-bundle\") pod \"ovn-controller-7jgdf\" (UID: \"7fc5dcd5-fde5-4a72-8d4a-cda51785e77f\") " pod="openstack/ovn-controller-7jgdf" Nov 28 13:35:13 crc kubenswrapper[4631]: I1128 13:35:13.791414 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/7fc5dcd5-fde5-4a72-8d4a-cda51785e77f-ovn-controller-tls-certs\") pod \"ovn-controller-7jgdf\" (UID: \"7fc5dcd5-fde5-4a72-8d4a-cda51785e77f\") " pod="openstack/ovn-controller-7jgdf" Nov 28 13:35:13 crc kubenswrapper[4631]: I1128 13:35:13.809258 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7fc5dcd5-fde5-4a72-8d4a-cda51785e77f-scripts\") pod \"ovn-controller-7jgdf\" (UID: \"7fc5dcd5-fde5-4a72-8d4a-cda51785e77f\") " pod="openstack/ovn-controller-7jgdf" Nov 28 13:35:13 crc kubenswrapper[4631]: I1128 13:35:13.898330 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g4qql\" (UniqueName: \"kubernetes.io/projected/7fc0ee4e-435d-4297-94eb-9e8c6fa90004-kube-api-access-g4qql\") pod \"ovn-controller-ovs-lskqw\" (UID: \"7fc0ee4e-435d-4297-94eb-9e8c6fa90004\") " pod="openstack/ovn-controller-ovs-lskqw" Nov 28 13:35:13 crc kubenswrapper[4631]: I1128 13:35:13.898489 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/7fc0ee4e-435d-4297-94eb-9e8c6fa90004-var-log\") pod \"ovn-controller-ovs-lskqw\" (UID: \"7fc0ee4e-435d-4297-94eb-9e8c6fa90004\") " pod="openstack/ovn-controller-ovs-lskqw" Nov 28 13:35:13 crc kubenswrapper[4631]: I1128 13:35:13.898690 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/7fc0ee4e-435d-4297-94eb-9e8c6fa90004-etc-ovs\") pod \"ovn-controller-ovs-lskqw\" (UID: \"7fc0ee4e-435d-4297-94eb-9e8c6fa90004\") " pod="openstack/ovn-controller-ovs-lskqw" Nov 28 13:35:13 crc kubenswrapper[4631]: I1128 13:35:13.898710 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/7fc0ee4e-435d-4297-94eb-9e8c6fa90004-var-run\") pod \"ovn-controller-ovs-lskqw\" (UID: \"7fc0ee4e-435d-4297-94eb-9e8c6fa90004\") " pod="openstack/ovn-controller-ovs-lskqw" Nov 28 13:35:13 crc kubenswrapper[4631]: I1128 13:35:13.898765 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/7fc0ee4e-435d-4297-94eb-9e8c6fa90004-var-lib\") pod \"ovn-controller-ovs-lskqw\" (UID: \"7fc0ee4e-435d-4297-94eb-9e8c6fa90004\") " pod="openstack/ovn-controller-ovs-lskqw" Nov 28 13:35:13 crc kubenswrapper[4631]: I1128 13:35:13.898906 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7fc0ee4e-435d-4297-94eb-9e8c6fa90004-scripts\") pod \"ovn-controller-ovs-lskqw\" (UID: \"7fc0ee4e-435d-4297-94eb-9e8c6fa90004\") " pod="openstack/ovn-controller-ovs-lskqw" Nov 28 13:35:13 crc kubenswrapper[4631]: I1128 13:35:13.901513 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7fc0ee4e-435d-4297-94eb-9e8c6fa90004-scripts\") pod \"ovn-controller-ovs-lskqw\" (UID: \"7fc0ee4e-435d-4297-94eb-9e8c6fa90004\") " pod="openstack/ovn-controller-ovs-lskqw" Nov 28 13:35:13 crc kubenswrapper[4631]: I1128 13:35:13.901643 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/7fc0ee4e-435d-4297-94eb-9e8c6fa90004-var-log\") pod \"ovn-controller-ovs-lskqw\" (UID: \"7fc0ee4e-435d-4297-94eb-9e8c6fa90004\") " pod="openstack/ovn-controller-ovs-lskqw" Nov 28 13:35:13 crc kubenswrapper[4631]: I1128 13:35:13.901749 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/7fc0ee4e-435d-4297-94eb-9e8c6fa90004-etc-ovs\") pod \"ovn-controller-ovs-lskqw\" (UID: \"7fc0ee4e-435d-4297-94eb-9e8c6fa90004\") " pod="openstack/ovn-controller-ovs-lskqw" Nov 28 13:35:13 crc kubenswrapper[4631]: I1128 13:35:13.901795 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/7fc0ee4e-435d-4297-94eb-9e8c6fa90004-var-run\") pod \"ovn-controller-ovs-lskqw\" (UID: \"7fc0ee4e-435d-4297-94eb-9e8c6fa90004\") " pod="openstack/ovn-controller-ovs-lskqw" Nov 28 13:35:13 crc kubenswrapper[4631]: I1128 13:35:13.901896 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/7fc0ee4e-435d-4297-94eb-9e8c6fa90004-var-lib\") pod \"ovn-controller-ovs-lskqw\" (UID: \"7fc0ee4e-435d-4297-94eb-9e8c6fa90004\") " pod="openstack/ovn-controller-ovs-lskqw" Nov 28 13:35:13 crc kubenswrapper[4631]: I1128 13:35:13.928730 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-56tdl\" (UniqueName: \"kubernetes.io/projected/7fc5dcd5-fde5-4a72-8d4a-cda51785e77f-kube-api-access-56tdl\") pod \"ovn-controller-7jgdf\" (UID: \"7fc5dcd5-fde5-4a72-8d4a-cda51785e77f\") " pod="openstack/ovn-controller-7jgdf" Nov 28 13:35:13 crc kubenswrapper[4631]: I1128 13:35:13.948041 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-7jgdf" Nov 28 13:35:13 crc kubenswrapper[4631]: I1128 13:35:13.964171 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g4qql\" (UniqueName: \"kubernetes.io/projected/7fc0ee4e-435d-4297-94eb-9e8c6fa90004-kube-api-access-g4qql\") pod \"ovn-controller-ovs-lskqw\" (UID: \"7fc0ee4e-435d-4297-94eb-9e8c6fa90004\") " pod="openstack/ovn-controller-ovs-lskqw" Nov 28 13:35:14 crc kubenswrapper[4631]: I1128 13:35:14.088547 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Nov 28 13:35:14 crc kubenswrapper[4631]: W1128 13:35:14.117508 4631 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podca11061d_b0a1_464d_9e1d_5572fc439580.slice/crio-d3b1f22eacc4e5fcae386f967b5370f34c393daf390d30d42bc3b73a64b0f0bf WatchSource:0}: Error finding container d3b1f22eacc4e5fcae386f967b5370f34c393daf390d30d42bc3b73a64b0f0bf: Status 404 returned error can't find the container with id d3b1f22eacc4e5fcae386f967b5370f34c393daf390d30d42bc3b73a64b0f0bf Nov 28 13:35:14 crc kubenswrapper[4631]: I1128 13:35:14.150788 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-lskqw" Nov 28 13:35:14 crc kubenswrapper[4631]: I1128 13:35:14.199346 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"ca11061d-b0a1-464d-9e1d-5572fc439580","Type":"ContainerStarted","Data":"d3b1f22eacc4e5fcae386f967b5370f34c393daf390d30d42bc3b73a64b0f0bf"} Nov 28 13:35:14 crc kubenswrapper[4631]: I1128 13:35:14.311894 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Nov 28 13:35:14 crc kubenswrapper[4631]: I1128 13:35:14.314541 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Nov 28 13:35:14 crc kubenswrapper[4631]: I1128 13:35:14.318607 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Nov 28 13:35:14 crc kubenswrapper[4631]: I1128 13:35:14.318927 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Nov 28 13:35:14 crc kubenswrapper[4631]: I1128 13:35:14.319101 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-84cfn" Nov 28 13:35:14 crc kubenswrapper[4631]: I1128 13:35:14.319226 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Nov 28 13:35:14 crc kubenswrapper[4631]: I1128 13:35:14.335097 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Nov 28 13:35:14 crc kubenswrapper[4631]: I1128 13:35:14.419896 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/1d3ba54c-cb26-4e04-8310-34e6b4206119-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"1d3ba54c-cb26-4e04-8310-34e6b4206119\") " pod="openstack/ovsdbserver-sb-0" Nov 28 13:35:14 crc kubenswrapper[4631]: I1128 13:35:14.419931 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j99sl\" (UniqueName: \"kubernetes.io/projected/1d3ba54c-cb26-4e04-8310-34e6b4206119-kube-api-access-j99sl\") pod \"ovsdbserver-sb-0\" (UID: \"1d3ba54c-cb26-4e04-8310-34e6b4206119\") " pod="openstack/ovsdbserver-sb-0" Nov 28 13:35:14 crc kubenswrapper[4631]: I1128 13:35:14.419981 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"ovsdbserver-sb-0\" (UID: \"1d3ba54c-cb26-4e04-8310-34e6b4206119\") " pod="openstack/ovsdbserver-sb-0" Nov 28 13:35:14 crc kubenswrapper[4631]: I1128 13:35:14.420018 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1d3ba54c-cb26-4e04-8310-34e6b4206119-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"1d3ba54c-cb26-4e04-8310-34e6b4206119\") " pod="openstack/ovsdbserver-sb-0" Nov 28 13:35:14 crc kubenswrapper[4631]: I1128 13:35:14.420044 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/1d3ba54c-cb26-4e04-8310-34e6b4206119-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"1d3ba54c-cb26-4e04-8310-34e6b4206119\") " pod="openstack/ovsdbserver-sb-0" Nov 28 13:35:14 crc kubenswrapper[4631]: I1128 13:35:14.420074 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1d3ba54c-cb26-4e04-8310-34e6b4206119-config\") pod \"ovsdbserver-sb-0\" (UID: \"1d3ba54c-cb26-4e04-8310-34e6b4206119\") " pod="openstack/ovsdbserver-sb-0" Nov 28 13:35:14 crc kubenswrapper[4631]: I1128 13:35:14.420094 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/1d3ba54c-cb26-4e04-8310-34e6b4206119-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"1d3ba54c-cb26-4e04-8310-34e6b4206119\") " pod="openstack/ovsdbserver-sb-0" Nov 28 13:35:14 crc kubenswrapper[4631]: I1128 13:35:14.420120 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d3ba54c-cb26-4e04-8310-34e6b4206119-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"1d3ba54c-cb26-4e04-8310-34e6b4206119\") " pod="openstack/ovsdbserver-sb-0" Nov 28 13:35:14 crc kubenswrapper[4631]: I1128 13:35:14.525844 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/1d3ba54c-cb26-4e04-8310-34e6b4206119-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"1d3ba54c-cb26-4e04-8310-34e6b4206119\") " pod="openstack/ovsdbserver-sb-0" Nov 28 13:35:14 crc kubenswrapper[4631]: I1128 13:35:14.526227 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j99sl\" (UniqueName: \"kubernetes.io/projected/1d3ba54c-cb26-4e04-8310-34e6b4206119-kube-api-access-j99sl\") pod \"ovsdbserver-sb-0\" (UID: \"1d3ba54c-cb26-4e04-8310-34e6b4206119\") " pod="openstack/ovsdbserver-sb-0" Nov 28 13:35:14 crc kubenswrapper[4631]: I1128 13:35:14.526272 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"ovsdbserver-sb-0\" (UID: \"1d3ba54c-cb26-4e04-8310-34e6b4206119\") " pod="openstack/ovsdbserver-sb-0" Nov 28 13:35:14 crc kubenswrapper[4631]: I1128 13:35:14.526318 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1d3ba54c-cb26-4e04-8310-34e6b4206119-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"1d3ba54c-cb26-4e04-8310-34e6b4206119\") " pod="openstack/ovsdbserver-sb-0" Nov 28 13:35:14 crc kubenswrapper[4631]: I1128 13:35:14.526351 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/1d3ba54c-cb26-4e04-8310-34e6b4206119-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"1d3ba54c-cb26-4e04-8310-34e6b4206119\") " pod="openstack/ovsdbserver-sb-0" Nov 28 13:35:14 crc kubenswrapper[4631]: I1128 13:35:14.526376 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1d3ba54c-cb26-4e04-8310-34e6b4206119-config\") pod \"ovsdbserver-sb-0\" (UID: \"1d3ba54c-cb26-4e04-8310-34e6b4206119\") " pod="openstack/ovsdbserver-sb-0" Nov 28 13:35:14 crc kubenswrapper[4631]: I1128 13:35:14.526416 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/1d3ba54c-cb26-4e04-8310-34e6b4206119-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"1d3ba54c-cb26-4e04-8310-34e6b4206119\") " pod="openstack/ovsdbserver-sb-0" Nov 28 13:35:14 crc kubenswrapper[4631]: I1128 13:35:14.526446 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d3ba54c-cb26-4e04-8310-34e6b4206119-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"1d3ba54c-cb26-4e04-8310-34e6b4206119\") " pod="openstack/ovsdbserver-sb-0" Nov 28 13:35:14 crc kubenswrapper[4631]: I1128 13:35:14.526970 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/1d3ba54c-cb26-4e04-8310-34e6b4206119-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"1d3ba54c-cb26-4e04-8310-34e6b4206119\") " pod="openstack/ovsdbserver-sb-0" Nov 28 13:35:14 crc kubenswrapper[4631]: I1128 13:35:14.528426 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1d3ba54c-cb26-4e04-8310-34e6b4206119-config\") pod \"ovsdbserver-sb-0\" (UID: \"1d3ba54c-cb26-4e04-8310-34e6b4206119\") " pod="openstack/ovsdbserver-sb-0" Nov 28 13:35:14 crc kubenswrapper[4631]: I1128 13:35:14.529125 4631 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"ovsdbserver-sb-0\" (UID: \"1d3ba54c-cb26-4e04-8310-34e6b4206119\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/ovsdbserver-sb-0" Nov 28 13:35:14 crc kubenswrapper[4631]: I1128 13:35:14.531971 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1d3ba54c-cb26-4e04-8310-34e6b4206119-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"1d3ba54c-cb26-4e04-8310-34e6b4206119\") " pod="openstack/ovsdbserver-sb-0" Nov 28 13:35:14 crc kubenswrapper[4631]: I1128 13:35:14.543944 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/1d3ba54c-cb26-4e04-8310-34e6b4206119-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"1d3ba54c-cb26-4e04-8310-34e6b4206119\") " pod="openstack/ovsdbserver-sb-0" Nov 28 13:35:14 crc kubenswrapper[4631]: I1128 13:35:14.565421 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/1d3ba54c-cb26-4e04-8310-34e6b4206119-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"1d3ba54c-cb26-4e04-8310-34e6b4206119\") " pod="openstack/ovsdbserver-sb-0" Nov 28 13:35:14 crc kubenswrapper[4631]: I1128 13:35:14.579848 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d3ba54c-cb26-4e04-8310-34e6b4206119-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"1d3ba54c-cb26-4e04-8310-34e6b4206119\") " pod="openstack/ovsdbserver-sb-0" Nov 28 13:35:14 crc kubenswrapper[4631]: I1128 13:35:14.580040 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j99sl\" (UniqueName: \"kubernetes.io/projected/1d3ba54c-cb26-4e04-8310-34e6b4206119-kube-api-access-j99sl\") pod \"ovsdbserver-sb-0\" (UID: \"1d3ba54c-cb26-4e04-8310-34e6b4206119\") " pod="openstack/ovsdbserver-sb-0" Nov 28 13:35:14 crc kubenswrapper[4631]: I1128 13:35:14.597253 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"ovsdbserver-sb-0\" (UID: \"1d3ba54c-cb26-4e04-8310-34e6b4206119\") " pod="openstack/ovsdbserver-sb-0" Nov 28 13:35:14 crc kubenswrapper[4631]: I1128 13:35:14.638892 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Nov 28 13:35:14 crc kubenswrapper[4631]: I1128 13:35:14.872032 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-7jgdf"] Nov 28 13:35:14 crc kubenswrapper[4631]: I1128 13:35:14.881426 4631 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-5kbsx" Nov 28 13:35:15 crc kubenswrapper[4631]: I1128 13:35:15.250530 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-5kbsx" Nov 28 13:35:15 crc kubenswrapper[4631]: I1128 13:35:15.455796 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-5kbsx"] Nov 28 13:35:16 crc kubenswrapper[4631]: I1128 13:35:16.271577 4631 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-5kbsx" podUID="03ba4ed9-11ea-4c20-baaf-8975710c6708" containerName="registry-server" containerID="cri-o://a83d391e2674fdc7da18b14130e6f1287d3f94f515d8a9241a80c096555d9ace" gracePeriod=2 Nov 28 13:35:16 crc kubenswrapper[4631]: W1128 13:35:16.571331 4631 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7fc5dcd5_fde5_4a72_8d4a_cda51785e77f.slice/crio-4f67eb54f5a065630c3584c35a19c7b873ae5d53deb73c18c426d7854941364d WatchSource:0}: Error finding container 4f67eb54f5a065630c3584c35a19c7b873ae5d53deb73c18c426d7854941364d: Status 404 returned error can't find the container with id 4f67eb54f5a065630c3584c35a19c7b873ae5d53deb73c18c426d7854941364d Nov 28 13:35:17 crc kubenswrapper[4631]: I1128 13:35:17.295984 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-7jgdf" event={"ID":"7fc5dcd5-fde5-4a72-8d4a-cda51785e77f","Type":"ContainerStarted","Data":"4f67eb54f5a065630c3584c35a19c7b873ae5d53deb73c18c426d7854941364d"} Nov 28 13:35:17 crc kubenswrapper[4631]: I1128 13:35:17.310959 4631 generic.go:334] "Generic (PLEG): container finished" podID="03ba4ed9-11ea-4c20-baaf-8975710c6708" containerID="a83d391e2674fdc7da18b14130e6f1287d3f94f515d8a9241a80c096555d9ace" exitCode=0 Nov 28 13:35:17 crc kubenswrapper[4631]: I1128 13:35:17.311010 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5kbsx" event={"ID":"03ba4ed9-11ea-4c20-baaf-8975710c6708","Type":"ContainerDied","Data":"a83d391e2674fdc7da18b14130e6f1287d3f94f515d8a9241a80c096555d9ace"} Nov 28 13:35:17 crc kubenswrapper[4631]: I1128 13:35:17.975559 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5kbsx" Nov 28 13:35:18 crc kubenswrapper[4631]: I1128 13:35:18.142588 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/03ba4ed9-11ea-4c20-baaf-8975710c6708-catalog-content\") pod \"03ba4ed9-11ea-4c20-baaf-8975710c6708\" (UID: \"03ba4ed9-11ea-4c20-baaf-8975710c6708\") " Nov 28 13:35:18 crc kubenswrapper[4631]: I1128 13:35:18.143450 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/03ba4ed9-11ea-4c20-baaf-8975710c6708-utilities\") pod \"03ba4ed9-11ea-4c20-baaf-8975710c6708\" (UID: \"03ba4ed9-11ea-4c20-baaf-8975710c6708\") " Nov 28 13:35:18 crc kubenswrapper[4631]: I1128 13:35:18.143516 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5lgkc\" (UniqueName: \"kubernetes.io/projected/03ba4ed9-11ea-4c20-baaf-8975710c6708-kube-api-access-5lgkc\") pod \"03ba4ed9-11ea-4c20-baaf-8975710c6708\" (UID: \"03ba4ed9-11ea-4c20-baaf-8975710c6708\") " Nov 28 13:35:18 crc kubenswrapper[4631]: I1128 13:35:18.145493 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/03ba4ed9-11ea-4c20-baaf-8975710c6708-utilities" (OuterVolumeSpecName: "utilities") pod "03ba4ed9-11ea-4c20-baaf-8975710c6708" (UID: "03ba4ed9-11ea-4c20-baaf-8975710c6708"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 13:35:18 crc kubenswrapper[4631]: I1128 13:35:18.164210 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/03ba4ed9-11ea-4c20-baaf-8975710c6708-kube-api-access-5lgkc" (OuterVolumeSpecName: "kube-api-access-5lgkc") pod "03ba4ed9-11ea-4c20-baaf-8975710c6708" (UID: "03ba4ed9-11ea-4c20-baaf-8975710c6708"). InnerVolumeSpecName "kube-api-access-5lgkc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:35:18 crc kubenswrapper[4631]: I1128 13:35:18.219987 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/03ba4ed9-11ea-4c20-baaf-8975710c6708-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "03ba4ed9-11ea-4c20-baaf-8975710c6708" (UID: "03ba4ed9-11ea-4c20-baaf-8975710c6708"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 13:35:18 crc kubenswrapper[4631]: I1128 13:35:18.246110 4631 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/03ba4ed9-11ea-4c20-baaf-8975710c6708-utilities\") on node \"crc\" DevicePath \"\"" Nov 28 13:35:18 crc kubenswrapper[4631]: I1128 13:35:18.246148 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5lgkc\" (UniqueName: \"kubernetes.io/projected/03ba4ed9-11ea-4c20-baaf-8975710c6708-kube-api-access-5lgkc\") on node \"crc\" DevicePath \"\"" Nov 28 13:35:18 crc kubenswrapper[4631]: I1128 13:35:18.246174 4631 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/03ba4ed9-11ea-4c20-baaf-8975710c6708-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 28 13:35:18 crc kubenswrapper[4631]: I1128 13:35:18.348160 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5kbsx" event={"ID":"03ba4ed9-11ea-4c20-baaf-8975710c6708","Type":"ContainerDied","Data":"1acd129545f28ef171bd341f19aa9260adee0a846dd7b798e4b6c5b9c8988ac1"} Nov 28 13:35:18 crc kubenswrapper[4631]: I1128 13:35:18.348181 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5kbsx" Nov 28 13:35:18 crc kubenswrapper[4631]: I1128 13:35:18.348217 4631 scope.go:117] "RemoveContainer" containerID="a83d391e2674fdc7da18b14130e6f1287d3f94f515d8a9241a80c096555d9ace" Nov 28 13:35:18 crc kubenswrapper[4631]: I1128 13:35:18.405474 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-5kbsx"] Nov 28 13:35:18 crc kubenswrapper[4631]: I1128 13:35:18.418651 4631 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-5kbsx"] Nov 28 13:35:18 crc kubenswrapper[4631]: I1128 13:35:18.632464 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Nov 28 13:35:19 crc kubenswrapper[4631]: I1128 13:35:19.411307 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"1d3ba54c-cb26-4e04-8310-34e6b4206119","Type":"ContainerStarted","Data":"3580cc22cd79c0794929bda5858fed7ff5d37d326e5bf56ffde0ba865d71dd2d"} Nov 28 13:35:19 crc kubenswrapper[4631]: I1128 13:35:19.436384 4631 scope.go:117] "RemoveContainer" containerID="8a5833bbb45a201fe6bdee78e6f456aa459ba86952faeeceda5f1bdd2fce2a52" Nov 28 13:35:19 crc kubenswrapper[4631]: I1128 13:35:19.534417 4631 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="03ba4ed9-11ea-4c20-baaf-8975710c6708" path="/var/lib/kubelet/pods/03ba4ed9-11ea-4c20-baaf-8975710c6708/volumes" Nov 28 13:35:20 crc kubenswrapper[4631]: I1128 13:35:20.158357 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-lskqw"] Nov 28 13:35:24 crc kubenswrapper[4631]: I1128 13:35:24.281388 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-p7l62"] Nov 28 13:35:24 crc kubenswrapper[4631]: E1128 13:35:24.282480 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="03ba4ed9-11ea-4c20-baaf-8975710c6708" containerName="registry-server" Nov 28 13:35:24 crc kubenswrapper[4631]: I1128 13:35:24.282494 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="03ba4ed9-11ea-4c20-baaf-8975710c6708" containerName="registry-server" Nov 28 13:35:24 crc kubenswrapper[4631]: E1128 13:35:24.282518 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="03ba4ed9-11ea-4c20-baaf-8975710c6708" containerName="extract-content" Nov 28 13:35:24 crc kubenswrapper[4631]: I1128 13:35:24.282525 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="03ba4ed9-11ea-4c20-baaf-8975710c6708" containerName="extract-content" Nov 28 13:35:24 crc kubenswrapper[4631]: E1128 13:35:24.282536 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="03ba4ed9-11ea-4c20-baaf-8975710c6708" containerName="extract-utilities" Nov 28 13:35:24 crc kubenswrapper[4631]: I1128 13:35:24.282542 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="03ba4ed9-11ea-4c20-baaf-8975710c6708" containerName="extract-utilities" Nov 28 13:35:24 crc kubenswrapper[4631]: I1128 13:35:24.282721 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="03ba4ed9-11ea-4c20-baaf-8975710c6708" containerName="registry-server" Nov 28 13:35:24 crc kubenswrapper[4631]: I1128 13:35:24.283410 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-p7l62" Nov 28 13:35:24 crc kubenswrapper[4631]: I1128 13:35:24.286295 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Nov 28 13:35:24 crc kubenswrapper[4631]: I1128 13:35:24.297438 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-p7l62"] Nov 28 13:35:24 crc kubenswrapper[4631]: I1128 13:35:24.371519 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pk88m\" (UniqueName: \"kubernetes.io/projected/6dfe4f1a-897a-4ece-a345-4a4f86901079-kube-api-access-pk88m\") pod \"ovn-controller-metrics-p7l62\" (UID: \"6dfe4f1a-897a-4ece-a345-4a4f86901079\") " pod="openstack/ovn-controller-metrics-p7l62" Nov 28 13:35:24 crc kubenswrapper[4631]: I1128 13:35:24.371584 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/6dfe4f1a-897a-4ece-a345-4a4f86901079-ovs-rundir\") pod \"ovn-controller-metrics-p7l62\" (UID: \"6dfe4f1a-897a-4ece-a345-4a4f86901079\") " pod="openstack/ovn-controller-metrics-p7l62" Nov 28 13:35:24 crc kubenswrapper[4631]: I1128 13:35:24.371704 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/6dfe4f1a-897a-4ece-a345-4a4f86901079-ovn-rundir\") pod \"ovn-controller-metrics-p7l62\" (UID: \"6dfe4f1a-897a-4ece-a345-4a4f86901079\") " pod="openstack/ovn-controller-metrics-p7l62" Nov 28 13:35:24 crc kubenswrapper[4631]: I1128 13:35:24.371752 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/6dfe4f1a-897a-4ece-a345-4a4f86901079-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-p7l62\" (UID: \"6dfe4f1a-897a-4ece-a345-4a4f86901079\") " pod="openstack/ovn-controller-metrics-p7l62" Nov 28 13:35:24 crc kubenswrapper[4631]: I1128 13:35:24.371819 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6dfe4f1a-897a-4ece-a345-4a4f86901079-config\") pod \"ovn-controller-metrics-p7l62\" (UID: \"6dfe4f1a-897a-4ece-a345-4a4f86901079\") " pod="openstack/ovn-controller-metrics-p7l62" Nov 28 13:35:24 crc kubenswrapper[4631]: I1128 13:35:24.371840 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6dfe4f1a-897a-4ece-a345-4a4f86901079-combined-ca-bundle\") pod \"ovn-controller-metrics-p7l62\" (UID: \"6dfe4f1a-897a-4ece-a345-4a4f86901079\") " pod="openstack/ovn-controller-metrics-p7l62" Nov 28 13:35:24 crc kubenswrapper[4631]: I1128 13:35:24.475340 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/6dfe4f1a-897a-4ece-a345-4a4f86901079-ovs-rundir\") pod \"ovn-controller-metrics-p7l62\" (UID: \"6dfe4f1a-897a-4ece-a345-4a4f86901079\") " pod="openstack/ovn-controller-metrics-p7l62" Nov 28 13:35:24 crc kubenswrapper[4631]: I1128 13:35:24.475421 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/6dfe4f1a-897a-4ece-a345-4a4f86901079-ovn-rundir\") pod \"ovn-controller-metrics-p7l62\" (UID: \"6dfe4f1a-897a-4ece-a345-4a4f86901079\") " pod="openstack/ovn-controller-metrics-p7l62" Nov 28 13:35:24 crc kubenswrapper[4631]: I1128 13:35:24.475462 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/6dfe4f1a-897a-4ece-a345-4a4f86901079-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-p7l62\" (UID: \"6dfe4f1a-897a-4ece-a345-4a4f86901079\") " pod="openstack/ovn-controller-metrics-p7l62" Nov 28 13:35:24 crc kubenswrapper[4631]: I1128 13:35:24.475523 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6dfe4f1a-897a-4ece-a345-4a4f86901079-config\") pod \"ovn-controller-metrics-p7l62\" (UID: \"6dfe4f1a-897a-4ece-a345-4a4f86901079\") " pod="openstack/ovn-controller-metrics-p7l62" Nov 28 13:35:24 crc kubenswrapper[4631]: I1128 13:35:24.475540 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6dfe4f1a-897a-4ece-a345-4a4f86901079-combined-ca-bundle\") pod \"ovn-controller-metrics-p7l62\" (UID: \"6dfe4f1a-897a-4ece-a345-4a4f86901079\") " pod="openstack/ovn-controller-metrics-p7l62" Nov 28 13:35:24 crc kubenswrapper[4631]: I1128 13:35:24.475573 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pk88m\" (UniqueName: \"kubernetes.io/projected/6dfe4f1a-897a-4ece-a345-4a4f86901079-kube-api-access-pk88m\") pod \"ovn-controller-metrics-p7l62\" (UID: \"6dfe4f1a-897a-4ece-a345-4a4f86901079\") " pod="openstack/ovn-controller-metrics-p7l62" Nov 28 13:35:24 crc kubenswrapper[4631]: I1128 13:35:24.476168 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/6dfe4f1a-897a-4ece-a345-4a4f86901079-ovs-rundir\") pod \"ovn-controller-metrics-p7l62\" (UID: \"6dfe4f1a-897a-4ece-a345-4a4f86901079\") " pod="openstack/ovn-controller-metrics-p7l62" Nov 28 13:35:24 crc kubenswrapper[4631]: I1128 13:35:24.476221 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/6dfe4f1a-897a-4ece-a345-4a4f86901079-ovn-rundir\") pod \"ovn-controller-metrics-p7l62\" (UID: \"6dfe4f1a-897a-4ece-a345-4a4f86901079\") " pod="openstack/ovn-controller-metrics-p7l62" Nov 28 13:35:24 crc kubenswrapper[4631]: I1128 13:35:24.477634 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6dfe4f1a-897a-4ece-a345-4a4f86901079-config\") pod \"ovn-controller-metrics-p7l62\" (UID: \"6dfe4f1a-897a-4ece-a345-4a4f86901079\") " pod="openstack/ovn-controller-metrics-p7l62" Nov 28 13:35:24 crc kubenswrapper[4631]: I1128 13:35:24.485449 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/6dfe4f1a-897a-4ece-a345-4a4f86901079-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-p7l62\" (UID: \"6dfe4f1a-897a-4ece-a345-4a4f86901079\") " pod="openstack/ovn-controller-metrics-p7l62" Nov 28 13:35:24 crc kubenswrapper[4631]: I1128 13:35:24.514165 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6dfe4f1a-897a-4ece-a345-4a4f86901079-combined-ca-bundle\") pod \"ovn-controller-metrics-p7l62\" (UID: \"6dfe4f1a-897a-4ece-a345-4a4f86901079\") " pod="openstack/ovn-controller-metrics-p7l62" Nov 28 13:35:24 crc kubenswrapper[4631]: I1128 13:35:24.519574 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pk88m\" (UniqueName: \"kubernetes.io/projected/6dfe4f1a-897a-4ece-a345-4a4f86901079-kube-api-access-pk88m\") pod \"ovn-controller-metrics-p7l62\" (UID: \"6dfe4f1a-897a-4ece-a345-4a4f86901079\") " pod="openstack/ovn-controller-metrics-p7l62" Nov 28 13:35:24 crc kubenswrapper[4631]: I1128 13:35:24.551770 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-jl8jp"] Nov 28 13:35:24 crc kubenswrapper[4631]: I1128 13:35:24.614000 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-fhmpd"] Nov 28 13:35:24 crc kubenswrapper[4631]: I1128 13:35:24.620692 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fd796d7df-fhmpd" Nov 28 13:35:24 crc kubenswrapper[4631]: I1128 13:35:24.623989 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-fhmpd"] Nov 28 13:35:24 crc kubenswrapper[4631]: I1128 13:35:24.624784 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-p7l62" Nov 28 13:35:24 crc kubenswrapper[4631]: I1128 13:35:24.628268 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Nov 28 13:35:24 crc kubenswrapper[4631]: I1128 13:35:24.679762 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mvcxt\" (UniqueName: \"kubernetes.io/projected/9594c34e-0cdb-4e81-9e4a-1f542741e05b-kube-api-access-mvcxt\") pod \"dnsmasq-dns-7fd796d7df-fhmpd\" (UID: \"9594c34e-0cdb-4e81-9e4a-1f542741e05b\") " pod="openstack/dnsmasq-dns-7fd796d7df-fhmpd" Nov 28 13:35:24 crc kubenswrapper[4631]: I1128 13:35:24.680207 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9594c34e-0cdb-4e81-9e4a-1f542741e05b-ovsdbserver-nb\") pod \"dnsmasq-dns-7fd796d7df-fhmpd\" (UID: \"9594c34e-0cdb-4e81-9e4a-1f542741e05b\") " pod="openstack/dnsmasq-dns-7fd796d7df-fhmpd" Nov 28 13:35:24 crc kubenswrapper[4631]: I1128 13:35:24.680278 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9594c34e-0cdb-4e81-9e4a-1f542741e05b-dns-svc\") pod \"dnsmasq-dns-7fd796d7df-fhmpd\" (UID: \"9594c34e-0cdb-4e81-9e4a-1f542741e05b\") " pod="openstack/dnsmasq-dns-7fd796d7df-fhmpd" Nov 28 13:35:24 crc kubenswrapper[4631]: I1128 13:35:24.680326 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9594c34e-0cdb-4e81-9e4a-1f542741e05b-config\") pod \"dnsmasq-dns-7fd796d7df-fhmpd\" (UID: \"9594c34e-0cdb-4e81-9e4a-1f542741e05b\") " pod="openstack/dnsmasq-dns-7fd796d7df-fhmpd" Nov 28 13:35:24 crc kubenswrapper[4631]: I1128 13:35:24.767078 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-clh2s"] Nov 28 13:35:24 crc kubenswrapper[4631]: I1128 13:35:24.782554 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mvcxt\" (UniqueName: \"kubernetes.io/projected/9594c34e-0cdb-4e81-9e4a-1f542741e05b-kube-api-access-mvcxt\") pod \"dnsmasq-dns-7fd796d7df-fhmpd\" (UID: \"9594c34e-0cdb-4e81-9e4a-1f542741e05b\") " pod="openstack/dnsmasq-dns-7fd796d7df-fhmpd" Nov 28 13:35:24 crc kubenswrapper[4631]: I1128 13:35:24.782742 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9594c34e-0cdb-4e81-9e4a-1f542741e05b-ovsdbserver-nb\") pod \"dnsmasq-dns-7fd796d7df-fhmpd\" (UID: \"9594c34e-0cdb-4e81-9e4a-1f542741e05b\") " pod="openstack/dnsmasq-dns-7fd796d7df-fhmpd" Nov 28 13:35:24 crc kubenswrapper[4631]: I1128 13:35:24.782812 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9594c34e-0cdb-4e81-9e4a-1f542741e05b-dns-svc\") pod \"dnsmasq-dns-7fd796d7df-fhmpd\" (UID: \"9594c34e-0cdb-4e81-9e4a-1f542741e05b\") " pod="openstack/dnsmasq-dns-7fd796d7df-fhmpd" Nov 28 13:35:24 crc kubenswrapper[4631]: I1128 13:35:24.782871 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9594c34e-0cdb-4e81-9e4a-1f542741e05b-config\") pod \"dnsmasq-dns-7fd796d7df-fhmpd\" (UID: \"9594c34e-0cdb-4e81-9e4a-1f542741e05b\") " pod="openstack/dnsmasq-dns-7fd796d7df-fhmpd" Nov 28 13:35:24 crc kubenswrapper[4631]: I1128 13:35:24.783732 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9594c34e-0cdb-4e81-9e4a-1f542741e05b-ovsdbserver-nb\") pod \"dnsmasq-dns-7fd796d7df-fhmpd\" (UID: \"9594c34e-0cdb-4e81-9e4a-1f542741e05b\") " pod="openstack/dnsmasq-dns-7fd796d7df-fhmpd" Nov 28 13:35:24 crc kubenswrapper[4631]: I1128 13:35:24.783844 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9594c34e-0cdb-4e81-9e4a-1f542741e05b-dns-svc\") pod \"dnsmasq-dns-7fd796d7df-fhmpd\" (UID: \"9594c34e-0cdb-4e81-9e4a-1f542741e05b\") " pod="openstack/dnsmasq-dns-7fd796d7df-fhmpd" Nov 28 13:35:24 crc kubenswrapper[4631]: I1128 13:35:24.783854 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9594c34e-0cdb-4e81-9e4a-1f542741e05b-config\") pod \"dnsmasq-dns-7fd796d7df-fhmpd\" (UID: \"9594c34e-0cdb-4e81-9e4a-1f542741e05b\") " pod="openstack/dnsmasq-dns-7fd796d7df-fhmpd" Nov 28 13:35:24 crc kubenswrapper[4631]: I1128 13:35:24.799970 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-49s6f"] Nov 28 13:35:24 crc kubenswrapper[4631]: I1128 13:35:24.801613 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-49s6f" Nov 28 13:35:24 crc kubenswrapper[4631]: I1128 13:35:24.805479 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Nov 28 13:35:24 crc kubenswrapper[4631]: I1128 13:35:24.809357 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mvcxt\" (UniqueName: \"kubernetes.io/projected/9594c34e-0cdb-4e81-9e4a-1f542741e05b-kube-api-access-mvcxt\") pod \"dnsmasq-dns-7fd796d7df-fhmpd\" (UID: \"9594c34e-0cdb-4e81-9e4a-1f542741e05b\") " pod="openstack/dnsmasq-dns-7fd796d7df-fhmpd" Nov 28 13:35:24 crc kubenswrapper[4631]: I1128 13:35:24.827571 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-49s6f"] Nov 28 13:35:24 crc kubenswrapper[4631]: I1128 13:35:24.886324 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1959423d-7154-41e1-80db-d5840a9c1702-ovsdbserver-nb\") pod \"dnsmasq-dns-86db49b7ff-49s6f\" (UID: \"1959423d-7154-41e1-80db-d5840a9c1702\") " pod="openstack/dnsmasq-dns-86db49b7ff-49s6f" Nov 28 13:35:24 crc kubenswrapper[4631]: I1128 13:35:24.886471 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j95lf\" (UniqueName: \"kubernetes.io/projected/1959423d-7154-41e1-80db-d5840a9c1702-kube-api-access-j95lf\") pod \"dnsmasq-dns-86db49b7ff-49s6f\" (UID: \"1959423d-7154-41e1-80db-d5840a9c1702\") " pod="openstack/dnsmasq-dns-86db49b7ff-49s6f" Nov 28 13:35:24 crc kubenswrapper[4631]: I1128 13:35:24.886507 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1959423d-7154-41e1-80db-d5840a9c1702-ovsdbserver-sb\") pod \"dnsmasq-dns-86db49b7ff-49s6f\" (UID: \"1959423d-7154-41e1-80db-d5840a9c1702\") " pod="openstack/dnsmasq-dns-86db49b7ff-49s6f" Nov 28 13:35:24 crc kubenswrapper[4631]: I1128 13:35:24.886575 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1959423d-7154-41e1-80db-d5840a9c1702-config\") pod \"dnsmasq-dns-86db49b7ff-49s6f\" (UID: \"1959423d-7154-41e1-80db-d5840a9c1702\") " pod="openstack/dnsmasq-dns-86db49b7ff-49s6f" Nov 28 13:35:24 crc kubenswrapper[4631]: I1128 13:35:24.886625 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1959423d-7154-41e1-80db-d5840a9c1702-dns-svc\") pod \"dnsmasq-dns-86db49b7ff-49s6f\" (UID: \"1959423d-7154-41e1-80db-d5840a9c1702\") " pod="openstack/dnsmasq-dns-86db49b7ff-49s6f" Nov 28 13:35:24 crc kubenswrapper[4631]: I1128 13:35:24.965913 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fd796d7df-fhmpd" Nov 28 13:35:24 crc kubenswrapper[4631]: I1128 13:35:24.988306 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1959423d-7154-41e1-80db-d5840a9c1702-config\") pod \"dnsmasq-dns-86db49b7ff-49s6f\" (UID: \"1959423d-7154-41e1-80db-d5840a9c1702\") " pod="openstack/dnsmasq-dns-86db49b7ff-49s6f" Nov 28 13:35:24 crc kubenswrapper[4631]: I1128 13:35:24.988381 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1959423d-7154-41e1-80db-d5840a9c1702-dns-svc\") pod \"dnsmasq-dns-86db49b7ff-49s6f\" (UID: \"1959423d-7154-41e1-80db-d5840a9c1702\") " pod="openstack/dnsmasq-dns-86db49b7ff-49s6f" Nov 28 13:35:24 crc kubenswrapper[4631]: I1128 13:35:24.988427 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1959423d-7154-41e1-80db-d5840a9c1702-ovsdbserver-nb\") pod \"dnsmasq-dns-86db49b7ff-49s6f\" (UID: \"1959423d-7154-41e1-80db-d5840a9c1702\") " pod="openstack/dnsmasq-dns-86db49b7ff-49s6f" Nov 28 13:35:24 crc kubenswrapper[4631]: I1128 13:35:24.988479 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j95lf\" (UniqueName: \"kubernetes.io/projected/1959423d-7154-41e1-80db-d5840a9c1702-kube-api-access-j95lf\") pod \"dnsmasq-dns-86db49b7ff-49s6f\" (UID: \"1959423d-7154-41e1-80db-d5840a9c1702\") " pod="openstack/dnsmasq-dns-86db49b7ff-49s6f" Nov 28 13:35:24 crc kubenswrapper[4631]: I1128 13:35:24.988497 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1959423d-7154-41e1-80db-d5840a9c1702-ovsdbserver-sb\") pod \"dnsmasq-dns-86db49b7ff-49s6f\" (UID: \"1959423d-7154-41e1-80db-d5840a9c1702\") " pod="openstack/dnsmasq-dns-86db49b7ff-49s6f" Nov 28 13:35:24 crc kubenswrapper[4631]: I1128 13:35:24.989221 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1959423d-7154-41e1-80db-d5840a9c1702-config\") pod \"dnsmasq-dns-86db49b7ff-49s6f\" (UID: \"1959423d-7154-41e1-80db-d5840a9c1702\") " pod="openstack/dnsmasq-dns-86db49b7ff-49s6f" Nov 28 13:35:24 crc kubenswrapper[4631]: I1128 13:35:24.989444 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1959423d-7154-41e1-80db-d5840a9c1702-ovsdbserver-sb\") pod \"dnsmasq-dns-86db49b7ff-49s6f\" (UID: \"1959423d-7154-41e1-80db-d5840a9c1702\") " pod="openstack/dnsmasq-dns-86db49b7ff-49s6f" Nov 28 13:35:24 crc kubenswrapper[4631]: I1128 13:35:24.989534 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1959423d-7154-41e1-80db-d5840a9c1702-dns-svc\") pod \"dnsmasq-dns-86db49b7ff-49s6f\" (UID: \"1959423d-7154-41e1-80db-d5840a9c1702\") " pod="openstack/dnsmasq-dns-86db49b7ff-49s6f" Nov 28 13:35:24 crc kubenswrapper[4631]: I1128 13:35:24.989559 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1959423d-7154-41e1-80db-d5840a9c1702-ovsdbserver-nb\") pod \"dnsmasq-dns-86db49b7ff-49s6f\" (UID: \"1959423d-7154-41e1-80db-d5840a9c1702\") " pod="openstack/dnsmasq-dns-86db49b7ff-49s6f" Nov 28 13:35:25 crc kubenswrapper[4631]: I1128 13:35:25.006137 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j95lf\" (UniqueName: \"kubernetes.io/projected/1959423d-7154-41e1-80db-d5840a9c1702-kube-api-access-j95lf\") pod \"dnsmasq-dns-86db49b7ff-49s6f\" (UID: \"1959423d-7154-41e1-80db-d5840a9c1702\") " pod="openstack/dnsmasq-dns-86db49b7ff-49s6f" Nov 28 13:35:25 crc kubenswrapper[4631]: I1128 13:35:25.151375 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-49s6f" Nov 28 13:35:27 crc kubenswrapper[4631]: W1128 13:35:27.546572 4631 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7fc0ee4e_435d_4297_94eb_9e8c6fa90004.slice/crio-81e7353e670e7dd6c3f146f9edc57b0aecb9e642574c2ce904f8264fc01647f7 WatchSource:0}: Error finding container 81e7353e670e7dd6c3f146f9edc57b0aecb9e642574c2ce904f8264fc01647f7: Status 404 returned error can't find the container with id 81e7353e670e7dd6c3f146f9edc57b0aecb9e642574c2ce904f8264fc01647f7 Nov 28 13:35:27 crc kubenswrapper[4631]: I1128 13:35:27.621855 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-lskqw" event={"ID":"7fc0ee4e-435d-4297-94eb-9e8c6fa90004","Type":"ContainerStarted","Data":"81e7353e670e7dd6c3f146f9edc57b0aecb9e642574c2ce904f8264fc01647f7"} Nov 28 13:35:31 crc kubenswrapper[4631]: I1128 13:35:31.842390 4631 scope.go:117] "RemoveContainer" containerID="17ab23e3b640d1fff2f89f44b9ccfaeaa88a21edb104d9e450e9cba4860a98c1" Nov 28 13:35:32 crc kubenswrapper[4631]: E1128 13:35:32.279910 4631 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.k8s.io/kube-state-metrics/kube-state-metrics:v2.15.0" Nov 28 13:35:32 crc kubenswrapper[4631]: E1128 13:35:32.280012 4631 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.k8s.io/kube-state-metrics/kube-state-metrics:v2.15.0" Nov 28 13:35:32 crc kubenswrapper[4631]: E1128 13:35:32.280248 4631 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-state-metrics,Image:registry.k8s.io/kube-state-metrics/kube-state-metrics:v2.15.0,Command:[],Args:[--resources=pods --namespaces=openstack],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:http-metrics,HostPort:0,ContainerPort:8080,Protocol:TCP,HostIP:,},ContainerPort{Name:telemetry,HostPort:0,ContainerPort:8081,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-xj59v,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/livez,Port:{0 8080 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:*true,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod kube-state-metrics-0_openstack(83d77da3-fbb7-434f-abaf-5be9b599a9bf): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 28 13:35:32 crc kubenswrapper[4631]: E1128 13:35:32.281529 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-state-metrics\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openstack/kube-state-metrics-0" podUID="83d77da3-fbb7-434f-abaf-5be9b599a9bf" Nov 28 13:35:32 crc kubenswrapper[4631]: E1128 13:35:32.673671 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-state-metrics\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.k8s.io/kube-state-metrics/kube-state-metrics:v2.15.0\\\"\"" pod="openstack/kube-state-metrics-0" podUID="83d77da3-fbb7-434f-abaf-5be9b599a9bf" Nov 28 13:35:33 crc kubenswrapper[4631]: E1128 13:35:33.498089 4631 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified" Nov 28 13:35:33 crc kubenswrapper[4631]: E1128 13:35:33.498349 4631 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:setup-container,Image:quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified,Command:[sh -c cp /tmp/erlang-cookie-secret/.erlang.cookie /var/lib/rabbitmq/.erlang.cookie && chmod 600 /var/lib/rabbitmq/.erlang.cookie ; cp /tmp/rabbitmq-plugins/enabled_plugins /operator/enabled_plugins ; echo '[default]' > /var/lib/rabbitmq/.rabbitmqadmin.conf && sed -e 's/default_user/username/' -e 's/default_pass/password/' /tmp/default_user.conf >> /var/lib/rabbitmq/.rabbitmqadmin.conf && chmod 600 /var/lib/rabbitmq/.rabbitmqadmin.conf ; sleep 30],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Requests:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:plugins-conf,ReadOnly:false,MountPath:/tmp/rabbitmq-plugins/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-erlang-cookie,ReadOnly:false,MountPath:/var/lib/rabbitmq/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:erlang-cookie-secret,ReadOnly:false,MountPath:/tmp/erlang-cookie-secret/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-plugins,ReadOnly:false,MountPath:/operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:persistence,ReadOnly:false,MountPath:/var/lib/rabbitmq/mnesia/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-confd,ReadOnly:false,MountPath:/tmp/default_user.conf,SubPath:default_user.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-4vt5k,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cell1-server-0_openstack(ed4141f2-5eff-4252-8dc7-ad60a1580189): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 28 13:35:33 crc kubenswrapper[4631]: E1128 13:35:33.499423 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/rabbitmq-cell1-server-0" podUID="ed4141f2-5eff-4252-8dc7-ad60a1580189" Nov 28 13:35:33 crc kubenswrapper[4631]: E1128 13:35:33.678472 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified\\\"\"" pod="openstack/rabbitmq-cell1-server-0" podUID="ed4141f2-5eff-4252-8dc7-ad60a1580189" Nov 28 13:35:34 crc kubenswrapper[4631]: E1128 13:35:34.296789 4631 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-memcached:current-podified" Nov 28 13:35:34 crc kubenswrapper[4631]: E1128 13:35:34.297515 4631 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:memcached,Image:quay.io/podified-antelope-centos9/openstack-memcached:current-podified,Command:[/usr/bin/dumb-init -- /usr/local/bin/kolla_start],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:memcached,HostPort:0,ContainerPort:11211,Protocol:TCP,HostIP:,},ContainerPort{Name:memcached-tls,HostPort:0,ContainerPort:11212,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:POD_IPS,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIPs,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:CONFIG_HASH,Value:nf7hdch576h687h97h75h656h59chd9h586hc7h65fh58fh658h58dh56fh669h59bh5d6h5f4h555h74h54fh68h58dh59fh698h9bhf8h6dh7h5dcq,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/src,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kolla-config,ReadOnly:true,MountPath:/var/lib/kolla/config_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:memcached-tls-certs,ReadOnly:true,MountPath:/var/lib/config-data/tls/certs/memcached.crt,SubPath:tls.crt,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:memcached-tls-certs,ReadOnly:true,MountPath:/var/lib/config-data/tls/private/memcached.key,SubPath:tls.key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-8ppj8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 11211 },Host:,},GRPC:nil,},InitialDelaySeconds:3,TimeoutSeconds:5,PeriodSeconds:3,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 11211 },Host:,},GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42457,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42457,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod memcached-0_openstack(d35228a7-a0fe-496d-9d62-5f1d5a44c72f): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 28 13:35:34 crc kubenswrapper[4631]: E1128 13:35:34.299554 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"memcached\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/memcached-0" podUID="d35228a7-a0fe-496d-9d62-5f1d5a44c72f" Nov 28 13:35:34 crc kubenswrapper[4631]: E1128 13:35:34.317948 4631 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified" Nov 28 13:35:34 crc kubenswrapper[4631]: E1128 13:35:34.318225 4631 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:setup-container,Image:quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified,Command:[sh -c cp /tmp/erlang-cookie-secret/.erlang.cookie /var/lib/rabbitmq/.erlang.cookie && chmod 600 /var/lib/rabbitmq/.erlang.cookie ; cp /tmp/rabbitmq-plugins/enabled_plugins /operator/enabled_plugins ; echo '[default]' > /var/lib/rabbitmq/.rabbitmqadmin.conf && sed -e 's/default_user/username/' -e 's/default_pass/password/' /tmp/default_user.conf >> /var/lib/rabbitmq/.rabbitmqadmin.conf && chmod 600 /var/lib/rabbitmq/.rabbitmqadmin.conf ; sleep 30],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Requests:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:plugins-conf,ReadOnly:false,MountPath:/tmp/rabbitmq-plugins/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-erlang-cookie,ReadOnly:false,MountPath:/var/lib/rabbitmq/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:erlang-cookie-secret,ReadOnly:false,MountPath:/tmp/erlang-cookie-secret/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-plugins,ReadOnly:false,MountPath:/operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:persistence,ReadOnly:false,MountPath:/var/lib/rabbitmq/mnesia/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-confd,ReadOnly:false,MountPath:/tmp/default_user.conf,SubPath:default_user.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-4x8mv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-server-0_openstack(3a91e9fa-34b7-4688-9171-56a62043759d): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 28 13:35:34 crc kubenswrapper[4631]: E1128 13:35:34.319552 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/rabbitmq-server-0" podUID="3a91e9fa-34b7-4688-9171-56a62043759d" Nov 28 13:35:34 crc kubenswrapper[4631]: E1128 13:35:34.685148 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"memcached\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-memcached:current-podified\\\"\"" pod="openstack/memcached-0" podUID="d35228a7-a0fe-496d-9d62-5f1d5a44c72f" Nov 28 13:35:34 crc kubenswrapper[4631]: E1128 13:35:34.685469 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified\\\"\"" pod="openstack/rabbitmq-server-0" podUID="3a91e9fa-34b7-4688-9171-56a62043759d" Nov 28 13:35:50 crc kubenswrapper[4631]: E1128 13:35:50.694716 4631 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-ovn-base:current-podified" Nov 28 13:35:50 crc kubenswrapper[4631]: E1128 13:35:50.695649 4631 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:ovsdb-server-init,Image:quay.io/podified-antelope-centos9/openstack-ovn-base:current-podified,Command:[/usr/local/bin/container-scripts/init-ovsdb-server.sh],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ndh65bhc5h78h678h5d8h64h596h595h7fh7h686h85h55fhb5h66bh588h5c5h5bbh5fhf5hd5h659hcfh55dhc6h696h5bdh5c5hfch586h578q,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-ovs,ReadOnly:false,MountPath:/etc/openvswitch,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:var-run,ReadOnly:false,MountPath:/var/run/openvswitch,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:var-log,ReadOnly:false,MountPath:/var/log/openvswitch,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:var-lib,ReadOnly:false,MountPath:/var/lib/openvswitch,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-g4qql,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[NET_ADMIN SYS_ADMIN SYS_NICE],Drop:[],},Privileged:*true,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovn-controller-ovs-lskqw_openstack(7fc0ee4e-435d-4297-94eb-9e8c6fa90004): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 28 13:35:50 crc kubenswrapper[4631]: E1128 13:35:50.696905 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovsdb-server-init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/ovn-controller-ovs-lskqw" podUID="7fc0ee4e-435d-4297-94eb-9e8c6fa90004" Nov 28 13:35:50 crc kubenswrapper[4631]: E1128 13:35:50.823606 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovsdb-server-init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-ovn-base:current-podified\\\"\"" pod="openstack/ovn-controller-ovs-lskqw" podUID="7fc0ee4e-435d-4297-94eb-9e8c6fa90004" Nov 28 13:35:51 crc kubenswrapper[4631]: E1128 13:35:51.002843 4631 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified" Nov 28 13:35:51 crc kubenswrapper[4631]: E1128 13:35:51.003164 4631 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:ovn-controller,Image:quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified,Command:[ovn-controller --pidfile unix:/run/openvswitch/db.sock --certificate=/etc/pki/tls/certs/ovndb.crt --private-key=/etc/pki/tls/private/ovndb.key --ca-cert=/etc/pki/tls/certs/ovndbca.crt],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ndh65bhc5h78h678h5d8h64h596h595h7fh7h686h85h55fhb5h66bh588h5c5h5bbh5fhf5hd5h659hcfh55dhc6h696h5bdh5c5hfch586h578q,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:var-run,ReadOnly:false,MountPath:/var/run/openvswitch,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:var-run-ovn,ReadOnly:false,MountPath:/var/run/ovn,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:var-log-ovn,ReadOnly:false,MountPath:/var/log/ovn,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovn-controller-tls-certs,ReadOnly:true,MountPath:/etc/pki/tls/certs/ovndb.crt,SubPath:tls.crt,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovn-controller-tls-certs,ReadOnly:true,MountPath:/etc/pki/tls/private/ovndb.key,SubPath:tls.key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovn-controller-tls-certs,ReadOnly:true,MountPath:/etc/pki/tls/certs/ovndbca.crt,SubPath:ca.crt,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-56tdl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/local/bin/container-scripts/ovn_controller_liveness.sh],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:30,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/local/bin/container-scripts/ovn_controller_readiness.sh],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:30,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:&Lifecycle{PostStart:nil,PreStop:&LifecycleHandler{Exec:&ExecAction{Command:[/usr/share/ovn/scripts/ovn-ctl stop_controller],},HTTPGet:nil,TCPSocket:nil,Sleep:nil,},},TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[NET_ADMIN SYS_ADMIN SYS_NICE],Drop:[],},Privileged:*true,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovn-controller-7jgdf_openstack(7fc5dcd5-fde5-4a72-8d4a-cda51785e77f): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 28 13:35:51 crc kubenswrapper[4631]: E1128 13:35:51.006476 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovn-controller\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/ovn-controller-7jgdf" podUID="7fc5dcd5-fde5-4a72-8d4a-cda51785e77f" Nov 28 13:35:51 crc kubenswrapper[4631]: E1128 13:35:51.273212 4631 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-ovn-nb-db-server:current-podified" Nov 28 13:35:51 crc kubenswrapper[4631]: E1128 13:35:51.273611 4631 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:ovsdbserver-nb,Image:quay.io/podified-antelope-centos9/openstack-ovn-nb-db-server:current-podified,Command:[/usr/bin/dumb-init],Args:[/usr/local/bin/container-scripts/setup.sh],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n588hb4hbfh5ffh584h5c9h96h5b6h78h88h5cch668h554h5bbh9fh88h56ch5c8h5fbh566h64dh57bh8bhc7hbdh694h68fh658h686h567hd8h696q,ValueFrom:nil,},EnvVar{Name:OVN_LOGDIR,Value:/tmp,ValueFrom:nil,},EnvVar{Name:OVN_RUNDIR,Value:/tmp,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovndbcluster-nb-etc-ovn,ReadOnly:false,MountPath:/etc/ovn,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovsdb-rundir,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovsdbserver-nb-tls-certs,ReadOnly:true,MountPath:/etc/pki/tls/certs/ovndb.crt,SubPath:tls.crt,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovsdbserver-nb-tls-certs,ReadOnly:true,MountPath:/etc/pki/tls/private/ovndb.key,SubPath:tls.key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovsdbserver-nb-tls-certs,ReadOnly:true,MountPath:/etc/pki/tls/certs/ovndbca.crt,SubPath:ca.crt,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-mqt7w,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/pidof ovsdb-server],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:3,TimeoutSeconds:5,PeriodSeconds:3,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/pidof ovsdb-server],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:&Lifecycle{PostStart:nil,PreStop:&LifecycleHandler{Exec:&ExecAction{Command:[/usr/local/bin/container-scripts/cleanup.sh],},HTTPGet:nil,TCPSocket:nil,Sleep:nil,},},TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/pidof ovsdb-server],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:3,TimeoutSeconds:5,PeriodSeconds:3,SuccessThreshold:1,FailureThreshold:20,TerminationGracePeriodSeconds:nil,},ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovsdbserver-nb-0_openstack(ca11061d-b0a1-464d-9e1d-5572fc439580): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 28 13:35:51 crc kubenswrapper[4631]: E1128 13:35:51.475166 4631 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-ovn-sb-db-server:current-podified" Nov 28 13:35:51 crc kubenswrapper[4631]: E1128 13:35:51.475415 4631 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:ovsdbserver-sb,Image:quay.io/podified-antelope-centos9/openstack-ovn-sb-db-server:current-podified,Command:[/usr/bin/dumb-init],Args:[/usr/local/bin/container-scripts/setup.sh],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n549h644h6ch697h5dh5c6hc4h7chcdhbch8fh677h5dh9fh57hcfhb5hf5h5c7h695h5d9h68h67bh5dch554h5b5h594h64fh9h5f7h57bh7dq,ValueFrom:nil,},EnvVar{Name:OVN_LOGDIR,Value:/tmp,ValueFrom:nil,},EnvVar{Name:OVN_RUNDIR,Value:/tmp,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovndbcluster-sb-etc-ovn,ReadOnly:false,MountPath:/etc/ovn,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovsdb-rundir,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovsdbserver-sb-tls-certs,ReadOnly:true,MountPath:/etc/pki/tls/certs/ovndb.crt,SubPath:tls.crt,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovsdbserver-sb-tls-certs,ReadOnly:true,MountPath:/etc/pki/tls/private/ovndb.key,SubPath:tls.key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovsdbserver-sb-tls-certs,ReadOnly:true,MountPath:/etc/pki/tls/certs/ovndbca.crt,SubPath:ca.crt,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-j99sl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/pidof ovsdb-server],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:3,TimeoutSeconds:5,PeriodSeconds:3,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/pidof ovsdb-server],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:&Lifecycle{PostStart:nil,PreStop:&LifecycleHandler{Exec:&ExecAction{Command:[/usr/local/bin/container-scripts/cleanup.sh],},HTTPGet:nil,TCPSocket:nil,Sleep:nil,},},TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/pidof ovsdb-server],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:3,TimeoutSeconds:5,PeriodSeconds:3,SuccessThreshold:1,FailureThreshold:20,TerminationGracePeriodSeconds:nil,},ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovsdbserver-sb-0_openstack(1d3ba54c-cb26-4e04-8310-34e6b4206119): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 28 13:35:51 crc kubenswrapper[4631]: E1128 13:35:51.834015 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovn-controller\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified\\\"\"" pod="openstack/ovn-controller-7jgdf" podUID="7fc5dcd5-fde5-4a72-8d4a-cda51785e77f" Nov 28 13:35:52 crc kubenswrapper[4631]: E1128 13:35:52.509124 4631 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Nov 28 13:35:52 crc kubenswrapper[4631]: E1128 13:35:52.509392 4631 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n68chd6h679hbfh55fhc6h5ffh5d8h94h56ch589hb4hc5h57bh677hcdh655h8dh667h675h654h66ch567h8fh659h5b4h675h566h55bh54h67dh6dq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-696s2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-666b6646f7-jl8jp_openstack(d605e389-3651-4b5c-a54d-a2013cbb460d): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 28 13:35:52 crc kubenswrapper[4631]: E1128 13:35:52.510656 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-666b6646f7-jl8jp" podUID="d605e389-3651-4b5c-a54d-a2013cbb460d" Nov 28 13:35:52 crc kubenswrapper[4631]: E1128 13:35:52.553877 4631 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Nov 28 13:35:52 crc kubenswrapper[4631]: E1128 13:35:52.554943 4631 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nffh5bdhf4h5f8h79h55h77h58fh56dh7bh6fh578hbch55dh68h56bhd9h65dh57ch658hc9h566h666h688h58h65dh684h5d7h6ch575h5d6h88q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-h49sd,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-675f4bcbfc-mnfbw_openstack(51fb0dc5-be7d-4ffb-a0e1-3050f570b1c3): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 28 13:35:52 crc kubenswrapper[4631]: E1128 13:35:52.556406 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-675f4bcbfc-mnfbw" podUID="51fb0dc5-be7d-4ffb-a0e1-3050f570b1c3" Nov 28 13:35:52 crc kubenswrapper[4631]: E1128 13:35:52.569400 4631 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Nov 28 13:35:52 crc kubenswrapper[4631]: E1128 13:35:52.569567 4631 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ndfhb5h667h568h584h5f9h58dh565h664h587h597h577h64bh5c4h66fh647hbdh68ch5c5h68dh686h5f7h64hd7hc6h55fh57bh98h57fh87h5fh57fq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-ln2hw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-78dd6ddcc-zmbf9_openstack(f884a85b-d7a2-4ca8-aa32-9c4e8578c064): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 28 13:35:52 crc kubenswrapper[4631]: E1128 13:35:52.570937 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-78dd6ddcc-zmbf9" podUID="f884a85b-d7a2-4ca8-aa32-9c4e8578c064" Nov 28 13:35:52 crc kubenswrapper[4631]: E1128 13:35:52.631251 4631 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Nov 28 13:35:52 crc kubenswrapper[4631]: E1128 13:35:52.631412 4631 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n659h4h664hbh658h587h67ch89h587h8fh679hc6hf9h55fh644h5d5h698h68dh5cdh5ffh669h54ch9h689hb8hd4h5bfhd8h5d7h5fh665h574q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-59vt5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-57d769cc4f-clh2s_openstack(a9f46a16-a4a9-4ea0-af16-6c530dcd6386): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 28 13:35:52 crc kubenswrapper[4631]: E1128 13:35:52.632731 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-57d769cc4f-clh2s" podUID="a9f46a16-a4a9-4ea0-af16-6c530dcd6386" Nov 28 13:35:53 crc kubenswrapper[4631]: I1128 13:35:53.224651 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-49s6f"] Nov 28 13:35:53 crc kubenswrapper[4631]: W1128 13:35:53.230249 4631 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1959423d_7154_41e1_80db_d5840a9c1702.slice/crio-9347a7722abe5bca8284ab39251419763b393fa66ee05fa9a291f8c8da05ee5e WatchSource:0}: Error finding container 9347a7722abe5bca8284ab39251419763b393fa66ee05fa9a291f8c8da05ee5e: Status 404 returned error can't find the container with id 9347a7722abe5bca8284ab39251419763b393fa66ee05fa9a291f8c8da05ee5e Nov 28 13:35:53 crc kubenswrapper[4631]: I1128 13:35:53.234117 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-fhmpd"] Nov 28 13:35:53 crc kubenswrapper[4631]: I1128 13:35:53.245246 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-p7l62"] Nov 28 13:35:53 crc kubenswrapper[4631]: W1128 13:35:53.264580 4631 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6dfe4f1a_897a_4ece_a345_4a4f86901079.slice/crio-40b21341244d124b78e981773b4b2331cb79b57cbb4c2fd20969c1dbfb1884eb WatchSource:0}: Error finding container 40b21341244d124b78e981773b4b2331cb79b57cbb4c2fd20969c1dbfb1884eb: Status 404 returned error can't find the container with id 40b21341244d124b78e981773b4b2331cb79b57cbb4c2fd20969c1dbfb1884eb Nov 28 13:35:53 crc kubenswrapper[4631]: I1128 13:35:53.879506 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"4bc20a10-95d5-4003-86d9-c6848bc1b188","Type":"ContainerStarted","Data":"1131c4723e39bfe7eb0b990d0ee799ed781935305e35b754cb9b4890d0de0c5d"} Nov 28 13:35:53 crc kubenswrapper[4631]: I1128 13:35:53.884888 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"1a711f62-44fc-4046-a3aa-b5ef77b45e62","Type":"ContainerStarted","Data":"40143f28bcb20c800b13f8a8474b3814df3d1e3a3516afa655f1a80bc7b35589"} Nov 28 13:35:53 crc kubenswrapper[4631]: I1128 13:35:53.887499 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"83d77da3-fbb7-434f-abaf-5be9b599a9bf","Type":"ContainerStarted","Data":"52d0c55ae2c92fdf222cd8f5c0b2b93b9073d9c66e613db020f87d5dd63419b7"} Nov 28 13:35:53 crc kubenswrapper[4631]: I1128 13:35:53.887860 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Nov 28 13:35:53 crc kubenswrapper[4631]: I1128 13:35:53.890710 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-zmbf9" event={"ID":"f884a85b-d7a2-4ca8-aa32-9c4e8578c064","Type":"ContainerDied","Data":"08468d82eb0f02639f9a54cc2d72b8dcc50ea479306820113b35cb321f72e5a1"} Nov 28 13:35:53 crc kubenswrapper[4631]: I1128 13:35:53.891097 4631 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="08468d82eb0f02639f9a54cc2d72b8dcc50ea479306820113b35cb321f72e5a1" Nov 28 13:35:53 crc kubenswrapper[4631]: I1128 13:35:53.892624 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-clh2s" event={"ID":"a9f46a16-a4a9-4ea0-af16-6c530dcd6386","Type":"ContainerDied","Data":"3be66d5894663175baf25766072009a3f44f58601ca4e7a07b87de050d976c94"} Nov 28 13:35:53 crc kubenswrapper[4631]: I1128 13:35:53.892645 4631 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3be66d5894663175baf25766072009a3f44f58601ca4e7a07b87de050d976c94" Nov 28 13:35:53 crc kubenswrapper[4631]: I1128 13:35:53.894567 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-jl8jp" event={"ID":"d605e389-3651-4b5c-a54d-a2013cbb460d","Type":"ContainerDied","Data":"5d750248066cb1fae2c16050241dc5845c87a70e14401d14bd23d5018a05c1c2"} Nov 28 13:35:53 crc kubenswrapper[4631]: I1128 13:35:53.894689 4631 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5d750248066cb1fae2c16050241dc5845c87a70e14401d14bd23d5018a05c1c2" Nov 28 13:35:53 crc kubenswrapper[4631]: I1128 13:35:53.896343 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-49s6f" event={"ID":"1959423d-7154-41e1-80db-d5840a9c1702","Type":"ContainerStarted","Data":"9347a7722abe5bca8284ab39251419763b393fa66ee05fa9a291f8c8da05ee5e"} Nov 28 13:35:53 crc kubenswrapper[4631]: I1128 13:35:53.897710 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-p7l62" event={"ID":"6dfe4f1a-897a-4ece-a345-4a4f86901079","Type":"ContainerStarted","Data":"40b21341244d124b78e981773b4b2331cb79b57cbb4c2fd20969c1dbfb1884eb"} Nov 28 13:35:53 crc kubenswrapper[4631]: I1128 13:35:53.918509 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"d35228a7-a0fe-496d-9d62-5f1d5a44c72f","Type":"ContainerStarted","Data":"4b9995f6fe28246c44e552e4b30b439a9d83120e6186d4d345cf23bf9372240a"} Nov 28 13:35:53 crc kubenswrapper[4631]: I1128 13:35:53.920586 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Nov 28 13:35:53 crc kubenswrapper[4631]: I1128 13:35:53.933614 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-mnfbw" event={"ID":"51fb0dc5-be7d-4ffb-a0e1-3050f570b1c3","Type":"ContainerDied","Data":"b1034ae3c0de521dfb8b9743d870163a09f3e2170f858fbb3354ecf377eb0b01"} Nov 28 13:35:53 crc kubenswrapper[4631]: I1128 13:35:53.933967 4631 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b1034ae3c0de521dfb8b9743d870163a09f3e2170f858fbb3354ecf377eb0b01" Nov 28 13:35:53 crc kubenswrapper[4631]: I1128 13:35:53.942407 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=2.9911735999999998 podStartE2EDuration="45.942386641s" podCreationTimestamp="2025-11-28 13:35:08 +0000 UTC" firstStartedPulling="2025-11-28 13:35:09.780429924 +0000 UTC m=+866.587733258" lastFinishedPulling="2025-11-28 13:35:52.731642955 +0000 UTC m=+909.538946299" observedRunningTime="2025-11-28 13:35:53.926510445 +0000 UTC m=+910.733813799" watchObservedRunningTime="2025-11-28 13:35:53.942386641 +0000 UTC m=+910.749689985" Nov 28 13:35:53 crc kubenswrapper[4631]: I1128 13:35:53.969298 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fd796d7df-fhmpd" event={"ID":"9594c34e-0cdb-4e81-9e4a-1f542741e05b","Type":"ContainerStarted","Data":"73ce92a6c4b73b8d15ebd733b8712eeb903bdde9bf5914245dd233724ade1795"} Nov 28 13:35:53 crc kubenswrapper[4631]: I1128 13:35:53.978534 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=3.754508856 podStartE2EDuration="48.978504481s" podCreationTimestamp="2025-11-28 13:35:05 +0000 UTC" firstStartedPulling="2025-11-28 13:35:07.330078174 +0000 UTC m=+864.137381518" lastFinishedPulling="2025-11-28 13:35:52.554073799 +0000 UTC m=+909.361377143" observedRunningTime="2025-11-28 13:35:53.976635445 +0000 UTC m=+910.783938789" watchObservedRunningTime="2025-11-28 13:35:53.978504481 +0000 UTC m=+910.785807825" Nov 28 13:35:54 crc kubenswrapper[4631]: I1128 13:35:54.035585 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-zmbf9" Nov 28 13:35:54 crc kubenswrapper[4631]: I1128 13:35:54.050128 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-mnfbw" Nov 28 13:35:54 crc kubenswrapper[4631]: I1128 13:35:54.066801 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-jl8jp" Nov 28 13:35:54 crc kubenswrapper[4631]: I1128 13:35:54.070650 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-clh2s" Nov 28 13:35:54 crc kubenswrapper[4631]: I1128 13:35:54.085892 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f884a85b-d7a2-4ca8-aa32-9c4e8578c064-dns-svc\") pod \"f884a85b-d7a2-4ca8-aa32-9c4e8578c064\" (UID: \"f884a85b-d7a2-4ca8-aa32-9c4e8578c064\") " Nov 28 13:35:54 crc kubenswrapper[4631]: I1128 13:35:54.086013 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ln2hw\" (UniqueName: \"kubernetes.io/projected/f884a85b-d7a2-4ca8-aa32-9c4e8578c064-kube-api-access-ln2hw\") pod \"f884a85b-d7a2-4ca8-aa32-9c4e8578c064\" (UID: \"f884a85b-d7a2-4ca8-aa32-9c4e8578c064\") " Nov 28 13:35:54 crc kubenswrapper[4631]: I1128 13:35:54.086047 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f884a85b-d7a2-4ca8-aa32-9c4e8578c064-config\") pod \"f884a85b-d7a2-4ca8-aa32-9c4e8578c064\" (UID: \"f884a85b-d7a2-4ca8-aa32-9c4e8578c064\") " Nov 28 13:35:54 crc kubenswrapper[4631]: I1128 13:35:54.086273 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h49sd\" (UniqueName: \"kubernetes.io/projected/51fb0dc5-be7d-4ffb-a0e1-3050f570b1c3-kube-api-access-h49sd\") pod \"51fb0dc5-be7d-4ffb-a0e1-3050f570b1c3\" (UID: \"51fb0dc5-be7d-4ffb-a0e1-3050f570b1c3\") " Nov 28 13:35:54 crc kubenswrapper[4631]: I1128 13:35:54.086322 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/51fb0dc5-be7d-4ffb-a0e1-3050f570b1c3-config\") pod \"51fb0dc5-be7d-4ffb-a0e1-3050f570b1c3\" (UID: \"51fb0dc5-be7d-4ffb-a0e1-3050f570b1c3\") " Nov 28 13:35:54 crc kubenswrapper[4631]: I1128 13:35:54.087051 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f884a85b-d7a2-4ca8-aa32-9c4e8578c064-config" (OuterVolumeSpecName: "config") pod "f884a85b-d7a2-4ca8-aa32-9c4e8578c064" (UID: "f884a85b-d7a2-4ca8-aa32-9c4e8578c064"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 13:35:54 crc kubenswrapper[4631]: I1128 13:35:54.087195 4631 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f884a85b-d7a2-4ca8-aa32-9c4e8578c064-config\") on node \"crc\" DevicePath \"\"" Nov 28 13:35:54 crc kubenswrapper[4631]: I1128 13:35:54.087787 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f884a85b-d7a2-4ca8-aa32-9c4e8578c064-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "f884a85b-d7a2-4ca8-aa32-9c4e8578c064" (UID: "f884a85b-d7a2-4ca8-aa32-9c4e8578c064"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 13:35:54 crc kubenswrapper[4631]: I1128 13:35:54.087950 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/51fb0dc5-be7d-4ffb-a0e1-3050f570b1c3-config" (OuterVolumeSpecName: "config") pod "51fb0dc5-be7d-4ffb-a0e1-3050f570b1c3" (UID: "51fb0dc5-be7d-4ffb-a0e1-3050f570b1c3"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 13:35:54 crc kubenswrapper[4631]: I1128 13:35:54.095246 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f884a85b-d7a2-4ca8-aa32-9c4e8578c064-kube-api-access-ln2hw" (OuterVolumeSpecName: "kube-api-access-ln2hw") pod "f884a85b-d7a2-4ca8-aa32-9c4e8578c064" (UID: "f884a85b-d7a2-4ca8-aa32-9c4e8578c064"). InnerVolumeSpecName "kube-api-access-ln2hw". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:35:54 crc kubenswrapper[4631]: I1128 13:35:54.117498 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/51fb0dc5-be7d-4ffb-a0e1-3050f570b1c3-kube-api-access-h49sd" (OuterVolumeSpecName: "kube-api-access-h49sd") pod "51fb0dc5-be7d-4ffb-a0e1-3050f570b1c3" (UID: "51fb0dc5-be7d-4ffb-a0e1-3050f570b1c3"). InnerVolumeSpecName "kube-api-access-h49sd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:35:54 crc kubenswrapper[4631]: I1128 13:35:54.187988 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a9f46a16-a4a9-4ea0-af16-6c530dcd6386-config\") pod \"a9f46a16-a4a9-4ea0-af16-6c530dcd6386\" (UID: \"a9f46a16-a4a9-4ea0-af16-6c530dcd6386\") " Nov 28 13:35:54 crc kubenswrapper[4631]: I1128 13:35:54.188147 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-696s2\" (UniqueName: \"kubernetes.io/projected/d605e389-3651-4b5c-a54d-a2013cbb460d-kube-api-access-696s2\") pod \"d605e389-3651-4b5c-a54d-a2013cbb460d\" (UID: \"d605e389-3651-4b5c-a54d-a2013cbb460d\") " Nov 28 13:35:54 crc kubenswrapper[4631]: I1128 13:35:54.188196 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d605e389-3651-4b5c-a54d-a2013cbb460d-dns-svc\") pod \"d605e389-3651-4b5c-a54d-a2013cbb460d\" (UID: \"d605e389-3651-4b5c-a54d-a2013cbb460d\") " Nov 28 13:35:54 crc kubenswrapper[4631]: I1128 13:35:54.188234 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-59vt5\" (UniqueName: \"kubernetes.io/projected/a9f46a16-a4a9-4ea0-af16-6c530dcd6386-kube-api-access-59vt5\") pod \"a9f46a16-a4a9-4ea0-af16-6c530dcd6386\" (UID: \"a9f46a16-a4a9-4ea0-af16-6c530dcd6386\") " Nov 28 13:35:54 crc kubenswrapper[4631]: I1128 13:35:54.188408 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a9f46a16-a4a9-4ea0-af16-6c530dcd6386-dns-svc\") pod \"a9f46a16-a4a9-4ea0-af16-6c530dcd6386\" (UID: \"a9f46a16-a4a9-4ea0-af16-6c530dcd6386\") " Nov 28 13:35:54 crc kubenswrapper[4631]: I1128 13:35:54.188460 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d605e389-3651-4b5c-a54d-a2013cbb460d-config\") pod \"d605e389-3651-4b5c-a54d-a2013cbb460d\" (UID: \"d605e389-3651-4b5c-a54d-a2013cbb460d\") " Nov 28 13:35:54 crc kubenswrapper[4631]: I1128 13:35:54.188811 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h49sd\" (UniqueName: \"kubernetes.io/projected/51fb0dc5-be7d-4ffb-a0e1-3050f570b1c3-kube-api-access-h49sd\") on node \"crc\" DevicePath \"\"" Nov 28 13:35:54 crc kubenswrapper[4631]: I1128 13:35:54.188828 4631 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/51fb0dc5-be7d-4ffb-a0e1-3050f570b1c3-config\") on node \"crc\" DevicePath \"\"" Nov 28 13:35:54 crc kubenswrapper[4631]: I1128 13:35:54.188840 4631 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f884a85b-d7a2-4ca8-aa32-9c4e8578c064-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 28 13:35:54 crc kubenswrapper[4631]: I1128 13:35:54.188850 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ln2hw\" (UniqueName: \"kubernetes.io/projected/f884a85b-d7a2-4ca8-aa32-9c4e8578c064-kube-api-access-ln2hw\") on node \"crc\" DevicePath \"\"" Nov 28 13:35:54 crc kubenswrapper[4631]: I1128 13:35:54.189026 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a9f46a16-a4a9-4ea0-af16-6c530dcd6386-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "a9f46a16-a4a9-4ea0-af16-6c530dcd6386" (UID: "a9f46a16-a4a9-4ea0-af16-6c530dcd6386"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 13:35:54 crc kubenswrapper[4631]: I1128 13:35:54.189064 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d605e389-3651-4b5c-a54d-a2013cbb460d-config" (OuterVolumeSpecName: "config") pod "d605e389-3651-4b5c-a54d-a2013cbb460d" (UID: "d605e389-3651-4b5c-a54d-a2013cbb460d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 13:35:54 crc kubenswrapper[4631]: I1128 13:35:54.189519 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d605e389-3651-4b5c-a54d-a2013cbb460d-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "d605e389-3651-4b5c-a54d-a2013cbb460d" (UID: "d605e389-3651-4b5c-a54d-a2013cbb460d"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 13:35:54 crc kubenswrapper[4631]: I1128 13:35:54.189910 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a9f46a16-a4a9-4ea0-af16-6c530dcd6386-config" (OuterVolumeSpecName: "config") pod "a9f46a16-a4a9-4ea0-af16-6c530dcd6386" (UID: "a9f46a16-a4a9-4ea0-af16-6c530dcd6386"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 13:35:54 crc kubenswrapper[4631]: I1128 13:35:54.192628 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a9f46a16-a4a9-4ea0-af16-6c530dcd6386-kube-api-access-59vt5" (OuterVolumeSpecName: "kube-api-access-59vt5") pod "a9f46a16-a4a9-4ea0-af16-6c530dcd6386" (UID: "a9f46a16-a4a9-4ea0-af16-6c530dcd6386"). InnerVolumeSpecName "kube-api-access-59vt5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:35:54 crc kubenswrapper[4631]: I1128 13:35:54.193899 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d605e389-3651-4b5c-a54d-a2013cbb460d-kube-api-access-696s2" (OuterVolumeSpecName: "kube-api-access-696s2") pod "d605e389-3651-4b5c-a54d-a2013cbb460d" (UID: "d605e389-3651-4b5c-a54d-a2013cbb460d"). InnerVolumeSpecName "kube-api-access-696s2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:35:54 crc kubenswrapper[4631]: I1128 13:35:54.290360 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-59vt5\" (UniqueName: \"kubernetes.io/projected/a9f46a16-a4a9-4ea0-af16-6c530dcd6386-kube-api-access-59vt5\") on node \"crc\" DevicePath \"\"" Nov 28 13:35:54 crc kubenswrapper[4631]: I1128 13:35:54.290740 4631 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a9f46a16-a4a9-4ea0-af16-6c530dcd6386-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 28 13:35:54 crc kubenswrapper[4631]: I1128 13:35:54.290752 4631 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d605e389-3651-4b5c-a54d-a2013cbb460d-config\") on node \"crc\" DevicePath \"\"" Nov 28 13:35:54 crc kubenswrapper[4631]: I1128 13:35:54.290769 4631 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a9f46a16-a4a9-4ea0-af16-6c530dcd6386-config\") on node \"crc\" DevicePath \"\"" Nov 28 13:35:54 crc kubenswrapper[4631]: I1128 13:35:54.290780 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-696s2\" (UniqueName: \"kubernetes.io/projected/d605e389-3651-4b5c-a54d-a2013cbb460d-kube-api-access-696s2\") on node \"crc\" DevicePath \"\"" Nov 28 13:35:54 crc kubenswrapper[4631]: I1128 13:35:54.290791 4631 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d605e389-3651-4b5c-a54d-a2013cbb460d-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 28 13:35:54 crc kubenswrapper[4631]: I1128 13:35:54.986351 4631 generic.go:334] "Generic (PLEG): container finished" podID="9594c34e-0cdb-4e81-9e4a-1f542741e05b" containerID="b02c8505e2ae8141c62f5628c67f3e83e9662cf7e379ad1640849c61ad28e2c1" exitCode=0 Nov 28 13:35:54 crc kubenswrapper[4631]: I1128 13:35:54.986541 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fd796d7df-fhmpd" event={"ID":"9594c34e-0cdb-4e81-9e4a-1f542741e05b","Type":"ContainerDied","Data":"b02c8505e2ae8141c62f5628c67f3e83e9662cf7e379ad1640849c61ad28e2c1"} Nov 28 13:35:54 crc kubenswrapper[4631]: I1128 13:35:54.989442 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"ed4141f2-5eff-4252-8dc7-ad60a1580189","Type":"ContainerStarted","Data":"41d8c218651f1d165fa68a2452a35246fa25140ebacda04483d76b1fc68b2e42"} Nov 28 13:35:54 crc kubenswrapper[4631]: I1128 13:35:54.994483 4631 generic.go:334] "Generic (PLEG): container finished" podID="1959423d-7154-41e1-80db-d5840a9c1702" containerID="763e2d68c0063a8e7bec77440ee5a31d6b77bc422f4805a4ce9c2183d94e6d55" exitCode=0 Nov 28 13:35:54 crc kubenswrapper[4631]: I1128 13:35:54.994558 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-49s6f" event={"ID":"1959423d-7154-41e1-80db-d5840a9c1702","Type":"ContainerDied","Data":"763e2d68c0063a8e7bec77440ee5a31d6b77bc422f4805a4ce9c2183d94e6d55"} Nov 28 13:35:54 crc kubenswrapper[4631]: I1128 13:35:54.998135 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"3a91e9fa-34b7-4688-9171-56a62043759d","Type":"ContainerStarted","Data":"5f2c613e618512327c6478adcd7de0b71456862e29e5333eaab2b49e071f2777"} Nov 28 13:35:54 crc kubenswrapper[4631]: I1128 13:35:54.998421 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-jl8jp" Nov 28 13:35:54 crc kubenswrapper[4631]: I1128 13:35:54.998502 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-mnfbw" Nov 28 13:35:54 crc kubenswrapper[4631]: I1128 13:35:54.998573 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-zmbf9" Nov 28 13:35:55 crc kubenswrapper[4631]: I1128 13:35:54.999534 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-clh2s" Nov 28 13:35:55 crc kubenswrapper[4631]: I1128 13:35:55.172724 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-jl8jp"] Nov 28 13:35:55 crc kubenswrapper[4631]: I1128 13:35:55.194448 4631 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-jl8jp"] Nov 28 13:35:55 crc kubenswrapper[4631]: I1128 13:35:55.222247 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-clh2s"] Nov 28 13:35:55 crc kubenswrapper[4631]: I1128 13:35:55.235989 4631 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-clh2s"] Nov 28 13:35:55 crc kubenswrapper[4631]: I1128 13:35:55.258893 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-mnfbw"] Nov 28 13:35:55 crc kubenswrapper[4631]: I1128 13:35:55.263792 4631 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-mnfbw"] Nov 28 13:35:55 crc kubenswrapper[4631]: I1128 13:35:55.275156 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-zmbf9"] Nov 28 13:35:55 crc kubenswrapper[4631]: I1128 13:35:55.280055 4631 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-zmbf9"] Nov 28 13:35:55 crc kubenswrapper[4631]: I1128 13:35:55.529264 4631 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="51fb0dc5-be7d-4ffb-a0e1-3050f570b1c3" path="/var/lib/kubelet/pods/51fb0dc5-be7d-4ffb-a0e1-3050f570b1c3/volumes" Nov 28 13:35:55 crc kubenswrapper[4631]: I1128 13:35:55.529836 4631 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a9f46a16-a4a9-4ea0-af16-6c530dcd6386" path="/var/lib/kubelet/pods/a9f46a16-a4a9-4ea0-af16-6c530dcd6386/volumes" Nov 28 13:35:55 crc kubenswrapper[4631]: I1128 13:35:55.530312 4631 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d605e389-3651-4b5c-a54d-a2013cbb460d" path="/var/lib/kubelet/pods/d605e389-3651-4b5c-a54d-a2013cbb460d/volumes" Nov 28 13:35:55 crc kubenswrapper[4631]: I1128 13:35:55.530880 4631 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f884a85b-d7a2-4ca8-aa32-9c4e8578c064" path="/var/lib/kubelet/pods/f884a85b-d7a2-4ca8-aa32-9c4e8578c064/volumes" Nov 28 13:35:57 crc kubenswrapper[4631]: E1128 13:35:57.315632 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovsdbserver-nb\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/ovsdbserver-nb-0" podUID="ca11061d-b0a1-464d-9e1d-5572fc439580" Nov 28 13:35:57 crc kubenswrapper[4631]: E1128 13:35:57.322314 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovsdbserver-sb\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/ovsdbserver-sb-0" podUID="1d3ba54c-cb26-4e04-8310-34e6b4206119" Nov 28 13:35:58 crc kubenswrapper[4631]: I1128 13:35:58.035789 4631 generic.go:334] "Generic (PLEG): container finished" podID="1a711f62-44fc-4046-a3aa-b5ef77b45e62" containerID="40143f28bcb20c800b13f8a8474b3814df3d1e3a3516afa655f1a80bc7b35589" exitCode=0 Nov 28 13:35:58 crc kubenswrapper[4631]: I1128 13:35:58.035864 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"1a711f62-44fc-4046-a3aa-b5ef77b45e62","Type":"ContainerDied","Data":"40143f28bcb20c800b13f8a8474b3814df3d1e3a3516afa655f1a80bc7b35589"} Nov 28 13:35:58 crc kubenswrapper[4631]: I1128 13:35:58.041202 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"ca11061d-b0a1-464d-9e1d-5572fc439580","Type":"ContainerStarted","Data":"dd2ce84ae98422b5a2bec25beb5c410bb15086f19c08d0f00fc92d76a122609f"} Nov 28 13:35:58 crc kubenswrapper[4631]: E1128 13:35:58.043273 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovsdbserver-nb\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-ovn-nb-db-server:current-podified\\\"\"" pod="openstack/ovsdbserver-nb-0" podUID="ca11061d-b0a1-464d-9e1d-5572fc439580" Nov 28 13:35:58 crc kubenswrapper[4631]: I1128 13:35:58.046598 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-49s6f" event={"ID":"1959423d-7154-41e1-80db-d5840a9c1702","Type":"ContainerStarted","Data":"5cbe99b53ff72888b0833a40a1bbcc9dd936f48191d7159210c2429b4be79dad"} Nov 28 13:35:58 crc kubenswrapper[4631]: I1128 13:35:58.046758 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-86db49b7ff-49s6f" Nov 28 13:35:58 crc kubenswrapper[4631]: I1128 13:35:58.049327 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"1d3ba54c-cb26-4e04-8310-34e6b4206119","Type":"ContainerStarted","Data":"59d669fa35926b85a57d640c7a7380a6764a5a2aa44f762393ce3ec3944b86ab"} Nov 28 13:35:58 crc kubenswrapper[4631]: I1128 13:35:58.053085 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-p7l62" event={"ID":"6dfe4f1a-897a-4ece-a345-4a4f86901079","Type":"ContainerStarted","Data":"04b3db0cbd9cd5b0d07f53fe2c2b2e01334cd3e8a60f3d0a88dd6c8c71c58317"} Nov 28 13:35:58 crc kubenswrapper[4631]: E1128 13:35:58.055189 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovsdbserver-sb\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-ovn-sb-db-server:current-podified\\\"\"" pod="openstack/ovsdbserver-sb-0" podUID="1d3ba54c-cb26-4e04-8310-34e6b4206119" Nov 28 13:35:58 crc kubenswrapper[4631]: I1128 13:35:58.057173 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fd796d7df-fhmpd" event={"ID":"9594c34e-0cdb-4e81-9e4a-1f542741e05b","Type":"ContainerStarted","Data":"265d530085b2141c8bbc6e5a0188ade7a1d462da994668c7ff64e5e292000555"} Nov 28 13:35:58 crc kubenswrapper[4631]: I1128 13:35:58.058609 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7fd796d7df-fhmpd" Nov 28 13:35:58 crc kubenswrapper[4631]: I1128 13:35:58.060676 4631 generic.go:334] "Generic (PLEG): container finished" podID="4bc20a10-95d5-4003-86d9-c6848bc1b188" containerID="1131c4723e39bfe7eb0b990d0ee799ed781935305e35b754cb9b4890d0de0c5d" exitCode=0 Nov 28 13:35:58 crc kubenswrapper[4631]: I1128 13:35:58.060721 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"4bc20a10-95d5-4003-86d9-c6848bc1b188","Type":"ContainerDied","Data":"1131c4723e39bfe7eb0b990d0ee799ed781935305e35b754cb9b4890d0de0c5d"} Nov 28 13:35:58 crc kubenswrapper[4631]: I1128 13:35:58.129280 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-p7l62" podStartSLOduration=30.361324606 podStartE2EDuration="34.129248594s" podCreationTimestamp="2025-11-28 13:35:24 +0000 UTC" firstStartedPulling="2025-11-28 13:35:53.274051239 +0000 UTC m=+910.081354583" lastFinishedPulling="2025-11-28 13:35:57.041975227 +0000 UTC m=+913.849278571" observedRunningTime="2025-11-28 13:35:58.116209789 +0000 UTC m=+914.923513133" watchObservedRunningTime="2025-11-28 13:35:58.129248594 +0000 UTC m=+914.936551948" Nov 28 13:35:58 crc kubenswrapper[4631]: I1128 13:35:58.170413 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-86db49b7ff-49s6f" podStartSLOduration=33.193207167 podStartE2EDuration="34.17039108s" podCreationTimestamp="2025-11-28 13:35:24 +0000 UTC" firstStartedPulling="2025-11-28 13:35:53.236376719 +0000 UTC m=+910.043680063" lastFinishedPulling="2025-11-28 13:35:54.213560632 +0000 UTC m=+911.020863976" observedRunningTime="2025-11-28 13:35:58.148217287 +0000 UTC m=+914.955520641" watchObservedRunningTime="2025-11-28 13:35:58.17039108 +0000 UTC m=+914.977694424" Nov 28 13:35:58 crc kubenswrapper[4631]: I1128 13:35:58.283343 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7fd796d7df-fhmpd" podStartSLOduration=32.898420698 podStartE2EDuration="34.283312325s" podCreationTimestamp="2025-11-28 13:35:24 +0000 UTC" firstStartedPulling="2025-11-28 13:35:53.259750842 +0000 UTC m=+910.067054186" lastFinishedPulling="2025-11-28 13:35:54.644642469 +0000 UTC m=+911.451945813" observedRunningTime="2025-11-28 13:35:58.267332156 +0000 UTC m=+915.074635500" watchObservedRunningTime="2025-11-28 13:35:58.283312325 +0000 UTC m=+915.090615669" Nov 28 13:35:58 crc kubenswrapper[4631]: I1128 13:35:58.603147 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Nov 28 13:35:59 crc kubenswrapper[4631]: I1128 13:35:59.071026 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"4bc20a10-95d5-4003-86d9-c6848bc1b188","Type":"ContainerStarted","Data":"94df700e4f40479d1c4e2802f1606cd6a78f01291947028531a3f8c8bbbc185c"} Nov 28 13:35:59 crc kubenswrapper[4631]: I1128 13:35:59.075181 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"1a711f62-44fc-4046-a3aa-b5ef77b45e62","Type":"ContainerStarted","Data":"08a376cf7b2ce6a36b6048fcd83ea8160710cbd45e50f1a370c1c32e7d1c736a"} Nov 28 13:35:59 crc kubenswrapper[4631]: E1128 13:35:59.077420 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovsdbserver-sb\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-ovn-sb-db-server:current-podified\\\"\"" pod="openstack/ovsdbserver-sb-0" podUID="1d3ba54c-cb26-4e04-8310-34e6b4206119" Nov 28 13:35:59 crc kubenswrapper[4631]: E1128 13:35:59.078156 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovsdbserver-nb\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-ovn-nb-db-server:current-podified\\\"\"" pod="openstack/ovsdbserver-nb-0" podUID="ca11061d-b0a1-464d-9e1d-5572fc439580" Nov 28 13:35:59 crc kubenswrapper[4631]: I1128 13:35:59.098944 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=11.717805808 podStartE2EDuration="56.098923099s" podCreationTimestamp="2025-11-28 13:35:03 +0000 UTC" firstStartedPulling="2025-11-28 13:35:07.079423845 +0000 UTC m=+863.886727189" lastFinishedPulling="2025-11-28 13:35:51.460541136 +0000 UTC m=+908.267844480" observedRunningTime="2025-11-28 13:35:59.094390466 +0000 UTC m=+915.901693820" watchObservedRunningTime="2025-11-28 13:35:59.098923099 +0000 UTC m=+915.906226433" Nov 28 13:35:59 crc kubenswrapper[4631]: I1128 13:35:59.180207 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=9.281359074 podStartE2EDuration="57.180187195s" podCreationTimestamp="2025-11-28 13:35:02 +0000 UTC" firstStartedPulling="2025-11-28 13:35:04.784062 +0000 UTC m=+861.591365434" lastFinishedPulling="2025-11-28 13:35:52.682890221 +0000 UTC m=+909.490193555" observedRunningTime="2025-11-28 13:35:59.168493754 +0000 UTC m=+915.975797098" watchObservedRunningTime="2025-11-28 13:35:59.180187195 +0000 UTC m=+915.987490539" Nov 28 13:36:00 crc kubenswrapper[4631]: I1128 13:36:00.951585 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Nov 28 13:36:04 crc kubenswrapper[4631]: I1128 13:36:04.068571 4631 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Nov 28 13:36:04 crc kubenswrapper[4631]: I1128 13:36:04.068975 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Nov 28 13:36:04 crc kubenswrapper[4631]: I1128 13:36:04.202217 4631 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Nov 28 13:36:04 crc kubenswrapper[4631]: I1128 13:36:04.300166 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Nov 28 13:36:04 crc kubenswrapper[4631]: I1128 13:36:04.967510 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7fd796d7df-fhmpd" Nov 28 13:36:05 crc kubenswrapper[4631]: I1128 13:36:05.153249 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-86db49b7ff-49s6f" Nov 28 13:36:05 crc kubenswrapper[4631]: I1128 13:36:05.256950 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-fhmpd"] Nov 28 13:36:05 crc kubenswrapper[4631]: I1128 13:36:05.261477 4631 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7fd796d7df-fhmpd" podUID="9594c34e-0cdb-4e81-9e4a-1f542741e05b" containerName="dnsmasq-dns" containerID="cri-o://265d530085b2141c8bbc6e5a0188ade7a1d462da994668c7ff64e5e292000555" gracePeriod=10 Nov 28 13:36:05 crc kubenswrapper[4631]: I1128 13:36:05.503677 4631 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Nov 28 13:36:05 crc kubenswrapper[4631]: I1128 13:36:05.503719 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Nov 28 13:36:05 crc kubenswrapper[4631]: I1128 13:36:05.592766 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-d5ea-account-create-update-cw75f"] Nov 28 13:36:05 crc kubenswrapper[4631]: I1128 13:36:05.598036 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-d5ea-account-create-update-cw75f"] Nov 28 13:36:05 crc kubenswrapper[4631]: I1128 13:36:05.599033 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-d5ea-account-create-update-cw75f" Nov 28 13:36:05 crc kubenswrapper[4631]: I1128 13:36:05.603955 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Nov 28 13:36:05 crc kubenswrapper[4631]: I1128 13:36:05.655702 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-9m2n2"] Nov 28 13:36:05 crc kubenswrapper[4631]: I1128 13:36:05.657547 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-9m2n2" Nov 28 13:36:05 crc kubenswrapper[4631]: I1128 13:36:05.681563 4631 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Nov 28 13:36:05 crc kubenswrapper[4631]: I1128 13:36:05.687776 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-9m2n2"] Nov 28 13:36:05 crc kubenswrapper[4631]: I1128 13:36:05.752410 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3edb21b5-935d-4b42-bc88-d590b7fb0e3d-operator-scripts\") pod \"keystone-db-create-9m2n2\" (UID: \"3edb21b5-935d-4b42-bc88-d590b7fb0e3d\") " pod="openstack/keystone-db-create-9m2n2" Nov 28 13:36:05 crc kubenswrapper[4631]: I1128 13:36:05.752459 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g9jlq\" (UniqueName: \"kubernetes.io/projected/3edb21b5-935d-4b42-bc88-d590b7fb0e3d-kube-api-access-g9jlq\") pod \"keystone-db-create-9m2n2\" (UID: \"3edb21b5-935d-4b42-bc88-d590b7fb0e3d\") " pod="openstack/keystone-db-create-9m2n2" Nov 28 13:36:05 crc kubenswrapper[4631]: I1128 13:36:05.752518 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dd84f393-e43d-45e8-8760-7e0643f7c30e-operator-scripts\") pod \"keystone-d5ea-account-create-update-cw75f\" (UID: \"dd84f393-e43d-45e8-8760-7e0643f7c30e\") " pod="openstack/keystone-d5ea-account-create-update-cw75f" Nov 28 13:36:05 crc kubenswrapper[4631]: I1128 13:36:05.752600 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9fmjt\" (UniqueName: \"kubernetes.io/projected/dd84f393-e43d-45e8-8760-7e0643f7c30e-kube-api-access-9fmjt\") pod \"keystone-d5ea-account-create-update-cw75f\" (UID: \"dd84f393-e43d-45e8-8760-7e0643f7c30e\") " pod="openstack/keystone-d5ea-account-create-update-cw75f" Nov 28 13:36:05 crc kubenswrapper[4631]: I1128 13:36:05.780564 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fd796d7df-fhmpd" Nov 28 13:36:05 crc kubenswrapper[4631]: I1128 13:36:05.854451 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dd84f393-e43d-45e8-8760-7e0643f7c30e-operator-scripts\") pod \"keystone-d5ea-account-create-update-cw75f\" (UID: \"dd84f393-e43d-45e8-8760-7e0643f7c30e\") " pod="openstack/keystone-d5ea-account-create-update-cw75f" Nov 28 13:36:05 crc kubenswrapper[4631]: I1128 13:36:05.854562 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9fmjt\" (UniqueName: \"kubernetes.io/projected/dd84f393-e43d-45e8-8760-7e0643f7c30e-kube-api-access-9fmjt\") pod \"keystone-d5ea-account-create-update-cw75f\" (UID: \"dd84f393-e43d-45e8-8760-7e0643f7c30e\") " pod="openstack/keystone-d5ea-account-create-update-cw75f" Nov 28 13:36:05 crc kubenswrapper[4631]: I1128 13:36:05.854602 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3edb21b5-935d-4b42-bc88-d590b7fb0e3d-operator-scripts\") pod \"keystone-db-create-9m2n2\" (UID: \"3edb21b5-935d-4b42-bc88-d590b7fb0e3d\") " pod="openstack/keystone-db-create-9m2n2" Nov 28 13:36:05 crc kubenswrapper[4631]: I1128 13:36:05.854632 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g9jlq\" (UniqueName: \"kubernetes.io/projected/3edb21b5-935d-4b42-bc88-d590b7fb0e3d-kube-api-access-g9jlq\") pod \"keystone-db-create-9m2n2\" (UID: \"3edb21b5-935d-4b42-bc88-d590b7fb0e3d\") " pod="openstack/keystone-db-create-9m2n2" Nov 28 13:36:05 crc kubenswrapper[4631]: I1128 13:36:05.855772 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3edb21b5-935d-4b42-bc88-d590b7fb0e3d-operator-scripts\") pod \"keystone-db-create-9m2n2\" (UID: \"3edb21b5-935d-4b42-bc88-d590b7fb0e3d\") " pod="openstack/keystone-db-create-9m2n2" Nov 28 13:36:05 crc kubenswrapper[4631]: I1128 13:36:05.855782 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dd84f393-e43d-45e8-8760-7e0643f7c30e-operator-scripts\") pod \"keystone-d5ea-account-create-update-cw75f\" (UID: \"dd84f393-e43d-45e8-8760-7e0643f7c30e\") " pod="openstack/keystone-d5ea-account-create-update-cw75f" Nov 28 13:36:05 crc kubenswrapper[4631]: I1128 13:36:05.873593 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9fmjt\" (UniqueName: \"kubernetes.io/projected/dd84f393-e43d-45e8-8760-7e0643f7c30e-kube-api-access-9fmjt\") pod \"keystone-d5ea-account-create-update-cw75f\" (UID: \"dd84f393-e43d-45e8-8760-7e0643f7c30e\") " pod="openstack/keystone-d5ea-account-create-update-cw75f" Nov 28 13:36:05 crc kubenswrapper[4631]: I1128 13:36:05.874097 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g9jlq\" (UniqueName: \"kubernetes.io/projected/3edb21b5-935d-4b42-bc88-d590b7fb0e3d-kube-api-access-g9jlq\") pod \"keystone-db-create-9m2n2\" (UID: \"3edb21b5-935d-4b42-bc88-d590b7fb0e3d\") " pod="openstack/keystone-db-create-9m2n2" Nov 28 13:36:05 crc kubenswrapper[4631]: I1128 13:36:05.931336 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-d5ea-account-create-update-cw75f" Nov 28 13:36:05 crc kubenswrapper[4631]: I1128 13:36:05.955400 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9594c34e-0cdb-4e81-9e4a-1f542741e05b-config\") pod \"9594c34e-0cdb-4e81-9e4a-1f542741e05b\" (UID: \"9594c34e-0cdb-4e81-9e4a-1f542741e05b\") " Nov 28 13:36:05 crc kubenswrapper[4631]: I1128 13:36:05.955512 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mvcxt\" (UniqueName: \"kubernetes.io/projected/9594c34e-0cdb-4e81-9e4a-1f542741e05b-kube-api-access-mvcxt\") pod \"9594c34e-0cdb-4e81-9e4a-1f542741e05b\" (UID: \"9594c34e-0cdb-4e81-9e4a-1f542741e05b\") " Nov 28 13:36:05 crc kubenswrapper[4631]: I1128 13:36:05.955654 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9594c34e-0cdb-4e81-9e4a-1f542741e05b-ovsdbserver-nb\") pod \"9594c34e-0cdb-4e81-9e4a-1f542741e05b\" (UID: \"9594c34e-0cdb-4e81-9e4a-1f542741e05b\") " Nov 28 13:36:05 crc kubenswrapper[4631]: I1128 13:36:05.955695 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9594c34e-0cdb-4e81-9e4a-1f542741e05b-dns-svc\") pod \"9594c34e-0cdb-4e81-9e4a-1f542741e05b\" (UID: \"9594c34e-0cdb-4e81-9e4a-1f542741e05b\") " Nov 28 13:36:05 crc kubenswrapper[4631]: I1128 13:36:05.961200 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9594c34e-0cdb-4e81-9e4a-1f542741e05b-kube-api-access-mvcxt" (OuterVolumeSpecName: "kube-api-access-mvcxt") pod "9594c34e-0cdb-4e81-9e4a-1f542741e05b" (UID: "9594c34e-0cdb-4e81-9e4a-1f542741e05b"). InnerVolumeSpecName "kube-api-access-mvcxt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:36:05 crc kubenswrapper[4631]: I1128 13:36:05.984513 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-9m2n2" Nov 28 13:36:05 crc kubenswrapper[4631]: I1128 13:36:05.999461 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9594c34e-0cdb-4e81-9e4a-1f542741e05b-config" (OuterVolumeSpecName: "config") pod "9594c34e-0cdb-4e81-9e4a-1f542741e05b" (UID: "9594c34e-0cdb-4e81-9e4a-1f542741e05b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 13:36:05 crc kubenswrapper[4631]: I1128 13:36:05.999643 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9594c34e-0cdb-4e81-9e4a-1f542741e05b-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "9594c34e-0cdb-4e81-9e4a-1f542741e05b" (UID: "9594c34e-0cdb-4e81-9e4a-1f542741e05b"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 13:36:05 crc kubenswrapper[4631]: I1128 13:36:05.999706 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9594c34e-0cdb-4e81-9e4a-1f542741e05b-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "9594c34e-0cdb-4e81-9e4a-1f542741e05b" (UID: "9594c34e-0cdb-4e81-9e4a-1f542741e05b"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 13:36:06 crc kubenswrapper[4631]: I1128 13:36:06.080442 4631 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9594c34e-0cdb-4e81-9e4a-1f542741e05b-config\") on node \"crc\" DevicePath \"\"" Nov 28 13:36:06 crc kubenswrapper[4631]: I1128 13:36:06.080479 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mvcxt\" (UniqueName: \"kubernetes.io/projected/9594c34e-0cdb-4e81-9e4a-1f542741e05b-kube-api-access-mvcxt\") on node \"crc\" DevicePath \"\"" Nov 28 13:36:06 crc kubenswrapper[4631]: I1128 13:36:06.080496 4631 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9594c34e-0cdb-4e81-9e4a-1f542741e05b-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 28 13:36:06 crc kubenswrapper[4631]: I1128 13:36:06.080507 4631 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9594c34e-0cdb-4e81-9e4a-1f542741e05b-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 28 13:36:06 crc kubenswrapper[4631]: I1128 13:36:06.127864 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-68dgq"] Nov 28 13:36:06 crc kubenswrapper[4631]: E1128 13:36:06.128344 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9594c34e-0cdb-4e81-9e4a-1f542741e05b" containerName="init" Nov 28 13:36:06 crc kubenswrapper[4631]: I1128 13:36:06.128356 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="9594c34e-0cdb-4e81-9e4a-1f542741e05b" containerName="init" Nov 28 13:36:06 crc kubenswrapper[4631]: E1128 13:36:06.128381 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9594c34e-0cdb-4e81-9e4a-1f542741e05b" containerName="dnsmasq-dns" Nov 28 13:36:06 crc kubenswrapper[4631]: I1128 13:36:06.128387 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="9594c34e-0cdb-4e81-9e4a-1f542741e05b" containerName="dnsmasq-dns" Nov 28 13:36:06 crc kubenswrapper[4631]: I1128 13:36:06.128553 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="9594c34e-0cdb-4e81-9e4a-1f542741e05b" containerName="dnsmasq-dns" Nov 28 13:36:06 crc kubenswrapper[4631]: I1128 13:36:06.129141 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-68dgq" Nov 28 13:36:06 crc kubenswrapper[4631]: I1128 13:36:06.138385 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-68dgq"] Nov 28 13:36:06 crc kubenswrapper[4631]: I1128 13:36:06.164366 4631 generic.go:334] "Generic (PLEG): container finished" podID="7fc0ee4e-435d-4297-94eb-9e8c6fa90004" containerID="628bada2fda6b7eaf632ce128d6ff5922c0f35209346cd69e0be28511babdf32" exitCode=0 Nov 28 13:36:06 crc kubenswrapper[4631]: I1128 13:36:06.164451 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-lskqw" event={"ID":"7fc0ee4e-435d-4297-94eb-9e8c6fa90004","Type":"ContainerDied","Data":"628bada2fda6b7eaf632ce128d6ff5922c0f35209346cd69e0be28511babdf32"} Nov 28 13:36:06 crc kubenswrapper[4631]: I1128 13:36:06.178029 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-7jgdf" event={"ID":"7fc5dcd5-fde5-4a72-8d4a-cda51785e77f","Type":"ContainerStarted","Data":"40c7e743adcf10afc63cdb02026c32d07c7139e06d55a3fd956df8f15af52403"} Nov 28 13:36:06 crc kubenswrapper[4631]: I1128 13:36:06.178343 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-7jgdf" Nov 28 13:36:06 crc kubenswrapper[4631]: I1128 13:36:06.192678 4631 generic.go:334] "Generic (PLEG): container finished" podID="9594c34e-0cdb-4e81-9e4a-1f542741e05b" containerID="265d530085b2141c8bbc6e5a0188ade7a1d462da994668c7ff64e5e292000555" exitCode=0 Nov 28 13:36:06 crc kubenswrapper[4631]: I1128 13:36:06.192739 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fd796d7df-fhmpd" Nov 28 13:36:06 crc kubenswrapper[4631]: I1128 13:36:06.192826 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fd796d7df-fhmpd" event={"ID":"9594c34e-0cdb-4e81-9e4a-1f542741e05b","Type":"ContainerDied","Data":"265d530085b2141c8bbc6e5a0188ade7a1d462da994668c7ff64e5e292000555"} Nov 28 13:36:06 crc kubenswrapper[4631]: I1128 13:36:06.192856 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fd796d7df-fhmpd" event={"ID":"9594c34e-0cdb-4e81-9e4a-1f542741e05b","Type":"ContainerDied","Data":"73ce92a6c4b73b8d15ebd733b8712eeb903bdde9bf5914245dd233724ade1795"} Nov 28 13:36:06 crc kubenswrapper[4631]: I1128 13:36:06.192872 4631 scope.go:117] "RemoveContainer" containerID="265d530085b2141c8bbc6e5a0188ade7a1d462da994668c7ff64e5e292000555" Nov 28 13:36:06 crc kubenswrapper[4631]: I1128 13:36:06.212404 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-addf-account-create-update-px8w7"] Nov 28 13:36:06 crc kubenswrapper[4631]: I1128 13:36:06.213700 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-addf-account-create-update-px8w7" Nov 28 13:36:06 crc kubenswrapper[4631]: I1128 13:36:06.219116 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Nov 28 13:36:06 crc kubenswrapper[4631]: I1128 13:36:06.233634 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-addf-account-create-update-px8w7"] Nov 28 13:36:06 crc kubenswrapper[4631]: I1128 13:36:06.236944 4631 scope.go:117] "RemoveContainer" containerID="b02c8505e2ae8141c62f5628c67f3e83e9662cf7e379ad1640849c61ad28e2c1" Nov 28 13:36:06 crc kubenswrapper[4631]: I1128 13:36:06.253463 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-7jgdf" podStartSLOduration=4.932946697 podStartE2EDuration="53.253422039s" podCreationTimestamp="2025-11-28 13:35:13 +0000 UTC" firstStartedPulling="2025-11-28 13:35:16.71936219 +0000 UTC m=+873.526665534" lastFinishedPulling="2025-11-28 13:36:05.039837532 +0000 UTC m=+921.847140876" observedRunningTime="2025-11-28 13:36:06.24503746 +0000 UTC m=+923.052340804" watchObservedRunningTime="2025-11-28 13:36:06.253422039 +0000 UTC m=+923.060725383" Nov 28 13:36:06 crc kubenswrapper[4631]: I1128 13:36:06.283694 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hp6dl\" (UniqueName: \"kubernetes.io/projected/07e8a124-1828-4a61-9b39-e1c8cf4cb2cc-kube-api-access-hp6dl\") pod \"placement-db-create-68dgq\" (UID: \"07e8a124-1828-4a61-9b39-e1c8cf4cb2cc\") " pod="openstack/placement-db-create-68dgq" Nov 28 13:36:06 crc kubenswrapper[4631]: I1128 13:36:06.283846 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/07e8a124-1828-4a61-9b39-e1c8cf4cb2cc-operator-scripts\") pod \"placement-db-create-68dgq\" (UID: \"07e8a124-1828-4a61-9b39-e1c8cf4cb2cc\") " pod="openstack/placement-db-create-68dgq" Nov 28 13:36:06 crc kubenswrapper[4631]: I1128 13:36:06.318674 4631 scope.go:117] "RemoveContainer" containerID="265d530085b2141c8bbc6e5a0188ade7a1d462da994668c7ff64e5e292000555" Nov 28 13:36:06 crc kubenswrapper[4631]: E1128 13:36:06.322107 4631 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"265d530085b2141c8bbc6e5a0188ade7a1d462da994668c7ff64e5e292000555\": container with ID starting with 265d530085b2141c8bbc6e5a0188ade7a1d462da994668c7ff64e5e292000555 not found: ID does not exist" containerID="265d530085b2141c8bbc6e5a0188ade7a1d462da994668c7ff64e5e292000555" Nov 28 13:36:06 crc kubenswrapper[4631]: I1128 13:36:06.322168 4631 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"265d530085b2141c8bbc6e5a0188ade7a1d462da994668c7ff64e5e292000555"} err="failed to get container status \"265d530085b2141c8bbc6e5a0188ade7a1d462da994668c7ff64e5e292000555\": rpc error: code = NotFound desc = could not find container \"265d530085b2141c8bbc6e5a0188ade7a1d462da994668c7ff64e5e292000555\": container with ID starting with 265d530085b2141c8bbc6e5a0188ade7a1d462da994668c7ff64e5e292000555 not found: ID does not exist" Nov 28 13:36:06 crc kubenswrapper[4631]: I1128 13:36:06.322203 4631 scope.go:117] "RemoveContainer" containerID="b02c8505e2ae8141c62f5628c67f3e83e9662cf7e379ad1640849c61ad28e2c1" Nov 28 13:36:06 crc kubenswrapper[4631]: E1128 13:36:06.323006 4631 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b02c8505e2ae8141c62f5628c67f3e83e9662cf7e379ad1640849c61ad28e2c1\": container with ID starting with b02c8505e2ae8141c62f5628c67f3e83e9662cf7e379ad1640849c61ad28e2c1 not found: ID does not exist" containerID="b02c8505e2ae8141c62f5628c67f3e83e9662cf7e379ad1640849c61ad28e2c1" Nov 28 13:36:06 crc kubenswrapper[4631]: I1128 13:36:06.323056 4631 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b02c8505e2ae8141c62f5628c67f3e83e9662cf7e379ad1640849c61ad28e2c1"} err="failed to get container status \"b02c8505e2ae8141c62f5628c67f3e83e9662cf7e379ad1640849c61ad28e2c1\": rpc error: code = NotFound desc = could not find container \"b02c8505e2ae8141c62f5628c67f3e83e9662cf7e379ad1640849c61ad28e2c1\": container with ID starting with b02c8505e2ae8141c62f5628c67f3e83e9662cf7e379ad1640849c61ad28e2c1 not found: ID does not exist" Nov 28 13:36:06 crc kubenswrapper[4631]: I1128 13:36:06.345655 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-fhmpd"] Nov 28 13:36:06 crc kubenswrapper[4631]: I1128 13:36:06.354202 4631 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-fhmpd"] Nov 28 13:36:06 crc kubenswrapper[4631]: I1128 13:36:06.374530 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Nov 28 13:36:06 crc kubenswrapper[4631]: I1128 13:36:06.385753 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hp6dl\" (UniqueName: \"kubernetes.io/projected/07e8a124-1828-4a61-9b39-e1c8cf4cb2cc-kube-api-access-hp6dl\") pod \"placement-db-create-68dgq\" (UID: \"07e8a124-1828-4a61-9b39-e1c8cf4cb2cc\") " pod="openstack/placement-db-create-68dgq" Nov 28 13:36:06 crc kubenswrapper[4631]: I1128 13:36:06.385885 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/07e8a124-1828-4a61-9b39-e1c8cf4cb2cc-operator-scripts\") pod \"placement-db-create-68dgq\" (UID: \"07e8a124-1828-4a61-9b39-e1c8cf4cb2cc\") " pod="openstack/placement-db-create-68dgq" Nov 28 13:36:06 crc kubenswrapper[4631]: I1128 13:36:06.385954 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/122f3637-af63-467a-9e98-6b5f48570d9d-operator-scripts\") pod \"placement-addf-account-create-update-px8w7\" (UID: \"122f3637-af63-467a-9e98-6b5f48570d9d\") " pod="openstack/placement-addf-account-create-update-px8w7" Nov 28 13:36:06 crc kubenswrapper[4631]: I1128 13:36:06.386141 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d6rtt\" (UniqueName: \"kubernetes.io/projected/122f3637-af63-467a-9e98-6b5f48570d9d-kube-api-access-d6rtt\") pod \"placement-addf-account-create-update-px8w7\" (UID: \"122f3637-af63-467a-9e98-6b5f48570d9d\") " pod="openstack/placement-addf-account-create-update-px8w7" Nov 28 13:36:06 crc kubenswrapper[4631]: I1128 13:36:06.387614 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/07e8a124-1828-4a61-9b39-e1c8cf4cb2cc-operator-scripts\") pod \"placement-db-create-68dgq\" (UID: \"07e8a124-1828-4a61-9b39-e1c8cf4cb2cc\") " pod="openstack/placement-db-create-68dgq" Nov 28 13:36:06 crc kubenswrapper[4631]: I1128 13:36:06.409377 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hp6dl\" (UniqueName: \"kubernetes.io/projected/07e8a124-1828-4a61-9b39-e1c8cf4cb2cc-kube-api-access-hp6dl\") pod \"placement-db-create-68dgq\" (UID: \"07e8a124-1828-4a61-9b39-e1c8cf4cb2cc\") " pod="openstack/placement-db-create-68dgq" Nov 28 13:36:06 crc kubenswrapper[4631]: I1128 13:36:06.467393 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-68dgq" Nov 28 13:36:06 crc kubenswrapper[4631]: I1128 13:36:06.489902 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d6rtt\" (UniqueName: \"kubernetes.io/projected/122f3637-af63-467a-9e98-6b5f48570d9d-kube-api-access-d6rtt\") pod \"placement-addf-account-create-update-px8w7\" (UID: \"122f3637-af63-467a-9e98-6b5f48570d9d\") " pod="openstack/placement-addf-account-create-update-px8w7" Nov 28 13:36:06 crc kubenswrapper[4631]: I1128 13:36:06.490130 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/122f3637-af63-467a-9e98-6b5f48570d9d-operator-scripts\") pod \"placement-addf-account-create-update-px8w7\" (UID: \"122f3637-af63-467a-9e98-6b5f48570d9d\") " pod="openstack/placement-addf-account-create-update-px8w7" Nov 28 13:36:06 crc kubenswrapper[4631]: I1128 13:36:06.495842 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/122f3637-af63-467a-9e98-6b5f48570d9d-operator-scripts\") pod \"placement-addf-account-create-update-px8w7\" (UID: \"122f3637-af63-467a-9e98-6b5f48570d9d\") " pod="openstack/placement-addf-account-create-update-px8w7" Nov 28 13:36:06 crc kubenswrapper[4631]: I1128 13:36:06.514322 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-d5ea-account-create-update-cw75f"] Nov 28 13:36:06 crc kubenswrapper[4631]: I1128 13:36:06.519674 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d6rtt\" (UniqueName: \"kubernetes.io/projected/122f3637-af63-467a-9e98-6b5f48570d9d-kube-api-access-d6rtt\") pod \"placement-addf-account-create-update-px8w7\" (UID: \"122f3637-af63-467a-9e98-6b5f48570d9d\") " pod="openstack/placement-addf-account-create-update-px8w7" Nov 28 13:36:06 crc kubenswrapper[4631]: W1128 13:36:06.535647 4631 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddd84f393_e43d_45e8_8760_7e0643f7c30e.slice/crio-95d7044657a319b28f89eb1e563ed29670c134d06016edf80bf44c7d58c995cc WatchSource:0}: Error finding container 95d7044657a319b28f89eb1e563ed29670c134d06016edf80bf44c7d58c995cc: Status 404 returned error can't find the container with id 95d7044657a319b28f89eb1e563ed29670c134d06016edf80bf44c7d58c995cc Nov 28 13:36:06 crc kubenswrapper[4631]: I1128 13:36:06.583368 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-addf-account-create-update-px8w7" Nov 28 13:36:06 crc kubenswrapper[4631]: I1128 13:36:06.783495 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-9m2n2"] Nov 28 13:36:07 crc kubenswrapper[4631]: I1128 13:36:07.053343 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-68dgq"] Nov 28 13:36:07 crc kubenswrapper[4631]: W1128 13:36:07.090225 4631 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod07e8a124_1828_4a61_9b39_e1c8cf4cb2cc.slice/crio-a84b4b018ef8064337c3f67ed9ed5334130687d1355bcee7f4ae9833692f1c6a WatchSource:0}: Error finding container a84b4b018ef8064337c3f67ed9ed5334130687d1355bcee7f4ae9833692f1c6a: Status 404 returned error can't find the container with id a84b4b018ef8064337c3f67ed9ed5334130687d1355bcee7f4ae9833692f1c6a Nov 28 13:36:07 crc kubenswrapper[4631]: I1128 13:36:07.164680 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-addf-account-create-update-px8w7"] Nov 28 13:36:07 crc kubenswrapper[4631]: I1128 13:36:07.209799 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-d5ea-account-create-update-cw75f" event={"ID":"dd84f393-e43d-45e8-8760-7e0643f7c30e","Type":"ContainerStarted","Data":"a1646b6221d6723f5f5f2cc3c750881618fc6f98c65326ef8d092470b7a48482"} Nov 28 13:36:07 crc kubenswrapper[4631]: I1128 13:36:07.209842 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-d5ea-account-create-update-cw75f" event={"ID":"dd84f393-e43d-45e8-8760-7e0643f7c30e","Type":"ContainerStarted","Data":"95d7044657a319b28f89eb1e563ed29670c134d06016edf80bf44c7d58c995cc"} Nov 28 13:36:07 crc kubenswrapper[4631]: I1128 13:36:07.213825 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-68dgq" event={"ID":"07e8a124-1828-4a61-9b39-e1c8cf4cb2cc","Type":"ContainerStarted","Data":"a84b4b018ef8064337c3f67ed9ed5334130687d1355bcee7f4ae9833692f1c6a"} Nov 28 13:36:07 crc kubenswrapper[4631]: I1128 13:36:07.214999 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-9m2n2" event={"ID":"3edb21b5-935d-4b42-bc88-d590b7fb0e3d","Type":"ContainerStarted","Data":"1cac3f2d3153a738c4382184487f445a93cf68bcfaf941c66086e3a38cf0dcf8"} Nov 28 13:36:07 crc kubenswrapper[4631]: I1128 13:36:07.215020 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-9m2n2" event={"ID":"3edb21b5-935d-4b42-bc88-d590b7fb0e3d","Type":"ContainerStarted","Data":"d324ef1a2c4bf53f291513227a62d874d10be702d9fb44157c872abc2859911a"} Nov 28 13:36:07 crc kubenswrapper[4631]: I1128 13:36:07.217669 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-lskqw" event={"ID":"7fc0ee4e-435d-4297-94eb-9e8c6fa90004","Type":"ContainerStarted","Data":"41733c7ff2d1d92e10501d54d0dbbedaf309c5aa3515bc42a4fba61e1c6ae340"} Nov 28 13:36:07 crc kubenswrapper[4631]: W1128 13:36:07.218710 4631 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod122f3637_af63_467a_9e98_6b5f48570d9d.slice/crio-ccc890bcfbd2b3f3dae5658bcf1316cc046a20dbd2917ae57a763d27445b7e31 WatchSource:0}: Error finding container ccc890bcfbd2b3f3dae5658bcf1316cc046a20dbd2917ae57a763d27445b7e31: Status 404 returned error can't find the container with id ccc890bcfbd2b3f3dae5658bcf1316cc046a20dbd2917ae57a763d27445b7e31 Nov 28 13:36:07 crc kubenswrapper[4631]: I1128 13:36:07.251256 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-create-9m2n2" podStartSLOduration=2.251237305 podStartE2EDuration="2.251237305s" podCreationTimestamp="2025-11-28 13:36:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 13:36:07.250179529 +0000 UTC m=+924.057482873" watchObservedRunningTime="2025-11-28 13:36:07.251237305 +0000 UTC m=+924.058540649" Nov 28 13:36:07 crc kubenswrapper[4631]: I1128 13:36:07.252550 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-d5ea-account-create-update-cw75f" podStartSLOduration=2.252543437 podStartE2EDuration="2.252543437s" podCreationTimestamp="2025-11-28 13:36:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 13:36:07.230449807 +0000 UTC m=+924.037753151" watchObservedRunningTime="2025-11-28 13:36:07.252543437 +0000 UTC m=+924.059846781" Nov 28 13:36:07 crc kubenswrapper[4631]: I1128 13:36:07.526157 4631 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9594c34e-0cdb-4e81-9e4a-1f542741e05b" path="/var/lib/kubelet/pods/9594c34e-0cdb-4e81-9e4a-1f542741e05b/volumes" Nov 28 13:36:08 crc kubenswrapper[4631]: I1128 13:36:08.230197 4631 generic.go:334] "Generic (PLEG): container finished" podID="dd84f393-e43d-45e8-8760-7e0643f7c30e" containerID="a1646b6221d6723f5f5f2cc3c750881618fc6f98c65326ef8d092470b7a48482" exitCode=0 Nov 28 13:36:08 crc kubenswrapper[4631]: I1128 13:36:08.230303 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-d5ea-account-create-update-cw75f" event={"ID":"dd84f393-e43d-45e8-8760-7e0643f7c30e","Type":"ContainerDied","Data":"a1646b6221d6723f5f5f2cc3c750881618fc6f98c65326ef8d092470b7a48482"} Nov 28 13:36:08 crc kubenswrapper[4631]: I1128 13:36:08.237597 4631 generic.go:334] "Generic (PLEG): container finished" podID="07e8a124-1828-4a61-9b39-e1c8cf4cb2cc" containerID="f1cab3e61e6936a06f7682175f6c6e2f4d0dc741337beca101fa628f19cce22d" exitCode=0 Nov 28 13:36:08 crc kubenswrapper[4631]: I1128 13:36:08.237694 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-68dgq" event={"ID":"07e8a124-1828-4a61-9b39-e1c8cf4cb2cc","Type":"ContainerDied","Data":"f1cab3e61e6936a06f7682175f6c6e2f4d0dc741337beca101fa628f19cce22d"} Nov 28 13:36:08 crc kubenswrapper[4631]: I1128 13:36:08.240842 4631 generic.go:334] "Generic (PLEG): container finished" podID="122f3637-af63-467a-9e98-6b5f48570d9d" containerID="b55a0c4fae79fdf4bb59d9a51cb2dc17f21eefaee959939aee4a11adfc266b4e" exitCode=0 Nov 28 13:36:08 crc kubenswrapper[4631]: I1128 13:36:08.241032 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-addf-account-create-update-px8w7" event={"ID":"122f3637-af63-467a-9e98-6b5f48570d9d","Type":"ContainerDied","Data":"b55a0c4fae79fdf4bb59d9a51cb2dc17f21eefaee959939aee4a11adfc266b4e"} Nov 28 13:36:08 crc kubenswrapper[4631]: I1128 13:36:08.241203 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-addf-account-create-update-px8w7" event={"ID":"122f3637-af63-467a-9e98-6b5f48570d9d","Type":"ContainerStarted","Data":"ccc890bcfbd2b3f3dae5658bcf1316cc046a20dbd2917ae57a763d27445b7e31"} Nov 28 13:36:08 crc kubenswrapper[4631]: I1128 13:36:08.251224 4631 generic.go:334] "Generic (PLEG): container finished" podID="3edb21b5-935d-4b42-bc88-d590b7fb0e3d" containerID="1cac3f2d3153a738c4382184487f445a93cf68bcfaf941c66086e3a38cf0dcf8" exitCode=0 Nov 28 13:36:08 crc kubenswrapper[4631]: I1128 13:36:08.251410 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-9m2n2" event={"ID":"3edb21b5-935d-4b42-bc88-d590b7fb0e3d","Type":"ContainerDied","Data":"1cac3f2d3153a738c4382184487f445a93cf68bcfaf941c66086e3a38cf0dcf8"} Nov 28 13:36:08 crc kubenswrapper[4631]: I1128 13:36:08.257805 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-lskqw" event={"ID":"7fc0ee4e-435d-4297-94eb-9e8c6fa90004","Type":"ContainerStarted","Data":"ee4b2d4c89049d02b5d31d36cfb30d1a32c2592b82a52ad7151940b0623eb9cc"} Nov 28 13:36:08 crc kubenswrapper[4631]: I1128 13:36:08.257958 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-lskqw" Nov 28 13:36:08 crc kubenswrapper[4631]: I1128 13:36:08.258033 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-lskqw" Nov 28 13:36:08 crc kubenswrapper[4631]: I1128 13:36:08.341580 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-lskqw" podStartSLOduration=17.858176205 podStartE2EDuration="55.341551517s" podCreationTimestamp="2025-11-28 13:35:13 +0000 UTC" firstStartedPulling="2025-11-28 13:35:27.548948863 +0000 UTC m=+884.356252207" lastFinishedPulling="2025-11-28 13:36:05.032324165 +0000 UTC m=+921.839627519" observedRunningTime="2025-11-28 13:36:08.331106178 +0000 UTC m=+925.138409562" watchObservedRunningTime="2025-11-28 13:36:08.341551517 +0000 UTC m=+925.148854881" Nov 28 13:36:08 crc kubenswrapper[4631]: I1128 13:36:08.612457 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-698758b865-m2zrc"] Nov 28 13:36:08 crc kubenswrapper[4631]: I1128 13:36:08.614915 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-m2zrc" Nov 28 13:36:08 crc kubenswrapper[4631]: I1128 13:36:08.647775 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-698758b865-m2zrc"] Nov 28 13:36:08 crc kubenswrapper[4631]: I1128 13:36:08.765355 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t8w2f\" (UniqueName: \"kubernetes.io/projected/ed04b83d-c180-4bf2-b6b9-b112a536d9da-kube-api-access-t8w2f\") pod \"dnsmasq-dns-698758b865-m2zrc\" (UID: \"ed04b83d-c180-4bf2-b6b9-b112a536d9da\") " pod="openstack/dnsmasq-dns-698758b865-m2zrc" Nov 28 13:36:08 crc kubenswrapper[4631]: I1128 13:36:08.765417 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ed04b83d-c180-4bf2-b6b9-b112a536d9da-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-m2zrc\" (UID: \"ed04b83d-c180-4bf2-b6b9-b112a536d9da\") " pod="openstack/dnsmasq-dns-698758b865-m2zrc" Nov 28 13:36:08 crc kubenswrapper[4631]: I1128 13:36:08.765455 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ed04b83d-c180-4bf2-b6b9-b112a536d9da-dns-svc\") pod \"dnsmasq-dns-698758b865-m2zrc\" (UID: \"ed04b83d-c180-4bf2-b6b9-b112a536d9da\") " pod="openstack/dnsmasq-dns-698758b865-m2zrc" Nov 28 13:36:08 crc kubenswrapper[4631]: I1128 13:36:08.765476 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ed04b83d-c180-4bf2-b6b9-b112a536d9da-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-m2zrc\" (UID: \"ed04b83d-c180-4bf2-b6b9-b112a536d9da\") " pod="openstack/dnsmasq-dns-698758b865-m2zrc" Nov 28 13:36:08 crc kubenswrapper[4631]: I1128 13:36:08.765527 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ed04b83d-c180-4bf2-b6b9-b112a536d9da-config\") pod \"dnsmasq-dns-698758b865-m2zrc\" (UID: \"ed04b83d-c180-4bf2-b6b9-b112a536d9da\") " pod="openstack/dnsmasq-dns-698758b865-m2zrc" Nov 28 13:36:08 crc kubenswrapper[4631]: I1128 13:36:08.870011 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ed04b83d-c180-4bf2-b6b9-b112a536d9da-dns-svc\") pod \"dnsmasq-dns-698758b865-m2zrc\" (UID: \"ed04b83d-c180-4bf2-b6b9-b112a536d9da\") " pod="openstack/dnsmasq-dns-698758b865-m2zrc" Nov 28 13:36:08 crc kubenswrapper[4631]: I1128 13:36:08.870118 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ed04b83d-c180-4bf2-b6b9-b112a536d9da-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-m2zrc\" (UID: \"ed04b83d-c180-4bf2-b6b9-b112a536d9da\") " pod="openstack/dnsmasq-dns-698758b865-m2zrc" Nov 28 13:36:08 crc kubenswrapper[4631]: I1128 13:36:08.870209 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ed04b83d-c180-4bf2-b6b9-b112a536d9da-config\") pod \"dnsmasq-dns-698758b865-m2zrc\" (UID: \"ed04b83d-c180-4bf2-b6b9-b112a536d9da\") " pod="openstack/dnsmasq-dns-698758b865-m2zrc" Nov 28 13:36:08 crc kubenswrapper[4631]: I1128 13:36:08.870332 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t8w2f\" (UniqueName: \"kubernetes.io/projected/ed04b83d-c180-4bf2-b6b9-b112a536d9da-kube-api-access-t8w2f\") pod \"dnsmasq-dns-698758b865-m2zrc\" (UID: \"ed04b83d-c180-4bf2-b6b9-b112a536d9da\") " pod="openstack/dnsmasq-dns-698758b865-m2zrc" Nov 28 13:36:08 crc kubenswrapper[4631]: I1128 13:36:08.870374 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ed04b83d-c180-4bf2-b6b9-b112a536d9da-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-m2zrc\" (UID: \"ed04b83d-c180-4bf2-b6b9-b112a536d9da\") " pod="openstack/dnsmasq-dns-698758b865-m2zrc" Nov 28 13:36:08 crc kubenswrapper[4631]: I1128 13:36:08.870379 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ed04b83d-c180-4bf2-b6b9-b112a536d9da-dns-svc\") pod \"dnsmasq-dns-698758b865-m2zrc\" (UID: \"ed04b83d-c180-4bf2-b6b9-b112a536d9da\") " pod="openstack/dnsmasq-dns-698758b865-m2zrc" Nov 28 13:36:08 crc kubenswrapper[4631]: I1128 13:36:08.871010 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ed04b83d-c180-4bf2-b6b9-b112a536d9da-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-m2zrc\" (UID: \"ed04b83d-c180-4bf2-b6b9-b112a536d9da\") " pod="openstack/dnsmasq-dns-698758b865-m2zrc" Nov 28 13:36:08 crc kubenswrapper[4631]: I1128 13:36:08.871834 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ed04b83d-c180-4bf2-b6b9-b112a536d9da-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-m2zrc\" (UID: \"ed04b83d-c180-4bf2-b6b9-b112a536d9da\") " pod="openstack/dnsmasq-dns-698758b865-m2zrc" Nov 28 13:36:08 crc kubenswrapper[4631]: I1128 13:36:08.872771 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ed04b83d-c180-4bf2-b6b9-b112a536d9da-config\") pod \"dnsmasq-dns-698758b865-m2zrc\" (UID: \"ed04b83d-c180-4bf2-b6b9-b112a536d9da\") " pod="openstack/dnsmasq-dns-698758b865-m2zrc" Nov 28 13:36:08 crc kubenswrapper[4631]: I1128 13:36:08.903793 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t8w2f\" (UniqueName: \"kubernetes.io/projected/ed04b83d-c180-4bf2-b6b9-b112a536d9da-kube-api-access-t8w2f\") pod \"dnsmasq-dns-698758b865-m2zrc\" (UID: \"ed04b83d-c180-4bf2-b6b9-b112a536d9da\") " pod="openstack/dnsmasq-dns-698758b865-m2zrc" Nov 28 13:36:08 crc kubenswrapper[4631]: I1128 13:36:08.952261 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-m2zrc" Nov 28 13:36:09 crc kubenswrapper[4631]: W1128 13:36:09.492183 4631 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poded04b83d_c180_4bf2_b6b9_b112a536d9da.slice/crio-6c9354e875a4d160ec866ea5f7f4933f80b648e00763d7c48e9e07ea5fac8369 WatchSource:0}: Error finding container 6c9354e875a4d160ec866ea5f7f4933f80b648e00763d7c48e9e07ea5fac8369: Status 404 returned error can't find the container with id 6c9354e875a4d160ec866ea5f7f4933f80b648e00763d7c48e9e07ea5fac8369 Nov 28 13:36:09 crc kubenswrapper[4631]: I1128 13:36:09.500028 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-698758b865-m2zrc"] Nov 28 13:36:09 crc kubenswrapper[4631]: I1128 13:36:09.569689 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-68dgq" Nov 28 13:36:09 crc kubenswrapper[4631]: I1128 13:36:09.713281 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hp6dl\" (UniqueName: \"kubernetes.io/projected/07e8a124-1828-4a61-9b39-e1c8cf4cb2cc-kube-api-access-hp6dl\") pod \"07e8a124-1828-4a61-9b39-e1c8cf4cb2cc\" (UID: \"07e8a124-1828-4a61-9b39-e1c8cf4cb2cc\") " Nov 28 13:36:09 crc kubenswrapper[4631]: I1128 13:36:09.714242 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/07e8a124-1828-4a61-9b39-e1c8cf4cb2cc-operator-scripts\") pod \"07e8a124-1828-4a61-9b39-e1c8cf4cb2cc\" (UID: \"07e8a124-1828-4a61-9b39-e1c8cf4cb2cc\") " Nov 28 13:36:09 crc kubenswrapper[4631]: I1128 13:36:09.715315 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/07e8a124-1828-4a61-9b39-e1c8cf4cb2cc-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "07e8a124-1828-4a61-9b39-e1c8cf4cb2cc" (UID: "07e8a124-1828-4a61-9b39-e1c8cf4cb2cc"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 13:36:09 crc kubenswrapper[4631]: I1128 13:36:09.724565 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/07e8a124-1828-4a61-9b39-e1c8cf4cb2cc-kube-api-access-hp6dl" (OuterVolumeSpecName: "kube-api-access-hp6dl") pod "07e8a124-1828-4a61-9b39-e1c8cf4cb2cc" (UID: "07e8a124-1828-4a61-9b39-e1c8cf4cb2cc"). InnerVolumeSpecName "kube-api-access-hp6dl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:36:09 crc kubenswrapper[4631]: I1128 13:36:09.780602 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-storage-0"] Nov 28 13:36:09 crc kubenswrapper[4631]: E1128 13:36:09.781078 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="07e8a124-1828-4a61-9b39-e1c8cf4cb2cc" containerName="mariadb-database-create" Nov 28 13:36:09 crc kubenswrapper[4631]: I1128 13:36:09.781094 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="07e8a124-1828-4a61-9b39-e1c8cf4cb2cc" containerName="mariadb-database-create" Nov 28 13:36:09 crc kubenswrapper[4631]: I1128 13:36:09.781304 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="07e8a124-1828-4a61-9b39-e1c8cf4cb2cc" containerName="mariadb-database-create" Nov 28 13:36:09 crc kubenswrapper[4631]: I1128 13:36:09.795073 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Nov 28 13:36:09 crc kubenswrapper[4631]: I1128 13:36:09.806076 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-swift-dockercfg-tzkn5" Nov 28 13:36:09 crc kubenswrapper[4631]: I1128 13:36:09.806223 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-conf" Nov 28 13:36:09 crc kubenswrapper[4631]: I1128 13:36:09.806378 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-storage-config-data" Nov 28 13:36:09 crc kubenswrapper[4631]: I1128 13:36:09.806509 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-files" Nov 28 13:36:09 crc kubenswrapper[4631]: I1128 13:36:09.816259 4631 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/07e8a124-1828-4a61-9b39-e1c8cf4cb2cc-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 13:36:09 crc kubenswrapper[4631]: I1128 13:36:09.816301 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hp6dl\" (UniqueName: \"kubernetes.io/projected/07e8a124-1828-4a61-9b39-e1c8cf4cb2cc-kube-api-access-hp6dl\") on node \"crc\" DevicePath \"\"" Nov 28 13:36:09 crc kubenswrapper[4631]: I1128 13:36:09.817842 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Nov 28 13:36:09 crc kubenswrapper[4631]: I1128 13:36:09.902298 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-d5ea-account-create-update-cw75f" Nov 28 13:36:09 crc kubenswrapper[4631]: I1128 13:36:09.918947 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/c3e1f6fb-1845-4a39-8eca-1262da1ff281-cache\") pod \"swift-storage-0\" (UID: \"c3e1f6fb-1845-4a39-8eca-1262da1ff281\") " pod="openstack/swift-storage-0" Nov 28 13:36:09 crc kubenswrapper[4631]: I1128 13:36:09.919010 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/c3e1f6fb-1845-4a39-8eca-1262da1ff281-lock\") pod \"swift-storage-0\" (UID: \"c3e1f6fb-1845-4a39-8eca-1262da1ff281\") " pod="openstack/swift-storage-0" Nov 28 13:36:09 crc kubenswrapper[4631]: I1128 13:36:09.919222 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/c3e1f6fb-1845-4a39-8eca-1262da1ff281-etc-swift\") pod \"swift-storage-0\" (UID: \"c3e1f6fb-1845-4a39-8eca-1262da1ff281\") " pod="openstack/swift-storage-0" Nov 28 13:36:09 crc kubenswrapper[4631]: I1128 13:36:09.919352 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rj2c9\" (UniqueName: \"kubernetes.io/projected/c3e1f6fb-1845-4a39-8eca-1262da1ff281-kube-api-access-rj2c9\") pod \"swift-storage-0\" (UID: \"c3e1f6fb-1845-4a39-8eca-1262da1ff281\") " pod="openstack/swift-storage-0" Nov 28 13:36:09 crc kubenswrapper[4631]: I1128 13:36:09.919610 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"swift-storage-0\" (UID: \"c3e1f6fb-1845-4a39-8eca-1262da1ff281\") " pod="openstack/swift-storage-0" Nov 28 13:36:09 crc kubenswrapper[4631]: I1128 13:36:09.958791 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-9m2n2" Nov 28 13:36:09 crc kubenswrapper[4631]: I1128 13:36:09.967005 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-addf-account-create-update-px8w7" Nov 28 13:36:10 crc kubenswrapper[4631]: I1128 13:36:10.020591 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dd84f393-e43d-45e8-8760-7e0643f7c30e-operator-scripts\") pod \"dd84f393-e43d-45e8-8760-7e0643f7c30e\" (UID: \"dd84f393-e43d-45e8-8760-7e0643f7c30e\") " Nov 28 13:36:10 crc kubenswrapper[4631]: I1128 13:36:10.020649 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9fmjt\" (UniqueName: \"kubernetes.io/projected/dd84f393-e43d-45e8-8760-7e0643f7c30e-kube-api-access-9fmjt\") pod \"dd84f393-e43d-45e8-8760-7e0643f7c30e\" (UID: \"dd84f393-e43d-45e8-8760-7e0643f7c30e\") " Nov 28 13:36:10 crc kubenswrapper[4631]: I1128 13:36:10.021347 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dd84f393-e43d-45e8-8760-7e0643f7c30e-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "dd84f393-e43d-45e8-8760-7e0643f7c30e" (UID: "dd84f393-e43d-45e8-8760-7e0643f7c30e"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 13:36:10 crc kubenswrapper[4631]: I1128 13:36:10.021753 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g9jlq\" (UniqueName: \"kubernetes.io/projected/3edb21b5-935d-4b42-bc88-d590b7fb0e3d-kube-api-access-g9jlq\") pod \"3edb21b5-935d-4b42-bc88-d590b7fb0e3d\" (UID: \"3edb21b5-935d-4b42-bc88-d590b7fb0e3d\") " Nov 28 13:36:10 crc kubenswrapper[4631]: I1128 13:36:10.021809 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3edb21b5-935d-4b42-bc88-d590b7fb0e3d-operator-scripts\") pod \"3edb21b5-935d-4b42-bc88-d590b7fb0e3d\" (UID: \"3edb21b5-935d-4b42-bc88-d590b7fb0e3d\") " Nov 28 13:36:10 crc kubenswrapper[4631]: I1128 13:36:10.022119 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"swift-storage-0\" (UID: \"c3e1f6fb-1845-4a39-8eca-1262da1ff281\") " pod="openstack/swift-storage-0" Nov 28 13:36:10 crc kubenswrapper[4631]: I1128 13:36:10.022165 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/c3e1f6fb-1845-4a39-8eca-1262da1ff281-cache\") pod \"swift-storage-0\" (UID: \"c3e1f6fb-1845-4a39-8eca-1262da1ff281\") " pod="openstack/swift-storage-0" Nov 28 13:36:10 crc kubenswrapper[4631]: I1128 13:36:10.022187 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/c3e1f6fb-1845-4a39-8eca-1262da1ff281-lock\") pod \"swift-storage-0\" (UID: \"c3e1f6fb-1845-4a39-8eca-1262da1ff281\") " pod="openstack/swift-storage-0" Nov 28 13:36:10 crc kubenswrapper[4631]: I1128 13:36:10.022227 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/c3e1f6fb-1845-4a39-8eca-1262da1ff281-etc-swift\") pod \"swift-storage-0\" (UID: \"c3e1f6fb-1845-4a39-8eca-1262da1ff281\") " pod="openstack/swift-storage-0" Nov 28 13:36:10 crc kubenswrapper[4631]: I1128 13:36:10.022268 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rj2c9\" (UniqueName: \"kubernetes.io/projected/c3e1f6fb-1845-4a39-8eca-1262da1ff281-kube-api-access-rj2c9\") pod \"swift-storage-0\" (UID: \"c3e1f6fb-1845-4a39-8eca-1262da1ff281\") " pod="openstack/swift-storage-0" Nov 28 13:36:10 crc kubenswrapper[4631]: I1128 13:36:10.022402 4631 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dd84f393-e43d-45e8-8760-7e0643f7c30e-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 13:36:10 crc kubenswrapper[4631]: I1128 13:36:10.023617 4631 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"swift-storage-0\" (UID: \"c3e1f6fb-1845-4a39-8eca-1262da1ff281\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/swift-storage-0" Nov 28 13:36:10 crc kubenswrapper[4631]: I1128 13:36:10.023815 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/c3e1f6fb-1845-4a39-8eca-1262da1ff281-cache\") pod \"swift-storage-0\" (UID: \"c3e1f6fb-1845-4a39-8eca-1262da1ff281\") " pod="openstack/swift-storage-0" Nov 28 13:36:10 crc kubenswrapper[4631]: I1128 13:36:10.023823 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/c3e1f6fb-1845-4a39-8eca-1262da1ff281-lock\") pod \"swift-storage-0\" (UID: \"c3e1f6fb-1845-4a39-8eca-1262da1ff281\") " pod="openstack/swift-storage-0" Nov 28 13:36:10 crc kubenswrapper[4631]: E1128 13:36:10.023948 4631 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Nov 28 13:36:10 crc kubenswrapper[4631]: E1128 13:36:10.023966 4631 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Nov 28 13:36:10 crc kubenswrapper[4631]: E1128 13:36:10.024012 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/c3e1f6fb-1845-4a39-8eca-1262da1ff281-etc-swift podName:c3e1f6fb-1845-4a39-8eca-1262da1ff281 nodeName:}" failed. No retries permitted until 2025-11-28 13:36:10.523995833 +0000 UTC m=+927.331299177 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/c3e1f6fb-1845-4a39-8eca-1262da1ff281-etc-swift") pod "swift-storage-0" (UID: "c3e1f6fb-1845-4a39-8eca-1262da1ff281") : configmap "swift-ring-files" not found Nov 28 13:36:10 crc kubenswrapper[4631]: I1128 13:36:10.024527 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3edb21b5-935d-4b42-bc88-d590b7fb0e3d-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "3edb21b5-935d-4b42-bc88-d590b7fb0e3d" (UID: "3edb21b5-935d-4b42-bc88-d590b7fb0e3d"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 13:36:10 crc kubenswrapper[4631]: I1128 13:36:10.034639 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dd84f393-e43d-45e8-8760-7e0643f7c30e-kube-api-access-9fmjt" (OuterVolumeSpecName: "kube-api-access-9fmjt") pod "dd84f393-e43d-45e8-8760-7e0643f7c30e" (UID: "dd84f393-e43d-45e8-8760-7e0643f7c30e"). InnerVolumeSpecName "kube-api-access-9fmjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:36:10 crc kubenswrapper[4631]: I1128 13:36:10.034684 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3edb21b5-935d-4b42-bc88-d590b7fb0e3d-kube-api-access-g9jlq" (OuterVolumeSpecName: "kube-api-access-g9jlq") pod "3edb21b5-935d-4b42-bc88-d590b7fb0e3d" (UID: "3edb21b5-935d-4b42-bc88-d590b7fb0e3d"). InnerVolumeSpecName "kube-api-access-g9jlq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:36:10 crc kubenswrapper[4631]: I1128 13:36:10.040577 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rj2c9\" (UniqueName: \"kubernetes.io/projected/c3e1f6fb-1845-4a39-8eca-1262da1ff281-kube-api-access-rj2c9\") pod \"swift-storage-0\" (UID: \"c3e1f6fb-1845-4a39-8eca-1262da1ff281\") " pod="openstack/swift-storage-0" Nov 28 13:36:10 crc kubenswrapper[4631]: I1128 13:36:10.044886 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"swift-storage-0\" (UID: \"c3e1f6fb-1845-4a39-8eca-1262da1ff281\") " pod="openstack/swift-storage-0" Nov 28 13:36:10 crc kubenswrapper[4631]: I1128 13:36:10.123412 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/122f3637-af63-467a-9e98-6b5f48570d9d-operator-scripts\") pod \"122f3637-af63-467a-9e98-6b5f48570d9d\" (UID: \"122f3637-af63-467a-9e98-6b5f48570d9d\") " Nov 28 13:36:10 crc kubenswrapper[4631]: I1128 13:36:10.123500 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6rtt\" (UniqueName: \"kubernetes.io/projected/122f3637-af63-467a-9e98-6b5f48570d9d-kube-api-access-d6rtt\") pod \"122f3637-af63-467a-9e98-6b5f48570d9d\" (UID: \"122f3637-af63-467a-9e98-6b5f48570d9d\") " Nov 28 13:36:10 crc kubenswrapper[4631]: I1128 13:36:10.123839 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g9jlq\" (UniqueName: \"kubernetes.io/projected/3edb21b5-935d-4b42-bc88-d590b7fb0e3d-kube-api-access-g9jlq\") on node \"crc\" DevicePath \"\"" Nov 28 13:36:10 crc kubenswrapper[4631]: I1128 13:36:10.123856 4631 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3edb21b5-935d-4b42-bc88-d590b7fb0e3d-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 13:36:10 crc kubenswrapper[4631]: I1128 13:36:10.123870 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9fmjt\" (UniqueName: \"kubernetes.io/projected/dd84f393-e43d-45e8-8760-7e0643f7c30e-kube-api-access-9fmjt\") on node \"crc\" DevicePath \"\"" Nov 28 13:36:10 crc kubenswrapper[4631]: I1128 13:36:10.123958 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/122f3637-af63-467a-9e98-6b5f48570d9d-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "122f3637-af63-467a-9e98-6b5f48570d9d" (UID: "122f3637-af63-467a-9e98-6b5f48570d9d"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 13:36:10 crc kubenswrapper[4631]: I1128 13:36:10.127456 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/122f3637-af63-467a-9e98-6b5f48570d9d-kube-api-access-d6rtt" (OuterVolumeSpecName: "kube-api-access-d6rtt") pod "122f3637-af63-467a-9e98-6b5f48570d9d" (UID: "122f3637-af63-467a-9e98-6b5f48570d9d"). InnerVolumeSpecName "kube-api-access-d6rtt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:36:10 crc kubenswrapper[4631]: I1128 13:36:10.226234 4631 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/122f3637-af63-467a-9e98-6b5f48570d9d-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 13:36:10 crc kubenswrapper[4631]: I1128 13:36:10.226809 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6rtt\" (UniqueName: \"kubernetes.io/projected/122f3637-af63-467a-9e98-6b5f48570d9d-kube-api-access-d6rtt\") on node \"crc\" DevicePath \"\"" Nov 28 13:36:10 crc kubenswrapper[4631]: I1128 13:36:10.279141 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-d5ea-account-create-update-cw75f" event={"ID":"dd84f393-e43d-45e8-8760-7e0643f7c30e","Type":"ContainerDied","Data":"95d7044657a319b28f89eb1e563ed29670c134d06016edf80bf44c7d58c995cc"} Nov 28 13:36:10 crc kubenswrapper[4631]: I1128 13:36:10.279189 4631 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="95d7044657a319b28f89eb1e563ed29670c134d06016edf80bf44c7d58c995cc" Nov 28 13:36:10 crc kubenswrapper[4631]: I1128 13:36:10.279192 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-d5ea-account-create-update-cw75f" Nov 28 13:36:10 crc kubenswrapper[4631]: I1128 13:36:10.281500 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-68dgq" Nov 28 13:36:10 crc kubenswrapper[4631]: I1128 13:36:10.281630 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-68dgq" event={"ID":"07e8a124-1828-4a61-9b39-e1c8cf4cb2cc","Type":"ContainerDied","Data":"a84b4b018ef8064337c3f67ed9ed5334130687d1355bcee7f4ae9833692f1c6a"} Nov 28 13:36:10 crc kubenswrapper[4631]: I1128 13:36:10.281726 4631 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a84b4b018ef8064337c3f67ed9ed5334130687d1355bcee7f4ae9833692f1c6a" Nov 28 13:36:10 crc kubenswrapper[4631]: I1128 13:36:10.283437 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-addf-account-create-update-px8w7" Nov 28 13:36:10 crc kubenswrapper[4631]: I1128 13:36:10.283567 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-addf-account-create-update-px8w7" event={"ID":"122f3637-af63-467a-9e98-6b5f48570d9d","Type":"ContainerDied","Data":"ccc890bcfbd2b3f3dae5658bcf1316cc046a20dbd2917ae57a763d27445b7e31"} Nov 28 13:36:10 crc kubenswrapper[4631]: I1128 13:36:10.283649 4631 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ccc890bcfbd2b3f3dae5658bcf1316cc046a20dbd2917ae57a763d27445b7e31" Nov 28 13:36:10 crc kubenswrapper[4631]: I1128 13:36:10.285486 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-9m2n2" event={"ID":"3edb21b5-935d-4b42-bc88-d590b7fb0e3d","Type":"ContainerDied","Data":"d324ef1a2c4bf53f291513227a62d874d10be702d9fb44157c872abc2859911a"} Nov 28 13:36:10 crc kubenswrapper[4631]: I1128 13:36:10.285526 4631 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d324ef1a2c4bf53f291513227a62d874d10be702d9fb44157c872abc2859911a" Nov 28 13:36:10 crc kubenswrapper[4631]: I1128 13:36:10.285636 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-9m2n2" Nov 28 13:36:10 crc kubenswrapper[4631]: I1128 13:36:10.296604 4631 generic.go:334] "Generic (PLEG): container finished" podID="ed04b83d-c180-4bf2-b6b9-b112a536d9da" containerID="eec174b25ab3c2391bf3c823680cdaa945d25dc0520c28fc082487d5af87f766" exitCode=0 Nov 28 13:36:10 crc kubenswrapper[4631]: I1128 13:36:10.296670 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-m2zrc" event={"ID":"ed04b83d-c180-4bf2-b6b9-b112a536d9da","Type":"ContainerDied","Data":"eec174b25ab3c2391bf3c823680cdaa945d25dc0520c28fc082487d5af87f766"} Nov 28 13:36:10 crc kubenswrapper[4631]: I1128 13:36:10.296947 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-m2zrc" event={"ID":"ed04b83d-c180-4bf2-b6b9-b112a536d9da","Type":"ContainerStarted","Data":"6c9354e875a4d160ec866ea5f7f4933f80b648e00763d7c48e9e07ea5fac8369"} Nov 28 13:36:10 crc kubenswrapper[4631]: I1128 13:36:10.532336 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/c3e1f6fb-1845-4a39-8eca-1262da1ff281-etc-swift\") pod \"swift-storage-0\" (UID: \"c3e1f6fb-1845-4a39-8eca-1262da1ff281\") " pod="openstack/swift-storage-0" Nov 28 13:36:10 crc kubenswrapper[4631]: E1128 13:36:10.532668 4631 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Nov 28 13:36:10 crc kubenswrapper[4631]: E1128 13:36:10.532722 4631 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Nov 28 13:36:10 crc kubenswrapper[4631]: E1128 13:36:10.532819 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/c3e1f6fb-1845-4a39-8eca-1262da1ff281-etc-swift podName:c3e1f6fb-1845-4a39-8eca-1262da1ff281 nodeName:}" failed. No retries permitted until 2025-11-28 13:36:11.532793367 +0000 UTC m=+928.340096721 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/c3e1f6fb-1845-4a39-8eca-1262da1ff281-etc-swift") pod "swift-storage-0" (UID: "c3e1f6fb-1845-4a39-8eca-1262da1ff281") : configmap "swift-ring-files" not found Nov 28 13:36:11 crc kubenswrapper[4631]: I1128 13:36:11.194157 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-gn4cj"] Nov 28 13:36:11 crc kubenswrapper[4631]: E1128 13:36:11.196401 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dd84f393-e43d-45e8-8760-7e0643f7c30e" containerName="mariadb-account-create-update" Nov 28 13:36:11 crc kubenswrapper[4631]: I1128 13:36:11.196440 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="dd84f393-e43d-45e8-8760-7e0643f7c30e" containerName="mariadb-account-create-update" Nov 28 13:36:11 crc kubenswrapper[4631]: E1128 13:36:11.196493 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="122f3637-af63-467a-9e98-6b5f48570d9d" containerName="mariadb-account-create-update" Nov 28 13:36:11 crc kubenswrapper[4631]: I1128 13:36:11.196508 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="122f3637-af63-467a-9e98-6b5f48570d9d" containerName="mariadb-account-create-update" Nov 28 13:36:11 crc kubenswrapper[4631]: E1128 13:36:11.196545 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3edb21b5-935d-4b42-bc88-d590b7fb0e3d" containerName="mariadb-database-create" Nov 28 13:36:11 crc kubenswrapper[4631]: I1128 13:36:11.196560 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="3edb21b5-935d-4b42-bc88-d590b7fb0e3d" containerName="mariadb-database-create" Nov 28 13:36:11 crc kubenswrapper[4631]: I1128 13:36:11.197471 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="122f3637-af63-467a-9e98-6b5f48570d9d" containerName="mariadb-account-create-update" Nov 28 13:36:11 crc kubenswrapper[4631]: I1128 13:36:11.197521 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="dd84f393-e43d-45e8-8760-7e0643f7c30e" containerName="mariadb-account-create-update" Nov 28 13:36:11 crc kubenswrapper[4631]: I1128 13:36:11.197596 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="3edb21b5-935d-4b42-bc88-d590b7fb0e3d" containerName="mariadb-database-create" Nov 28 13:36:11 crc kubenswrapper[4631]: I1128 13:36:11.199421 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-gn4cj" Nov 28 13:36:11 crc kubenswrapper[4631]: I1128 13:36:11.259856 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-gn4cj"] Nov 28 13:36:11 crc kubenswrapper[4631]: I1128 13:36:11.308832 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-m2zrc" event={"ID":"ed04b83d-c180-4bf2-b6b9-b112a536d9da","Type":"ContainerStarted","Data":"6abb47410b6f4f8e4c483aea263ea839b52160f49e3c028e22b32ec972d55780"} Nov 28 13:36:11 crc kubenswrapper[4631]: I1128 13:36:11.309016 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-698758b865-m2zrc" Nov 28 13:36:11 crc kubenswrapper[4631]: I1128 13:36:11.334884 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-698758b865-m2zrc" podStartSLOduration=3.334860404 podStartE2EDuration="3.334860404s" podCreationTimestamp="2025-11-28 13:36:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 13:36:11.328049854 +0000 UTC m=+928.135353198" watchObservedRunningTime="2025-11-28 13:36:11.334860404 +0000 UTC m=+928.142163748" Nov 28 13:36:11 crc kubenswrapper[4631]: I1128 13:36:11.362378 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c20311ad-58ca-4d2a-b19d-3b5291464558-operator-scripts\") pod \"glance-db-create-gn4cj\" (UID: \"c20311ad-58ca-4d2a-b19d-3b5291464558\") " pod="openstack/glance-db-create-gn4cj" Nov 28 13:36:11 crc kubenswrapper[4631]: I1128 13:36:11.362446 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8dbrr\" (UniqueName: \"kubernetes.io/projected/c20311ad-58ca-4d2a-b19d-3b5291464558-kube-api-access-8dbrr\") pod \"glance-db-create-gn4cj\" (UID: \"c20311ad-58ca-4d2a-b19d-3b5291464558\") " pod="openstack/glance-db-create-gn4cj" Nov 28 13:36:11 crc kubenswrapper[4631]: I1128 13:36:11.392639 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-771b-account-create-update-wzs8f"] Nov 28 13:36:11 crc kubenswrapper[4631]: I1128 13:36:11.394324 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-771b-account-create-update-wzs8f" Nov 28 13:36:11 crc kubenswrapper[4631]: I1128 13:36:11.396461 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Nov 28 13:36:11 crc kubenswrapper[4631]: I1128 13:36:11.403554 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-771b-account-create-update-wzs8f"] Nov 28 13:36:11 crc kubenswrapper[4631]: I1128 13:36:11.464423 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c20311ad-58ca-4d2a-b19d-3b5291464558-operator-scripts\") pod \"glance-db-create-gn4cj\" (UID: \"c20311ad-58ca-4d2a-b19d-3b5291464558\") " pod="openstack/glance-db-create-gn4cj" Nov 28 13:36:11 crc kubenswrapper[4631]: I1128 13:36:11.464474 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8dbrr\" (UniqueName: \"kubernetes.io/projected/c20311ad-58ca-4d2a-b19d-3b5291464558-kube-api-access-8dbrr\") pod \"glance-db-create-gn4cj\" (UID: \"c20311ad-58ca-4d2a-b19d-3b5291464558\") " pod="openstack/glance-db-create-gn4cj" Nov 28 13:36:11 crc kubenswrapper[4631]: I1128 13:36:11.466590 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c20311ad-58ca-4d2a-b19d-3b5291464558-operator-scripts\") pod \"glance-db-create-gn4cj\" (UID: \"c20311ad-58ca-4d2a-b19d-3b5291464558\") " pod="openstack/glance-db-create-gn4cj" Nov 28 13:36:11 crc kubenswrapper[4631]: I1128 13:36:11.485383 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8dbrr\" (UniqueName: \"kubernetes.io/projected/c20311ad-58ca-4d2a-b19d-3b5291464558-kube-api-access-8dbrr\") pod \"glance-db-create-gn4cj\" (UID: \"c20311ad-58ca-4d2a-b19d-3b5291464558\") " pod="openstack/glance-db-create-gn4cj" Nov 28 13:36:11 crc kubenswrapper[4631]: I1128 13:36:11.541089 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-gn4cj" Nov 28 13:36:11 crc kubenswrapper[4631]: I1128 13:36:11.565687 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/c3e1f6fb-1845-4a39-8eca-1262da1ff281-etc-swift\") pod \"swift-storage-0\" (UID: \"c3e1f6fb-1845-4a39-8eca-1262da1ff281\") " pod="openstack/swift-storage-0" Nov 28 13:36:11 crc kubenswrapper[4631]: I1128 13:36:11.565786 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7lbp2\" (UniqueName: \"kubernetes.io/projected/20f86721-2a44-4cca-976b-9d3132583903-kube-api-access-7lbp2\") pod \"glance-771b-account-create-update-wzs8f\" (UID: \"20f86721-2a44-4cca-976b-9d3132583903\") " pod="openstack/glance-771b-account-create-update-wzs8f" Nov 28 13:36:11 crc kubenswrapper[4631]: I1128 13:36:11.565854 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/20f86721-2a44-4cca-976b-9d3132583903-operator-scripts\") pod \"glance-771b-account-create-update-wzs8f\" (UID: \"20f86721-2a44-4cca-976b-9d3132583903\") " pod="openstack/glance-771b-account-create-update-wzs8f" Nov 28 13:36:11 crc kubenswrapper[4631]: E1128 13:36:11.566024 4631 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Nov 28 13:36:11 crc kubenswrapper[4631]: E1128 13:36:11.566037 4631 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Nov 28 13:36:11 crc kubenswrapper[4631]: E1128 13:36:11.566076 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/c3e1f6fb-1845-4a39-8eca-1262da1ff281-etc-swift podName:c3e1f6fb-1845-4a39-8eca-1262da1ff281 nodeName:}" failed. No retries permitted until 2025-11-28 13:36:13.566063288 +0000 UTC m=+930.373366632 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/c3e1f6fb-1845-4a39-8eca-1262da1ff281-etc-swift") pod "swift-storage-0" (UID: "c3e1f6fb-1845-4a39-8eca-1262da1ff281") : configmap "swift-ring-files" not found Nov 28 13:36:11 crc kubenswrapper[4631]: I1128 13:36:11.667373 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7lbp2\" (UniqueName: \"kubernetes.io/projected/20f86721-2a44-4cca-976b-9d3132583903-kube-api-access-7lbp2\") pod \"glance-771b-account-create-update-wzs8f\" (UID: \"20f86721-2a44-4cca-976b-9d3132583903\") " pod="openstack/glance-771b-account-create-update-wzs8f" Nov 28 13:36:11 crc kubenswrapper[4631]: I1128 13:36:11.667733 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/20f86721-2a44-4cca-976b-9d3132583903-operator-scripts\") pod \"glance-771b-account-create-update-wzs8f\" (UID: \"20f86721-2a44-4cca-976b-9d3132583903\") " pod="openstack/glance-771b-account-create-update-wzs8f" Nov 28 13:36:11 crc kubenswrapper[4631]: I1128 13:36:11.668602 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/20f86721-2a44-4cca-976b-9d3132583903-operator-scripts\") pod \"glance-771b-account-create-update-wzs8f\" (UID: \"20f86721-2a44-4cca-976b-9d3132583903\") " pod="openstack/glance-771b-account-create-update-wzs8f" Nov 28 13:36:11 crc kubenswrapper[4631]: I1128 13:36:11.714880 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7lbp2\" (UniqueName: \"kubernetes.io/projected/20f86721-2a44-4cca-976b-9d3132583903-kube-api-access-7lbp2\") pod \"glance-771b-account-create-update-wzs8f\" (UID: \"20f86721-2a44-4cca-976b-9d3132583903\") " pod="openstack/glance-771b-account-create-update-wzs8f" Nov 28 13:36:11 crc kubenswrapper[4631]: I1128 13:36:11.717688 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-771b-account-create-update-wzs8f" Nov 28 13:36:12 crc kubenswrapper[4631]: I1128 13:36:12.037175 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-gn4cj"] Nov 28 13:36:12 crc kubenswrapper[4631]: I1128 13:36:12.241068 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-771b-account-create-update-wzs8f"] Nov 28 13:36:12 crc kubenswrapper[4631]: W1128 13:36:12.246430 4631 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod20f86721_2a44_4cca_976b_9d3132583903.slice/crio-d100210a7834bc833c0e5297cfcab0d22be62a714b033356cf805bfc2e8512ca WatchSource:0}: Error finding container d100210a7834bc833c0e5297cfcab0d22be62a714b033356cf805bfc2e8512ca: Status 404 returned error can't find the container with id d100210a7834bc833c0e5297cfcab0d22be62a714b033356cf805bfc2e8512ca Nov 28 13:36:12 crc kubenswrapper[4631]: I1128 13:36:12.323221 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-gn4cj" event={"ID":"c20311ad-58ca-4d2a-b19d-3b5291464558","Type":"ContainerStarted","Data":"701d20c69a4e68bd6d91547ecc784338578cec4318a80b2af1cf939907a0d8a0"} Nov 28 13:36:12 crc kubenswrapper[4631]: I1128 13:36:12.325972 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-771b-account-create-update-wzs8f" event={"ID":"20f86721-2a44-4cca-976b-9d3132583903","Type":"ContainerStarted","Data":"d100210a7834bc833c0e5297cfcab0d22be62a714b033356cf805bfc2e8512ca"} Nov 28 13:36:13 crc kubenswrapper[4631]: I1128 13:36:13.341664 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"ca11061d-b0a1-464d-9e1d-5572fc439580","Type":"ContainerStarted","Data":"6d30eeb2b883897542d979b46c290ee18fd1d05aff2c2452b874580bd36f2d72"} Nov 28 13:36:13 crc kubenswrapper[4631]: I1128 13:36:13.346044 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-771b-account-create-update-wzs8f" event={"ID":"20f86721-2a44-4cca-976b-9d3132583903","Type":"ContainerDied","Data":"94079e09b20e24fe68c2194a6d9eeb2494fdc13708f300c19079b06192e7eda9"} Nov 28 13:36:13 crc kubenswrapper[4631]: I1128 13:36:13.345960 4631 generic.go:334] "Generic (PLEG): container finished" podID="20f86721-2a44-4cca-976b-9d3132583903" containerID="94079e09b20e24fe68c2194a6d9eeb2494fdc13708f300c19079b06192e7eda9" exitCode=0 Nov 28 13:36:13 crc kubenswrapper[4631]: I1128 13:36:13.349188 4631 generic.go:334] "Generic (PLEG): container finished" podID="c20311ad-58ca-4d2a-b19d-3b5291464558" containerID="dc82724e514dbd2f2f7f5b8d2e7c30e436f62a081faae75f298fdffee1ee9332" exitCode=0 Nov 28 13:36:13 crc kubenswrapper[4631]: I1128 13:36:13.349269 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-gn4cj" event={"ID":"c20311ad-58ca-4d2a-b19d-3b5291464558","Type":"ContainerDied","Data":"dc82724e514dbd2f2f7f5b8d2e7c30e436f62a081faae75f298fdffee1ee9332"} Nov 28 13:36:13 crc kubenswrapper[4631]: I1128 13:36:13.378429 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=4.574006923 podStartE2EDuration="1m3.378397772s" podCreationTimestamp="2025-11-28 13:35:10 +0000 UTC" firstStartedPulling="2025-11-28 13:35:14.175705513 +0000 UTC m=+870.983008857" lastFinishedPulling="2025-11-28 13:36:12.980096362 +0000 UTC m=+929.787399706" observedRunningTime="2025-11-28 13:36:13.371583782 +0000 UTC m=+930.178887146" watchObservedRunningTime="2025-11-28 13:36:13.378397772 +0000 UTC m=+930.185701126" Nov 28 13:36:13 crc kubenswrapper[4631]: I1128 13:36:13.609365 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/c3e1f6fb-1845-4a39-8eca-1262da1ff281-etc-swift\") pod \"swift-storage-0\" (UID: \"c3e1f6fb-1845-4a39-8eca-1262da1ff281\") " pod="openstack/swift-storage-0" Nov 28 13:36:13 crc kubenswrapper[4631]: E1128 13:36:13.610429 4631 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Nov 28 13:36:13 crc kubenswrapper[4631]: E1128 13:36:13.610450 4631 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Nov 28 13:36:13 crc kubenswrapper[4631]: E1128 13:36:13.610489 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/c3e1f6fb-1845-4a39-8eca-1262da1ff281-etc-swift podName:c3e1f6fb-1845-4a39-8eca-1262da1ff281 nodeName:}" failed. No retries permitted until 2025-11-28 13:36:17.610475478 +0000 UTC m=+934.417778822 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/c3e1f6fb-1845-4a39-8eca-1262da1ff281-etc-swift") pod "swift-storage-0" (UID: "c3e1f6fb-1845-4a39-8eca-1262da1ff281") : configmap "swift-ring-files" not found Nov 28 13:36:13 crc kubenswrapper[4631]: I1128 13:36:13.687507 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-wh7th"] Nov 28 13:36:13 crc kubenswrapper[4631]: I1128 13:36:13.689312 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-wh7th" Nov 28 13:36:13 crc kubenswrapper[4631]: I1128 13:36:13.692181 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Nov 28 13:36:13 crc kubenswrapper[4631]: I1128 13:36:13.692711 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Nov 28 13:36:13 crc kubenswrapper[4631]: I1128 13:36:13.699820 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Nov 28 13:36:13 crc kubenswrapper[4631]: I1128 13:36:13.704504 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-wh7th"] Nov 28 13:36:13 crc kubenswrapper[4631]: I1128 13:36:13.814027 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/780a47b4-0078-4f97-a5f5-ad4e1b2015b4-scripts\") pod \"swift-ring-rebalance-wh7th\" (UID: \"780a47b4-0078-4f97-a5f5-ad4e1b2015b4\") " pod="openstack/swift-ring-rebalance-wh7th" Nov 28 13:36:13 crc kubenswrapper[4631]: I1128 13:36:13.814103 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bhkk6\" (UniqueName: \"kubernetes.io/projected/780a47b4-0078-4f97-a5f5-ad4e1b2015b4-kube-api-access-bhkk6\") pod \"swift-ring-rebalance-wh7th\" (UID: \"780a47b4-0078-4f97-a5f5-ad4e1b2015b4\") " pod="openstack/swift-ring-rebalance-wh7th" Nov 28 13:36:13 crc kubenswrapper[4631]: I1128 13:36:13.814384 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/780a47b4-0078-4f97-a5f5-ad4e1b2015b4-etc-swift\") pod \"swift-ring-rebalance-wh7th\" (UID: \"780a47b4-0078-4f97-a5f5-ad4e1b2015b4\") " pod="openstack/swift-ring-rebalance-wh7th" Nov 28 13:36:13 crc kubenswrapper[4631]: I1128 13:36:13.814602 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/780a47b4-0078-4f97-a5f5-ad4e1b2015b4-dispersionconf\") pod \"swift-ring-rebalance-wh7th\" (UID: \"780a47b4-0078-4f97-a5f5-ad4e1b2015b4\") " pod="openstack/swift-ring-rebalance-wh7th" Nov 28 13:36:13 crc kubenswrapper[4631]: I1128 13:36:13.814641 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/780a47b4-0078-4f97-a5f5-ad4e1b2015b4-ring-data-devices\") pod \"swift-ring-rebalance-wh7th\" (UID: \"780a47b4-0078-4f97-a5f5-ad4e1b2015b4\") " pod="openstack/swift-ring-rebalance-wh7th" Nov 28 13:36:13 crc kubenswrapper[4631]: I1128 13:36:13.814730 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/780a47b4-0078-4f97-a5f5-ad4e1b2015b4-combined-ca-bundle\") pod \"swift-ring-rebalance-wh7th\" (UID: \"780a47b4-0078-4f97-a5f5-ad4e1b2015b4\") " pod="openstack/swift-ring-rebalance-wh7th" Nov 28 13:36:13 crc kubenswrapper[4631]: I1128 13:36:13.814852 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/780a47b4-0078-4f97-a5f5-ad4e1b2015b4-swiftconf\") pod \"swift-ring-rebalance-wh7th\" (UID: \"780a47b4-0078-4f97-a5f5-ad4e1b2015b4\") " pod="openstack/swift-ring-rebalance-wh7th" Nov 28 13:36:13 crc kubenswrapper[4631]: I1128 13:36:13.917241 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/780a47b4-0078-4f97-a5f5-ad4e1b2015b4-etc-swift\") pod \"swift-ring-rebalance-wh7th\" (UID: \"780a47b4-0078-4f97-a5f5-ad4e1b2015b4\") " pod="openstack/swift-ring-rebalance-wh7th" Nov 28 13:36:13 crc kubenswrapper[4631]: I1128 13:36:13.917379 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/780a47b4-0078-4f97-a5f5-ad4e1b2015b4-dispersionconf\") pod \"swift-ring-rebalance-wh7th\" (UID: \"780a47b4-0078-4f97-a5f5-ad4e1b2015b4\") " pod="openstack/swift-ring-rebalance-wh7th" Nov 28 13:36:13 crc kubenswrapper[4631]: I1128 13:36:13.917400 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/780a47b4-0078-4f97-a5f5-ad4e1b2015b4-ring-data-devices\") pod \"swift-ring-rebalance-wh7th\" (UID: \"780a47b4-0078-4f97-a5f5-ad4e1b2015b4\") " pod="openstack/swift-ring-rebalance-wh7th" Nov 28 13:36:13 crc kubenswrapper[4631]: I1128 13:36:13.917448 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/780a47b4-0078-4f97-a5f5-ad4e1b2015b4-combined-ca-bundle\") pod \"swift-ring-rebalance-wh7th\" (UID: \"780a47b4-0078-4f97-a5f5-ad4e1b2015b4\") " pod="openstack/swift-ring-rebalance-wh7th" Nov 28 13:36:13 crc kubenswrapper[4631]: I1128 13:36:13.917490 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/780a47b4-0078-4f97-a5f5-ad4e1b2015b4-swiftconf\") pod \"swift-ring-rebalance-wh7th\" (UID: \"780a47b4-0078-4f97-a5f5-ad4e1b2015b4\") " pod="openstack/swift-ring-rebalance-wh7th" Nov 28 13:36:13 crc kubenswrapper[4631]: I1128 13:36:13.917539 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/780a47b4-0078-4f97-a5f5-ad4e1b2015b4-scripts\") pod \"swift-ring-rebalance-wh7th\" (UID: \"780a47b4-0078-4f97-a5f5-ad4e1b2015b4\") " pod="openstack/swift-ring-rebalance-wh7th" Nov 28 13:36:13 crc kubenswrapper[4631]: I1128 13:36:13.917559 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bhkk6\" (UniqueName: \"kubernetes.io/projected/780a47b4-0078-4f97-a5f5-ad4e1b2015b4-kube-api-access-bhkk6\") pod \"swift-ring-rebalance-wh7th\" (UID: \"780a47b4-0078-4f97-a5f5-ad4e1b2015b4\") " pod="openstack/swift-ring-rebalance-wh7th" Nov 28 13:36:13 crc kubenswrapper[4631]: I1128 13:36:13.917979 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/780a47b4-0078-4f97-a5f5-ad4e1b2015b4-etc-swift\") pod \"swift-ring-rebalance-wh7th\" (UID: \"780a47b4-0078-4f97-a5f5-ad4e1b2015b4\") " pod="openstack/swift-ring-rebalance-wh7th" Nov 28 13:36:13 crc kubenswrapper[4631]: I1128 13:36:13.918812 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/780a47b4-0078-4f97-a5f5-ad4e1b2015b4-scripts\") pod \"swift-ring-rebalance-wh7th\" (UID: \"780a47b4-0078-4f97-a5f5-ad4e1b2015b4\") " pod="openstack/swift-ring-rebalance-wh7th" Nov 28 13:36:13 crc kubenswrapper[4631]: I1128 13:36:13.919715 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/780a47b4-0078-4f97-a5f5-ad4e1b2015b4-ring-data-devices\") pod \"swift-ring-rebalance-wh7th\" (UID: \"780a47b4-0078-4f97-a5f5-ad4e1b2015b4\") " pod="openstack/swift-ring-rebalance-wh7th" Nov 28 13:36:13 crc kubenswrapper[4631]: I1128 13:36:13.923262 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/780a47b4-0078-4f97-a5f5-ad4e1b2015b4-dispersionconf\") pod \"swift-ring-rebalance-wh7th\" (UID: \"780a47b4-0078-4f97-a5f5-ad4e1b2015b4\") " pod="openstack/swift-ring-rebalance-wh7th" Nov 28 13:36:13 crc kubenswrapper[4631]: I1128 13:36:13.924069 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/780a47b4-0078-4f97-a5f5-ad4e1b2015b4-combined-ca-bundle\") pod \"swift-ring-rebalance-wh7th\" (UID: \"780a47b4-0078-4f97-a5f5-ad4e1b2015b4\") " pod="openstack/swift-ring-rebalance-wh7th" Nov 28 13:36:13 crc kubenswrapper[4631]: I1128 13:36:13.927498 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/780a47b4-0078-4f97-a5f5-ad4e1b2015b4-swiftconf\") pod \"swift-ring-rebalance-wh7th\" (UID: \"780a47b4-0078-4f97-a5f5-ad4e1b2015b4\") " pod="openstack/swift-ring-rebalance-wh7th" Nov 28 13:36:13 crc kubenswrapper[4631]: I1128 13:36:13.940153 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bhkk6\" (UniqueName: \"kubernetes.io/projected/780a47b4-0078-4f97-a5f5-ad4e1b2015b4-kube-api-access-bhkk6\") pod \"swift-ring-rebalance-wh7th\" (UID: \"780a47b4-0078-4f97-a5f5-ad4e1b2015b4\") " pod="openstack/swift-ring-rebalance-wh7th" Nov 28 13:36:14 crc kubenswrapper[4631]: I1128 13:36:14.009851 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-wh7th" Nov 28 13:36:14 crc kubenswrapper[4631]: I1128 13:36:14.784580 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-gn4cj" Nov 28 13:36:14 crc kubenswrapper[4631]: I1128 13:36:14.859067 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8dbrr\" (UniqueName: \"kubernetes.io/projected/c20311ad-58ca-4d2a-b19d-3b5291464558-kube-api-access-8dbrr\") pod \"c20311ad-58ca-4d2a-b19d-3b5291464558\" (UID: \"c20311ad-58ca-4d2a-b19d-3b5291464558\") " Nov 28 13:36:14 crc kubenswrapper[4631]: I1128 13:36:14.859433 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c20311ad-58ca-4d2a-b19d-3b5291464558-operator-scripts\") pod \"c20311ad-58ca-4d2a-b19d-3b5291464558\" (UID: \"c20311ad-58ca-4d2a-b19d-3b5291464558\") " Nov 28 13:36:14 crc kubenswrapper[4631]: I1128 13:36:14.861023 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c20311ad-58ca-4d2a-b19d-3b5291464558-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "c20311ad-58ca-4d2a-b19d-3b5291464558" (UID: "c20311ad-58ca-4d2a-b19d-3b5291464558"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 13:36:14 crc kubenswrapper[4631]: I1128 13:36:14.867510 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c20311ad-58ca-4d2a-b19d-3b5291464558-kube-api-access-8dbrr" (OuterVolumeSpecName: "kube-api-access-8dbrr") pod "c20311ad-58ca-4d2a-b19d-3b5291464558" (UID: "c20311ad-58ca-4d2a-b19d-3b5291464558"). InnerVolumeSpecName "kube-api-access-8dbrr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:36:14 crc kubenswrapper[4631]: I1128 13:36:14.880315 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-wh7th"] Nov 28 13:36:14 crc kubenswrapper[4631]: I1128 13:36:14.941364 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-771b-account-create-update-wzs8f" Nov 28 13:36:14 crc kubenswrapper[4631]: I1128 13:36:14.964830 4631 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c20311ad-58ca-4d2a-b19d-3b5291464558-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 13:36:14 crc kubenswrapper[4631]: I1128 13:36:14.964881 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8dbrr\" (UniqueName: \"kubernetes.io/projected/c20311ad-58ca-4d2a-b19d-3b5291464558-kube-api-access-8dbrr\") on node \"crc\" DevicePath \"\"" Nov 28 13:36:15 crc kubenswrapper[4631]: I1128 13:36:15.066746 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7lbp2\" (UniqueName: \"kubernetes.io/projected/20f86721-2a44-4cca-976b-9d3132583903-kube-api-access-7lbp2\") pod \"20f86721-2a44-4cca-976b-9d3132583903\" (UID: \"20f86721-2a44-4cca-976b-9d3132583903\") " Nov 28 13:36:15 crc kubenswrapper[4631]: I1128 13:36:15.066826 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/20f86721-2a44-4cca-976b-9d3132583903-operator-scripts\") pod \"20f86721-2a44-4cca-976b-9d3132583903\" (UID: \"20f86721-2a44-4cca-976b-9d3132583903\") " Nov 28 13:36:15 crc kubenswrapper[4631]: I1128 13:36:15.067556 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/20f86721-2a44-4cca-976b-9d3132583903-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "20f86721-2a44-4cca-976b-9d3132583903" (UID: "20f86721-2a44-4cca-976b-9d3132583903"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 13:36:15 crc kubenswrapper[4631]: I1128 13:36:15.072765 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20f86721-2a44-4cca-976b-9d3132583903-kube-api-access-7lbp2" (OuterVolumeSpecName: "kube-api-access-7lbp2") pod "20f86721-2a44-4cca-976b-9d3132583903" (UID: "20f86721-2a44-4cca-976b-9d3132583903"). InnerVolumeSpecName "kube-api-access-7lbp2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:36:15 crc kubenswrapper[4631]: I1128 13:36:15.168544 4631 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/20f86721-2a44-4cca-976b-9d3132583903-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 13:36:15 crc kubenswrapper[4631]: I1128 13:36:15.169321 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7lbp2\" (UniqueName: \"kubernetes.io/projected/20f86721-2a44-4cca-976b-9d3132583903-kube-api-access-7lbp2\") on node \"crc\" DevicePath \"\"" Nov 28 13:36:15 crc kubenswrapper[4631]: I1128 13:36:15.369420 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-gn4cj" event={"ID":"c20311ad-58ca-4d2a-b19d-3b5291464558","Type":"ContainerDied","Data":"701d20c69a4e68bd6d91547ecc784338578cec4318a80b2af1cf939907a0d8a0"} Nov 28 13:36:15 crc kubenswrapper[4631]: I1128 13:36:15.369561 4631 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="701d20c69a4e68bd6d91547ecc784338578cec4318a80b2af1cf939907a0d8a0" Nov 28 13:36:15 crc kubenswrapper[4631]: I1128 13:36:15.369424 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-gn4cj" Nov 28 13:36:15 crc kubenswrapper[4631]: I1128 13:36:15.373367 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"1d3ba54c-cb26-4e04-8310-34e6b4206119","Type":"ContainerStarted","Data":"4ba58cc63d8102b7d38fc8ae8096fb7e5a5ca2417ea11fae3499d55269b92a7d"} Nov 28 13:36:15 crc kubenswrapper[4631]: I1128 13:36:15.375708 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-771b-account-create-update-wzs8f" event={"ID":"20f86721-2a44-4cca-976b-9d3132583903","Type":"ContainerDied","Data":"d100210a7834bc833c0e5297cfcab0d22be62a714b033356cf805bfc2e8512ca"} Nov 28 13:36:15 crc kubenswrapper[4631]: I1128 13:36:15.375736 4631 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d100210a7834bc833c0e5297cfcab0d22be62a714b033356cf805bfc2e8512ca" Nov 28 13:36:15 crc kubenswrapper[4631]: I1128 13:36:15.375791 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-771b-account-create-update-wzs8f" Nov 28 13:36:15 crc kubenswrapper[4631]: I1128 13:36:15.384385 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-wh7th" event={"ID":"780a47b4-0078-4f97-a5f5-ad4e1b2015b4","Type":"ContainerStarted","Data":"118997c10f58da2a447a3350629eb702b5ec5515b0dee5b2216d41ea15e8681f"} Nov 28 13:36:15 crc kubenswrapper[4631]: I1128 13:36:15.407941 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=7.5968420089999995 podStartE2EDuration="1m2.40791732s" podCreationTimestamp="2025-11-28 13:35:13 +0000 UTC" firstStartedPulling="2025-11-28 13:35:19.323412422 +0000 UTC m=+876.130715766" lastFinishedPulling="2025-11-28 13:36:14.134487733 +0000 UTC m=+930.941791077" observedRunningTime="2025-11-28 13:36:15.400753182 +0000 UTC m=+932.208056586" watchObservedRunningTime="2025-11-28 13:36:15.40791732 +0000 UTC m=+932.215220654" Nov 28 13:36:15 crc kubenswrapper[4631]: I1128 13:36:15.490630 4631 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Nov 28 13:36:16 crc kubenswrapper[4631]: I1128 13:36:16.547472 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-g7926"] Nov 28 13:36:16 crc kubenswrapper[4631]: E1128 13:36:16.549082 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c20311ad-58ca-4d2a-b19d-3b5291464558" containerName="mariadb-database-create" Nov 28 13:36:16 crc kubenswrapper[4631]: I1128 13:36:16.549098 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="c20311ad-58ca-4d2a-b19d-3b5291464558" containerName="mariadb-database-create" Nov 28 13:36:16 crc kubenswrapper[4631]: E1128 13:36:16.549124 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="20f86721-2a44-4cca-976b-9d3132583903" containerName="mariadb-account-create-update" Nov 28 13:36:16 crc kubenswrapper[4631]: I1128 13:36:16.549130 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="20f86721-2a44-4cca-976b-9d3132583903" containerName="mariadb-account-create-update" Nov 28 13:36:16 crc kubenswrapper[4631]: I1128 13:36:16.549320 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="c20311ad-58ca-4d2a-b19d-3b5291464558" containerName="mariadb-database-create" Nov 28 13:36:16 crc kubenswrapper[4631]: I1128 13:36:16.549337 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="20f86721-2a44-4cca-976b-9d3132583903" containerName="mariadb-account-create-update" Nov 28 13:36:16 crc kubenswrapper[4631]: I1128 13:36:16.549936 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-g7926" Nov 28 13:36:16 crc kubenswrapper[4631]: I1128 13:36:16.552480 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-f22sn" Nov 28 13:36:16 crc kubenswrapper[4631]: I1128 13:36:16.554223 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Nov 28 13:36:16 crc kubenswrapper[4631]: I1128 13:36:16.565541 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-g7926"] Nov 28 13:36:16 crc kubenswrapper[4631]: I1128 13:36:16.608212 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/387c22fd-1a37-4031-a53c-1a5d9e322bed-combined-ca-bundle\") pod \"glance-db-sync-g7926\" (UID: \"387c22fd-1a37-4031-a53c-1a5d9e322bed\") " pod="openstack/glance-db-sync-g7926" Nov 28 13:36:16 crc kubenswrapper[4631]: I1128 13:36:16.608333 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/387c22fd-1a37-4031-a53c-1a5d9e322bed-config-data\") pod \"glance-db-sync-g7926\" (UID: \"387c22fd-1a37-4031-a53c-1a5d9e322bed\") " pod="openstack/glance-db-sync-g7926" Nov 28 13:36:16 crc kubenswrapper[4631]: I1128 13:36:16.608376 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9g5xx\" (UniqueName: \"kubernetes.io/projected/387c22fd-1a37-4031-a53c-1a5d9e322bed-kube-api-access-9g5xx\") pod \"glance-db-sync-g7926\" (UID: \"387c22fd-1a37-4031-a53c-1a5d9e322bed\") " pod="openstack/glance-db-sync-g7926" Nov 28 13:36:16 crc kubenswrapper[4631]: I1128 13:36:16.608578 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/387c22fd-1a37-4031-a53c-1a5d9e322bed-db-sync-config-data\") pod \"glance-db-sync-g7926\" (UID: \"387c22fd-1a37-4031-a53c-1a5d9e322bed\") " pod="openstack/glance-db-sync-g7926" Nov 28 13:36:16 crc kubenswrapper[4631]: I1128 13:36:16.711592 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/387c22fd-1a37-4031-a53c-1a5d9e322bed-combined-ca-bundle\") pod \"glance-db-sync-g7926\" (UID: \"387c22fd-1a37-4031-a53c-1a5d9e322bed\") " pod="openstack/glance-db-sync-g7926" Nov 28 13:36:16 crc kubenswrapper[4631]: I1128 13:36:16.711729 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/387c22fd-1a37-4031-a53c-1a5d9e322bed-config-data\") pod \"glance-db-sync-g7926\" (UID: \"387c22fd-1a37-4031-a53c-1a5d9e322bed\") " pod="openstack/glance-db-sync-g7926" Nov 28 13:36:16 crc kubenswrapper[4631]: I1128 13:36:16.711799 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9g5xx\" (UniqueName: \"kubernetes.io/projected/387c22fd-1a37-4031-a53c-1a5d9e322bed-kube-api-access-9g5xx\") pod \"glance-db-sync-g7926\" (UID: \"387c22fd-1a37-4031-a53c-1a5d9e322bed\") " pod="openstack/glance-db-sync-g7926" Nov 28 13:36:16 crc kubenswrapper[4631]: I1128 13:36:16.711888 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/387c22fd-1a37-4031-a53c-1a5d9e322bed-db-sync-config-data\") pod \"glance-db-sync-g7926\" (UID: \"387c22fd-1a37-4031-a53c-1a5d9e322bed\") " pod="openstack/glance-db-sync-g7926" Nov 28 13:36:16 crc kubenswrapper[4631]: I1128 13:36:16.723165 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/387c22fd-1a37-4031-a53c-1a5d9e322bed-combined-ca-bundle\") pod \"glance-db-sync-g7926\" (UID: \"387c22fd-1a37-4031-a53c-1a5d9e322bed\") " pod="openstack/glance-db-sync-g7926" Nov 28 13:36:16 crc kubenswrapper[4631]: I1128 13:36:16.732263 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/387c22fd-1a37-4031-a53c-1a5d9e322bed-db-sync-config-data\") pod \"glance-db-sync-g7926\" (UID: \"387c22fd-1a37-4031-a53c-1a5d9e322bed\") " pod="openstack/glance-db-sync-g7926" Nov 28 13:36:16 crc kubenswrapper[4631]: I1128 13:36:16.732780 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/387c22fd-1a37-4031-a53c-1a5d9e322bed-config-data\") pod \"glance-db-sync-g7926\" (UID: \"387c22fd-1a37-4031-a53c-1a5d9e322bed\") " pod="openstack/glance-db-sync-g7926" Nov 28 13:36:16 crc kubenswrapper[4631]: I1128 13:36:16.736900 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9g5xx\" (UniqueName: \"kubernetes.io/projected/387c22fd-1a37-4031-a53c-1a5d9e322bed-kube-api-access-9g5xx\") pod \"glance-db-sync-g7926\" (UID: \"387c22fd-1a37-4031-a53c-1a5d9e322bed\") " pod="openstack/glance-db-sync-g7926" Nov 28 13:36:16 crc kubenswrapper[4631]: I1128 13:36:16.929387 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-g7926" Nov 28 13:36:17 crc kubenswrapper[4631]: I1128 13:36:17.490669 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Nov 28 13:36:17 crc kubenswrapper[4631]: I1128 13:36:17.639630 4631 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Nov 28 13:36:17 crc kubenswrapper[4631]: I1128 13:36:17.642914 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/c3e1f6fb-1845-4a39-8eca-1262da1ff281-etc-swift\") pod \"swift-storage-0\" (UID: \"c3e1f6fb-1845-4a39-8eca-1262da1ff281\") " pod="openstack/swift-storage-0" Nov 28 13:36:17 crc kubenswrapper[4631]: E1128 13:36:17.643130 4631 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Nov 28 13:36:17 crc kubenswrapper[4631]: E1128 13:36:17.643155 4631 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Nov 28 13:36:17 crc kubenswrapper[4631]: E1128 13:36:17.643937 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/c3e1f6fb-1845-4a39-8eca-1262da1ff281-etc-swift podName:c3e1f6fb-1845-4a39-8eca-1262da1ff281 nodeName:}" failed. No retries permitted until 2025-11-28 13:36:25.643919417 +0000 UTC m=+942.451222921 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/c3e1f6fb-1845-4a39-8eca-1262da1ff281-etc-swift") pod "swift-storage-0" (UID: "c3e1f6fb-1845-4a39-8eca-1262da1ff281") : configmap "swift-ring-files" not found Nov 28 13:36:18 crc kubenswrapper[4631]: I1128 13:36:18.528808 4631 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Nov 28 13:36:18 crc kubenswrapper[4631]: I1128 13:36:18.585191 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Nov 28 13:36:18 crc kubenswrapper[4631]: I1128 13:36:18.955982 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-698758b865-m2zrc" Nov 28 13:36:19 crc kubenswrapper[4631]: I1128 13:36:19.010431 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-49s6f"] Nov 28 13:36:19 crc kubenswrapper[4631]: I1128 13:36:19.010699 4631 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-86db49b7ff-49s6f" podUID="1959423d-7154-41e1-80db-d5840a9c1702" containerName="dnsmasq-dns" containerID="cri-o://5cbe99b53ff72888b0833a40a1bbcc9dd936f48191d7159210c2429b4be79dad" gracePeriod=10 Nov 28 13:36:19 crc kubenswrapper[4631]: E1128 13:36:19.212315 4631 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1959423d_7154_41e1_80db_d5840a9c1702.slice/crio-5cbe99b53ff72888b0833a40a1bbcc9dd936f48191d7159210c2429b4be79dad.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1959423d_7154_41e1_80db_d5840a9c1702.slice/crio-conmon-5cbe99b53ff72888b0833a40a1bbcc9dd936f48191d7159210c2429b4be79dad.scope\": RecentStats: unable to find data in memory cache]" Nov 28 13:36:19 crc kubenswrapper[4631]: I1128 13:36:19.261423 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-g7926"] Nov 28 13:36:19 crc kubenswrapper[4631]: I1128 13:36:19.419574 4631 generic.go:334] "Generic (PLEG): container finished" podID="1959423d-7154-41e1-80db-d5840a9c1702" containerID="5cbe99b53ff72888b0833a40a1bbcc9dd936f48191d7159210c2429b4be79dad" exitCode=0 Nov 28 13:36:19 crc kubenswrapper[4631]: I1128 13:36:19.419675 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-49s6f" event={"ID":"1959423d-7154-41e1-80db-d5840a9c1702","Type":"ContainerDied","Data":"5cbe99b53ff72888b0833a40a1bbcc9dd936f48191d7159210c2429b4be79dad"} Nov 28 13:36:19 crc kubenswrapper[4631]: I1128 13:36:19.421860 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-wh7th" event={"ID":"780a47b4-0078-4f97-a5f5-ad4e1b2015b4","Type":"ContainerStarted","Data":"6a9fe81dc28b378017a91c6e0546bf8ece70a411c480b3ffd07430a03a003cd8"} Nov 28 13:36:19 crc kubenswrapper[4631]: I1128 13:36:19.425201 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-g7926" event={"ID":"387c22fd-1a37-4031-a53c-1a5d9e322bed","Type":"ContainerStarted","Data":"0c145e40542d56fae99ece808ecb6f3fe1b858b045c8401bf0b58516d854fb81"} Nov 28 13:36:19 crc kubenswrapper[4631]: I1128 13:36:19.444820 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-wh7th" podStartSLOduration=2.637185797 podStartE2EDuration="6.444720123s" podCreationTimestamp="2025-11-28 13:36:13 +0000 UTC" firstStartedPulling="2025-11-28 13:36:14.900529941 +0000 UTC m=+931.707833285" lastFinishedPulling="2025-11-28 13:36:18.708064257 +0000 UTC m=+935.515367611" observedRunningTime="2025-11-28 13:36:19.444323483 +0000 UTC m=+936.251626847" watchObservedRunningTime="2025-11-28 13:36:19.444720123 +0000 UTC m=+936.252023467" Nov 28 13:36:19 crc kubenswrapper[4631]: I1128 13:36:19.468684 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-49s6f" Nov 28 13:36:19 crc kubenswrapper[4631]: I1128 13:36:19.588022 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1959423d-7154-41e1-80db-d5840a9c1702-dns-svc\") pod \"1959423d-7154-41e1-80db-d5840a9c1702\" (UID: \"1959423d-7154-41e1-80db-d5840a9c1702\") " Nov 28 13:36:19 crc kubenswrapper[4631]: I1128 13:36:19.588085 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1959423d-7154-41e1-80db-d5840a9c1702-ovsdbserver-nb\") pod \"1959423d-7154-41e1-80db-d5840a9c1702\" (UID: \"1959423d-7154-41e1-80db-d5840a9c1702\") " Nov 28 13:36:19 crc kubenswrapper[4631]: I1128 13:36:19.588131 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1959423d-7154-41e1-80db-d5840a9c1702-ovsdbserver-sb\") pod \"1959423d-7154-41e1-80db-d5840a9c1702\" (UID: \"1959423d-7154-41e1-80db-d5840a9c1702\") " Nov 28 13:36:19 crc kubenswrapper[4631]: I1128 13:36:19.588179 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j95lf\" (UniqueName: \"kubernetes.io/projected/1959423d-7154-41e1-80db-d5840a9c1702-kube-api-access-j95lf\") pod \"1959423d-7154-41e1-80db-d5840a9c1702\" (UID: \"1959423d-7154-41e1-80db-d5840a9c1702\") " Nov 28 13:36:19 crc kubenswrapper[4631]: I1128 13:36:19.588249 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1959423d-7154-41e1-80db-d5840a9c1702-config\") pod \"1959423d-7154-41e1-80db-d5840a9c1702\" (UID: \"1959423d-7154-41e1-80db-d5840a9c1702\") " Nov 28 13:36:19 crc kubenswrapper[4631]: I1128 13:36:19.594677 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1959423d-7154-41e1-80db-d5840a9c1702-kube-api-access-j95lf" (OuterVolumeSpecName: "kube-api-access-j95lf") pod "1959423d-7154-41e1-80db-d5840a9c1702" (UID: "1959423d-7154-41e1-80db-d5840a9c1702"). InnerVolumeSpecName "kube-api-access-j95lf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:36:19 crc kubenswrapper[4631]: I1128 13:36:19.639815 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Nov 28 13:36:19 crc kubenswrapper[4631]: I1128 13:36:19.646912 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1959423d-7154-41e1-80db-d5840a9c1702-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "1959423d-7154-41e1-80db-d5840a9c1702" (UID: "1959423d-7154-41e1-80db-d5840a9c1702"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 13:36:19 crc kubenswrapper[4631]: I1128 13:36:19.657996 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1959423d-7154-41e1-80db-d5840a9c1702-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "1959423d-7154-41e1-80db-d5840a9c1702" (UID: "1959423d-7154-41e1-80db-d5840a9c1702"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 13:36:19 crc kubenswrapper[4631]: I1128 13:36:19.664873 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1959423d-7154-41e1-80db-d5840a9c1702-config" (OuterVolumeSpecName: "config") pod "1959423d-7154-41e1-80db-d5840a9c1702" (UID: "1959423d-7154-41e1-80db-d5840a9c1702"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 13:36:19 crc kubenswrapper[4631]: I1128 13:36:19.678409 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1959423d-7154-41e1-80db-d5840a9c1702-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "1959423d-7154-41e1-80db-d5840a9c1702" (UID: "1959423d-7154-41e1-80db-d5840a9c1702"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 13:36:19 crc kubenswrapper[4631]: I1128 13:36:19.690828 4631 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1959423d-7154-41e1-80db-d5840a9c1702-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 28 13:36:19 crc kubenswrapper[4631]: I1128 13:36:19.690872 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j95lf\" (UniqueName: \"kubernetes.io/projected/1959423d-7154-41e1-80db-d5840a9c1702-kube-api-access-j95lf\") on node \"crc\" DevicePath \"\"" Nov 28 13:36:19 crc kubenswrapper[4631]: I1128 13:36:19.690886 4631 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1959423d-7154-41e1-80db-d5840a9c1702-config\") on node \"crc\" DevicePath \"\"" Nov 28 13:36:19 crc kubenswrapper[4631]: I1128 13:36:19.690895 4631 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1959423d-7154-41e1-80db-d5840a9c1702-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 28 13:36:19 crc kubenswrapper[4631]: I1128 13:36:19.690906 4631 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1959423d-7154-41e1-80db-d5840a9c1702-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 28 13:36:20 crc kubenswrapper[4631]: I1128 13:36:20.439399 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-49s6f" Nov 28 13:36:20 crc kubenswrapper[4631]: I1128 13:36:20.440233 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-49s6f" event={"ID":"1959423d-7154-41e1-80db-d5840a9c1702","Type":"ContainerDied","Data":"9347a7722abe5bca8284ab39251419763b393fa66ee05fa9a291f8c8da05ee5e"} Nov 28 13:36:20 crc kubenswrapper[4631]: I1128 13:36:20.440340 4631 scope.go:117] "RemoveContainer" containerID="5cbe99b53ff72888b0833a40a1bbcc9dd936f48191d7159210c2429b4be79dad" Nov 28 13:36:20 crc kubenswrapper[4631]: I1128 13:36:20.485637 4631 scope.go:117] "RemoveContainer" containerID="763e2d68c0063a8e7bec77440ee5a31d6b77bc422f4805a4ce9c2183d94e6d55" Nov 28 13:36:20 crc kubenswrapper[4631]: I1128 13:36:20.497667 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-49s6f"] Nov 28 13:36:20 crc kubenswrapper[4631]: I1128 13:36:20.505383 4631 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-49s6f"] Nov 28 13:36:20 crc kubenswrapper[4631]: I1128 13:36:20.687693 4631 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Nov 28 13:36:20 crc kubenswrapper[4631]: I1128 13:36:20.737078 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Nov 28 13:36:20 crc kubenswrapper[4631]: I1128 13:36:20.947182 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Nov 28 13:36:20 crc kubenswrapper[4631]: E1128 13:36:20.947714 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1959423d-7154-41e1-80db-d5840a9c1702" containerName="dnsmasq-dns" Nov 28 13:36:20 crc kubenswrapper[4631]: I1128 13:36:20.947742 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="1959423d-7154-41e1-80db-d5840a9c1702" containerName="dnsmasq-dns" Nov 28 13:36:20 crc kubenswrapper[4631]: E1128 13:36:20.947780 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1959423d-7154-41e1-80db-d5840a9c1702" containerName="init" Nov 28 13:36:20 crc kubenswrapper[4631]: I1128 13:36:20.947790 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="1959423d-7154-41e1-80db-d5840a9c1702" containerName="init" Nov 28 13:36:20 crc kubenswrapper[4631]: I1128 13:36:20.948022 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="1959423d-7154-41e1-80db-d5840a9c1702" containerName="dnsmasq-dns" Nov 28 13:36:20 crc kubenswrapper[4631]: I1128 13:36:20.949857 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Nov 28 13:36:20 crc kubenswrapper[4631]: I1128 13:36:20.952557 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Nov 28 13:36:20 crc kubenswrapper[4631]: I1128 13:36:20.952847 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Nov 28 13:36:20 crc kubenswrapper[4631]: I1128 13:36:20.957577 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-fg68q" Nov 28 13:36:20 crc kubenswrapper[4631]: I1128 13:36:20.957918 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Nov 28 13:36:20 crc kubenswrapper[4631]: I1128 13:36:20.968934 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Nov 28 13:36:21 crc kubenswrapper[4631]: I1128 13:36:21.016641 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1418c5fe-bc14-4aeb-9449-30c85e26aed2-config\") pod \"ovn-northd-0\" (UID: \"1418c5fe-bc14-4aeb-9449-30c85e26aed2\") " pod="openstack/ovn-northd-0" Nov 28 13:36:21 crc kubenswrapper[4631]: I1128 13:36:21.016726 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/1418c5fe-bc14-4aeb-9449-30c85e26aed2-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"1418c5fe-bc14-4aeb-9449-30c85e26aed2\") " pod="openstack/ovn-northd-0" Nov 28 13:36:21 crc kubenswrapper[4631]: I1128 13:36:21.016762 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/1418c5fe-bc14-4aeb-9449-30c85e26aed2-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"1418c5fe-bc14-4aeb-9449-30c85e26aed2\") " pod="openstack/ovn-northd-0" Nov 28 13:36:21 crc kubenswrapper[4631]: I1128 13:36:21.016813 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/1418c5fe-bc14-4aeb-9449-30c85e26aed2-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"1418c5fe-bc14-4aeb-9449-30c85e26aed2\") " pod="openstack/ovn-northd-0" Nov 28 13:36:21 crc kubenswrapper[4631]: I1128 13:36:21.016849 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gxlhz\" (UniqueName: \"kubernetes.io/projected/1418c5fe-bc14-4aeb-9449-30c85e26aed2-kube-api-access-gxlhz\") pod \"ovn-northd-0\" (UID: \"1418c5fe-bc14-4aeb-9449-30c85e26aed2\") " pod="openstack/ovn-northd-0" Nov 28 13:36:21 crc kubenswrapper[4631]: I1128 13:36:21.016919 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1418c5fe-bc14-4aeb-9449-30c85e26aed2-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"1418c5fe-bc14-4aeb-9449-30c85e26aed2\") " pod="openstack/ovn-northd-0" Nov 28 13:36:21 crc kubenswrapper[4631]: I1128 13:36:21.017006 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1418c5fe-bc14-4aeb-9449-30c85e26aed2-scripts\") pod \"ovn-northd-0\" (UID: \"1418c5fe-bc14-4aeb-9449-30c85e26aed2\") " pod="openstack/ovn-northd-0" Nov 28 13:36:21 crc kubenswrapper[4631]: I1128 13:36:21.119003 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/1418c5fe-bc14-4aeb-9449-30c85e26aed2-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"1418c5fe-bc14-4aeb-9449-30c85e26aed2\") " pod="openstack/ovn-northd-0" Nov 28 13:36:21 crc kubenswrapper[4631]: I1128 13:36:21.119050 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/1418c5fe-bc14-4aeb-9449-30c85e26aed2-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"1418c5fe-bc14-4aeb-9449-30c85e26aed2\") " pod="openstack/ovn-northd-0" Nov 28 13:36:21 crc kubenswrapper[4631]: I1128 13:36:21.119097 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/1418c5fe-bc14-4aeb-9449-30c85e26aed2-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"1418c5fe-bc14-4aeb-9449-30c85e26aed2\") " pod="openstack/ovn-northd-0" Nov 28 13:36:21 crc kubenswrapper[4631]: I1128 13:36:21.119123 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gxlhz\" (UniqueName: \"kubernetes.io/projected/1418c5fe-bc14-4aeb-9449-30c85e26aed2-kube-api-access-gxlhz\") pod \"ovn-northd-0\" (UID: \"1418c5fe-bc14-4aeb-9449-30c85e26aed2\") " pod="openstack/ovn-northd-0" Nov 28 13:36:21 crc kubenswrapper[4631]: I1128 13:36:21.119177 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1418c5fe-bc14-4aeb-9449-30c85e26aed2-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"1418c5fe-bc14-4aeb-9449-30c85e26aed2\") " pod="openstack/ovn-northd-0" Nov 28 13:36:21 crc kubenswrapper[4631]: I1128 13:36:21.119236 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1418c5fe-bc14-4aeb-9449-30c85e26aed2-scripts\") pod \"ovn-northd-0\" (UID: \"1418c5fe-bc14-4aeb-9449-30c85e26aed2\") " pod="openstack/ovn-northd-0" Nov 28 13:36:21 crc kubenswrapper[4631]: I1128 13:36:21.119272 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1418c5fe-bc14-4aeb-9449-30c85e26aed2-config\") pod \"ovn-northd-0\" (UID: \"1418c5fe-bc14-4aeb-9449-30c85e26aed2\") " pod="openstack/ovn-northd-0" Nov 28 13:36:21 crc kubenswrapper[4631]: I1128 13:36:21.119722 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/1418c5fe-bc14-4aeb-9449-30c85e26aed2-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"1418c5fe-bc14-4aeb-9449-30c85e26aed2\") " pod="openstack/ovn-northd-0" Nov 28 13:36:21 crc kubenswrapper[4631]: I1128 13:36:21.120273 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1418c5fe-bc14-4aeb-9449-30c85e26aed2-config\") pod \"ovn-northd-0\" (UID: \"1418c5fe-bc14-4aeb-9449-30c85e26aed2\") " pod="openstack/ovn-northd-0" Nov 28 13:36:21 crc kubenswrapper[4631]: I1128 13:36:21.120919 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1418c5fe-bc14-4aeb-9449-30c85e26aed2-scripts\") pod \"ovn-northd-0\" (UID: \"1418c5fe-bc14-4aeb-9449-30c85e26aed2\") " pod="openstack/ovn-northd-0" Nov 28 13:36:21 crc kubenswrapper[4631]: I1128 13:36:21.128058 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/1418c5fe-bc14-4aeb-9449-30c85e26aed2-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"1418c5fe-bc14-4aeb-9449-30c85e26aed2\") " pod="openstack/ovn-northd-0" Nov 28 13:36:21 crc kubenswrapper[4631]: I1128 13:36:21.136716 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/1418c5fe-bc14-4aeb-9449-30c85e26aed2-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"1418c5fe-bc14-4aeb-9449-30c85e26aed2\") " pod="openstack/ovn-northd-0" Nov 28 13:36:21 crc kubenswrapper[4631]: I1128 13:36:21.139597 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1418c5fe-bc14-4aeb-9449-30c85e26aed2-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"1418c5fe-bc14-4aeb-9449-30c85e26aed2\") " pod="openstack/ovn-northd-0" Nov 28 13:36:21 crc kubenswrapper[4631]: I1128 13:36:21.155321 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gxlhz\" (UniqueName: \"kubernetes.io/projected/1418c5fe-bc14-4aeb-9449-30c85e26aed2-kube-api-access-gxlhz\") pod \"ovn-northd-0\" (UID: \"1418c5fe-bc14-4aeb-9449-30c85e26aed2\") " pod="openstack/ovn-northd-0" Nov 28 13:36:21 crc kubenswrapper[4631]: I1128 13:36:21.272422 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Nov 28 13:36:21 crc kubenswrapper[4631]: I1128 13:36:21.531091 4631 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1959423d-7154-41e1-80db-d5840a9c1702" path="/var/lib/kubelet/pods/1959423d-7154-41e1-80db-d5840a9c1702/volumes" Nov 28 13:36:21 crc kubenswrapper[4631]: I1128 13:36:21.774412 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Nov 28 13:36:22 crc kubenswrapper[4631]: I1128 13:36:22.484604 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"1418c5fe-bc14-4aeb-9449-30c85e26aed2","Type":"ContainerStarted","Data":"941d26ebcec7356d35bd38a96fbc3bf728c51df1dd4dc1f9ea012e143709feaf"} Nov 28 13:36:23 crc kubenswrapper[4631]: I1128 13:36:23.496973 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"1418c5fe-bc14-4aeb-9449-30c85e26aed2","Type":"ContainerStarted","Data":"754e9e0be4d03d5f9154b5cfc00c27b2170a140ae4edf1ad7ac8f13d9467a5c5"} Nov 28 13:36:23 crc kubenswrapper[4631]: I1128 13:36:23.497608 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Nov 28 13:36:23 crc kubenswrapper[4631]: I1128 13:36:23.497682 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"1418c5fe-bc14-4aeb-9449-30c85e26aed2","Type":"ContainerStarted","Data":"2c7512427201e8e8c61ce34c2fdd1128251dae7c1354a8267e8b65444d51d427"} Nov 28 13:36:23 crc kubenswrapper[4631]: I1128 13:36:23.527892 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=2.298399099 podStartE2EDuration="3.527872861s" podCreationTimestamp="2025-11-28 13:36:20 +0000 UTC" firstStartedPulling="2025-11-28 13:36:21.785643375 +0000 UTC m=+938.592946729" lastFinishedPulling="2025-11-28 13:36:23.015117147 +0000 UTC m=+939.822420491" observedRunningTime="2025-11-28 13:36:23.524417704 +0000 UTC m=+940.331721058" watchObservedRunningTime="2025-11-28 13:36:23.527872861 +0000 UTC m=+940.335176205" Nov 28 13:36:25 crc kubenswrapper[4631]: I1128 13:36:25.725627 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/c3e1f6fb-1845-4a39-8eca-1262da1ff281-etc-swift\") pod \"swift-storage-0\" (UID: \"c3e1f6fb-1845-4a39-8eca-1262da1ff281\") " pod="openstack/swift-storage-0" Nov 28 13:36:25 crc kubenswrapper[4631]: E1128 13:36:25.726372 4631 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Nov 28 13:36:25 crc kubenswrapper[4631]: E1128 13:36:25.726395 4631 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Nov 28 13:36:25 crc kubenswrapper[4631]: E1128 13:36:25.726468 4631 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/c3e1f6fb-1845-4a39-8eca-1262da1ff281-etc-swift podName:c3e1f6fb-1845-4a39-8eca-1262da1ff281 nodeName:}" failed. No retries permitted until 2025-11-28 13:36:41.726443334 +0000 UTC m=+958.533746688 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/c3e1f6fb-1845-4a39-8eca-1262da1ff281-etc-swift") pod "swift-storage-0" (UID: "c3e1f6fb-1845-4a39-8eca-1262da1ff281") : configmap "swift-ring-files" not found Nov 28 13:36:26 crc kubenswrapper[4631]: I1128 13:36:26.522136 4631 generic.go:334] "Generic (PLEG): container finished" podID="3a91e9fa-34b7-4688-9171-56a62043759d" containerID="5f2c613e618512327c6478adcd7de0b71456862e29e5333eaab2b49e071f2777" exitCode=0 Nov 28 13:36:26 crc kubenswrapper[4631]: I1128 13:36:26.522235 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"3a91e9fa-34b7-4688-9171-56a62043759d","Type":"ContainerDied","Data":"5f2c613e618512327c6478adcd7de0b71456862e29e5333eaab2b49e071f2777"} Nov 28 13:36:26 crc kubenswrapper[4631]: I1128 13:36:26.524272 4631 generic.go:334] "Generic (PLEG): container finished" podID="ed4141f2-5eff-4252-8dc7-ad60a1580189" containerID="41d8c218651f1d165fa68a2452a35246fa25140ebacda04483d76b1fc68b2e42" exitCode=0 Nov 28 13:36:26 crc kubenswrapper[4631]: I1128 13:36:26.524460 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"ed4141f2-5eff-4252-8dc7-ad60a1580189","Type":"ContainerDied","Data":"41d8c218651f1d165fa68a2452a35246fa25140ebacda04483d76b1fc68b2e42"} Nov 28 13:36:27 crc kubenswrapper[4631]: I1128 13:36:27.538527 4631 generic.go:334] "Generic (PLEG): container finished" podID="780a47b4-0078-4f97-a5f5-ad4e1b2015b4" containerID="6a9fe81dc28b378017a91c6e0546bf8ece70a411c480b3ffd07430a03a003cd8" exitCode=0 Nov 28 13:36:27 crc kubenswrapper[4631]: I1128 13:36:27.538599 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-wh7th" event={"ID":"780a47b4-0078-4f97-a5f5-ad4e1b2015b4","Type":"ContainerDied","Data":"6a9fe81dc28b378017a91c6e0546bf8ece70a411c480b3ffd07430a03a003cd8"} Nov 28 13:36:33 crc kubenswrapper[4631]: I1128 13:36:33.373013 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-wh7th" Nov 28 13:36:33 crc kubenswrapper[4631]: I1128 13:36:33.518527 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/780a47b4-0078-4f97-a5f5-ad4e1b2015b4-swiftconf\") pod \"780a47b4-0078-4f97-a5f5-ad4e1b2015b4\" (UID: \"780a47b4-0078-4f97-a5f5-ad4e1b2015b4\") " Nov 28 13:36:33 crc kubenswrapper[4631]: I1128 13:36:33.518974 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/780a47b4-0078-4f97-a5f5-ad4e1b2015b4-ring-data-devices\") pod \"780a47b4-0078-4f97-a5f5-ad4e1b2015b4\" (UID: \"780a47b4-0078-4f97-a5f5-ad4e1b2015b4\") " Nov 28 13:36:33 crc kubenswrapper[4631]: I1128 13:36:33.519096 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/780a47b4-0078-4f97-a5f5-ad4e1b2015b4-scripts\") pod \"780a47b4-0078-4f97-a5f5-ad4e1b2015b4\" (UID: \"780a47b4-0078-4f97-a5f5-ad4e1b2015b4\") " Nov 28 13:36:33 crc kubenswrapper[4631]: I1128 13:36:33.519140 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/780a47b4-0078-4f97-a5f5-ad4e1b2015b4-combined-ca-bundle\") pod \"780a47b4-0078-4f97-a5f5-ad4e1b2015b4\" (UID: \"780a47b4-0078-4f97-a5f5-ad4e1b2015b4\") " Nov 28 13:36:33 crc kubenswrapper[4631]: I1128 13:36:33.519201 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/780a47b4-0078-4f97-a5f5-ad4e1b2015b4-etc-swift\") pod \"780a47b4-0078-4f97-a5f5-ad4e1b2015b4\" (UID: \"780a47b4-0078-4f97-a5f5-ad4e1b2015b4\") " Nov 28 13:36:33 crc kubenswrapper[4631]: I1128 13:36:33.519239 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bhkk6\" (UniqueName: \"kubernetes.io/projected/780a47b4-0078-4f97-a5f5-ad4e1b2015b4-kube-api-access-bhkk6\") pod \"780a47b4-0078-4f97-a5f5-ad4e1b2015b4\" (UID: \"780a47b4-0078-4f97-a5f5-ad4e1b2015b4\") " Nov 28 13:36:33 crc kubenswrapper[4631]: I1128 13:36:33.519261 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/780a47b4-0078-4f97-a5f5-ad4e1b2015b4-dispersionconf\") pod \"780a47b4-0078-4f97-a5f5-ad4e1b2015b4\" (UID: \"780a47b4-0078-4f97-a5f5-ad4e1b2015b4\") " Nov 28 13:36:33 crc kubenswrapper[4631]: I1128 13:36:33.520105 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/780a47b4-0078-4f97-a5f5-ad4e1b2015b4-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "780a47b4-0078-4f97-a5f5-ad4e1b2015b4" (UID: "780a47b4-0078-4f97-a5f5-ad4e1b2015b4"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 13:36:33 crc kubenswrapper[4631]: I1128 13:36:33.525588 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/780a47b4-0078-4f97-a5f5-ad4e1b2015b4-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "780a47b4-0078-4f97-a5f5-ad4e1b2015b4" (UID: "780a47b4-0078-4f97-a5f5-ad4e1b2015b4"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 13:36:33 crc kubenswrapper[4631]: I1128 13:36:33.531850 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/780a47b4-0078-4f97-a5f5-ad4e1b2015b4-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "780a47b4-0078-4f97-a5f5-ad4e1b2015b4" (UID: "780a47b4-0078-4f97-a5f5-ad4e1b2015b4"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:36:33 crc kubenswrapper[4631]: I1128 13:36:33.537699 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/780a47b4-0078-4f97-a5f5-ad4e1b2015b4-kube-api-access-bhkk6" (OuterVolumeSpecName: "kube-api-access-bhkk6") pod "780a47b4-0078-4f97-a5f5-ad4e1b2015b4" (UID: "780a47b4-0078-4f97-a5f5-ad4e1b2015b4"). InnerVolumeSpecName "kube-api-access-bhkk6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:36:33 crc kubenswrapper[4631]: I1128 13:36:33.544656 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/780a47b4-0078-4f97-a5f5-ad4e1b2015b4-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "780a47b4-0078-4f97-a5f5-ad4e1b2015b4" (UID: "780a47b4-0078-4f97-a5f5-ad4e1b2015b4"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:36:33 crc kubenswrapper[4631]: I1128 13:36:33.556112 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/780a47b4-0078-4f97-a5f5-ad4e1b2015b4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "780a47b4-0078-4f97-a5f5-ad4e1b2015b4" (UID: "780a47b4-0078-4f97-a5f5-ad4e1b2015b4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:36:33 crc kubenswrapper[4631]: I1128 13:36:33.560075 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/780a47b4-0078-4f97-a5f5-ad4e1b2015b4-scripts" (OuterVolumeSpecName: "scripts") pod "780a47b4-0078-4f97-a5f5-ad4e1b2015b4" (UID: "780a47b4-0078-4f97-a5f5-ad4e1b2015b4"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 13:36:33 crc kubenswrapper[4631]: I1128 13:36:33.608433 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"3a91e9fa-34b7-4688-9171-56a62043759d","Type":"ContainerStarted","Data":"798f0920b82c4321f979cbb346e8d3258fa7abb13c5c825bae8d9f3b06fd453e"} Nov 28 13:36:33 crc kubenswrapper[4631]: I1128 13:36:33.608759 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Nov 28 13:36:33 crc kubenswrapper[4631]: I1128 13:36:33.611219 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-wh7th" event={"ID":"780a47b4-0078-4f97-a5f5-ad4e1b2015b4","Type":"ContainerDied","Data":"118997c10f58da2a447a3350629eb702b5ec5515b0dee5b2216d41ea15e8681f"} Nov 28 13:36:33 crc kubenswrapper[4631]: I1128 13:36:33.611247 4631 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="118997c10f58da2a447a3350629eb702b5ec5515b0dee5b2216d41ea15e8681f" Nov 28 13:36:33 crc kubenswrapper[4631]: I1128 13:36:33.611322 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-wh7th" Nov 28 13:36:33 crc kubenswrapper[4631]: I1128 13:36:33.616033 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"ed4141f2-5eff-4252-8dc7-ad60a1580189","Type":"ContainerStarted","Data":"bc7825d2e428610e7d264dba3f92af76fc65d8e94ba32e808f0899e4d096a085"} Nov 28 13:36:33 crc kubenswrapper[4631]: I1128 13:36:33.616969 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Nov 28 13:36:33 crc kubenswrapper[4631]: I1128 13:36:33.621417 4631 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/780a47b4-0078-4f97-a5f5-ad4e1b2015b4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 13:36:33 crc kubenswrapper[4631]: I1128 13:36:33.621449 4631 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/780a47b4-0078-4f97-a5f5-ad4e1b2015b4-etc-swift\") on node \"crc\" DevicePath \"\"" Nov 28 13:36:33 crc kubenswrapper[4631]: I1128 13:36:33.621461 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bhkk6\" (UniqueName: \"kubernetes.io/projected/780a47b4-0078-4f97-a5f5-ad4e1b2015b4-kube-api-access-bhkk6\") on node \"crc\" DevicePath \"\"" Nov 28 13:36:33 crc kubenswrapper[4631]: I1128 13:36:33.621472 4631 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/780a47b4-0078-4f97-a5f5-ad4e1b2015b4-dispersionconf\") on node \"crc\" DevicePath \"\"" Nov 28 13:36:33 crc kubenswrapper[4631]: I1128 13:36:33.621485 4631 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/780a47b4-0078-4f97-a5f5-ad4e1b2015b4-swiftconf\") on node \"crc\" DevicePath \"\"" Nov 28 13:36:33 crc kubenswrapper[4631]: I1128 13:36:33.621498 4631 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/780a47b4-0078-4f97-a5f5-ad4e1b2015b4-ring-data-devices\") on node \"crc\" DevicePath \"\"" Nov 28 13:36:33 crc kubenswrapper[4631]: I1128 13:36:33.621513 4631 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/780a47b4-0078-4f97-a5f5-ad4e1b2015b4-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 13:36:33 crc kubenswrapper[4631]: I1128 13:36:33.672054 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=44.128009169 podStartE2EDuration="1m32.672037006s" podCreationTimestamp="2025-11-28 13:35:01 +0000 UTC" firstStartedPulling="2025-11-28 13:35:04.04004642 +0000 UTC m=+860.847349764" lastFinishedPulling="2025-11-28 13:35:52.584074257 +0000 UTC m=+909.391377601" observedRunningTime="2025-11-28 13:36:33.669698688 +0000 UTC m=+950.477002042" watchObservedRunningTime="2025-11-28 13:36:33.672037006 +0000 UTC m=+950.479340350" Nov 28 13:36:33 crc kubenswrapper[4631]: I1128 13:36:33.674895 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=44.514751531 podStartE2EDuration="1m33.674886397s" podCreationTimestamp="2025-11-28 13:35:00 +0000 UTC" firstStartedPulling="2025-11-28 13:35:03.387970634 +0000 UTC m=+860.195273978" lastFinishedPulling="2025-11-28 13:35:52.54810549 +0000 UTC m=+909.355408844" observedRunningTime="2025-11-28 13:36:33.637949176 +0000 UTC m=+950.445252530" watchObservedRunningTime="2025-11-28 13:36:33.674886397 +0000 UTC m=+950.482189741" Nov 28 13:36:34 crc kubenswrapper[4631]: I1128 13:36:34.626419 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-g7926" event={"ID":"387c22fd-1a37-4031-a53c-1a5d9e322bed","Type":"ContainerStarted","Data":"354420ec5de1c2a4fa0ce766623d0472250845914103d054c4b82bac7368edb8"} Nov 28 13:36:34 crc kubenswrapper[4631]: I1128 13:36:34.661887 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-g7926" podStartSLOduration=4.667449367 podStartE2EDuration="18.661864874s" podCreationTimestamp="2025-11-28 13:36:16 +0000 UTC" firstStartedPulling="2025-11-28 13:36:19.268052548 +0000 UTC m=+936.075355892" lastFinishedPulling="2025-11-28 13:36:33.262468055 +0000 UTC m=+950.069771399" observedRunningTime="2025-11-28 13:36:34.661799822 +0000 UTC m=+951.469103176" watchObservedRunningTime="2025-11-28 13:36:34.661864874 +0000 UTC m=+951.469168218" Nov 28 13:36:36 crc kubenswrapper[4631]: I1128 13:36:36.352809 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Nov 28 13:36:38 crc kubenswrapper[4631]: I1128 13:36:38.984437 4631 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-7jgdf" podUID="7fc5dcd5-fde5-4a72-8d4a-cda51785e77f" containerName="ovn-controller" probeResult="failure" output=< Nov 28 13:36:38 crc kubenswrapper[4631]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Nov 28 13:36:38 crc kubenswrapper[4631]: > Nov 28 13:36:39 crc kubenswrapper[4631]: I1128 13:36:39.190025 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-lskqw" Nov 28 13:36:39 crc kubenswrapper[4631]: I1128 13:36:39.193983 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-lskqw" Nov 28 13:36:39 crc kubenswrapper[4631]: I1128 13:36:39.442528 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-7jgdf-config-gv897"] Nov 28 13:36:39 crc kubenswrapper[4631]: E1128 13:36:39.442891 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="780a47b4-0078-4f97-a5f5-ad4e1b2015b4" containerName="swift-ring-rebalance" Nov 28 13:36:39 crc kubenswrapper[4631]: I1128 13:36:39.442906 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="780a47b4-0078-4f97-a5f5-ad4e1b2015b4" containerName="swift-ring-rebalance" Nov 28 13:36:39 crc kubenswrapper[4631]: I1128 13:36:39.443099 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="780a47b4-0078-4f97-a5f5-ad4e1b2015b4" containerName="swift-ring-rebalance" Nov 28 13:36:39 crc kubenswrapper[4631]: I1128 13:36:39.443662 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-7jgdf-config-gv897" Nov 28 13:36:39 crc kubenswrapper[4631]: I1128 13:36:39.446250 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Nov 28 13:36:39 crc kubenswrapper[4631]: I1128 13:36:39.550893 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-7jgdf-config-gv897"] Nov 28 13:36:39 crc kubenswrapper[4631]: I1128 13:36:39.557359 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sbf72\" (UniqueName: \"kubernetes.io/projected/72e434fb-dcb5-485f-9236-7e23d77f84fc-kube-api-access-sbf72\") pod \"ovn-controller-7jgdf-config-gv897\" (UID: \"72e434fb-dcb5-485f-9236-7e23d77f84fc\") " pod="openstack/ovn-controller-7jgdf-config-gv897" Nov 28 13:36:39 crc kubenswrapper[4631]: I1128 13:36:39.557849 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/72e434fb-dcb5-485f-9236-7e23d77f84fc-scripts\") pod \"ovn-controller-7jgdf-config-gv897\" (UID: \"72e434fb-dcb5-485f-9236-7e23d77f84fc\") " pod="openstack/ovn-controller-7jgdf-config-gv897" Nov 28 13:36:39 crc kubenswrapper[4631]: I1128 13:36:39.558233 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/72e434fb-dcb5-485f-9236-7e23d77f84fc-var-log-ovn\") pod \"ovn-controller-7jgdf-config-gv897\" (UID: \"72e434fb-dcb5-485f-9236-7e23d77f84fc\") " pod="openstack/ovn-controller-7jgdf-config-gv897" Nov 28 13:36:39 crc kubenswrapper[4631]: I1128 13:36:39.558495 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/72e434fb-dcb5-485f-9236-7e23d77f84fc-var-run\") pod \"ovn-controller-7jgdf-config-gv897\" (UID: \"72e434fb-dcb5-485f-9236-7e23d77f84fc\") " pod="openstack/ovn-controller-7jgdf-config-gv897" Nov 28 13:36:39 crc kubenswrapper[4631]: I1128 13:36:39.558624 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/72e434fb-dcb5-485f-9236-7e23d77f84fc-var-run-ovn\") pod \"ovn-controller-7jgdf-config-gv897\" (UID: \"72e434fb-dcb5-485f-9236-7e23d77f84fc\") " pod="openstack/ovn-controller-7jgdf-config-gv897" Nov 28 13:36:39 crc kubenswrapper[4631]: I1128 13:36:39.558827 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/72e434fb-dcb5-485f-9236-7e23d77f84fc-additional-scripts\") pod \"ovn-controller-7jgdf-config-gv897\" (UID: \"72e434fb-dcb5-485f-9236-7e23d77f84fc\") " pod="openstack/ovn-controller-7jgdf-config-gv897" Nov 28 13:36:39 crc kubenswrapper[4631]: I1128 13:36:39.662631 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sbf72\" (UniqueName: \"kubernetes.io/projected/72e434fb-dcb5-485f-9236-7e23d77f84fc-kube-api-access-sbf72\") pod \"ovn-controller-7jgdf-config-gv897\" (UID: \"72e434fb-dcb5-485f-9236-7e23d77f84fc\") " pod="openstack/ovn-controller-7jgdf-config-gv897" Nov 28 13:36:39 crc kubenswrapper[4631]: I1128 13:36:39.662713 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/72e434fb-dcb5-485f-9236-7e23d77f84fc-scripts\") pod \"ovn-controller-7jgdf-config-gv897\" (UID: \"72e434fb-dcb5-485f-9236-7e23d77f84fc\") " pod="openstack/ovn-controller-7jgdf-config-gv897" Nov 28 13:36:39 crc kubenswrapper[4631]: I1128 13:36:39.662756 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/72e434fb-dcb5-485f-9236-7e23d77f84fc-var-log-ovn\") pod \"ovn-controller-7jgdf-config-gv897\" (UID: \"72e434fb-dcb5-485f-9236-7e23d77f84fc\") " pod="openstack/ovn-controller-7jgdf-config-gv897" Nov 28 13:36:39 crc kubenswrapper[4631]: I1128 13:36:39.662795 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/72e434fb-dcb5-485f-9236-7e23d77f84fc-var-run\") pod \"ovn-controller-7jgdf-config-gv897\" (UID: \"72e434fb-dcb5-485f-9236-7e23d77f84fc\") " pod="openstack/ovn-controller-7jgdf-config-gv897" Nov 28 13:36:39 crc kubenswrapper[4631]: I1128 13:36:39.662816 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/72e434fb-dcb5-485f-9236-7e23d77f84fc-var-run-ovn\") pod \"ovn-controller-7jgdf-config-gv897\" (UID: \"72e434fb-dcb5-485f-9236-7e23d77f84fc\") " pod="openstack/ovn-controller-7jgdf-config-gv897" Nov 28 13:36:39 crc kubenswrapper[4631]: I1128 13:36:39.662851 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/72e434fb-dcb5-485f-9236-7e23d77f84fc-additional-scripts\") pod \"ovn-controller-7jgdf-config-gv897\" (UID: \"72e434fb-dcb5-485f-9236-7e23d77f84fc\") " pod="openstack/ovn-controller-7jgdf-config-gv897" Nov 28 13:36:39 crc kubenswrapper[4631]: I1128 13:36:39.664052 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/72e434fb-dcb5-485f-9236-7e23d77f84fc-var-log-ovn\") pod \"ovn-controller-7jgdf-config-gv897\" (UID: \"72e434fb-dcb5-485f-9236-7e23d77f84fc\") " pod="openstack/ovn-controller-7jgdf-config-gv897" Nov 28 13:36:39 crc kubenswrapper[4631]: I1128 13:36:39.664105 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/72e434fb-dcb5-485f-9236-7e23d77f84fc-var-run\") pod \"ovn-controller-7jgdf-config-gv897\" (UID: \"72e434fb-dcb5-485f-9236-7e23d77f84fc\") " pod="openstack/ovn-controller-7jgdf-config-gv897" Nov 28 13:36:39 crc kubenswrapper[4631]: I1128 13:36:39.664142 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/72e434fb-dcb5-485f-9236-7e23d77f84fc-var-run-ovn\") pod \"ovn-controller-7jgdf-config-gv897\" (UID: \"72e434fb-dcb5-485f-9236-7e23d77f84fc\") " pod="openstack/ovn-controller-7jgdf-config-gv897" Nov 28 13:36:39 crc kubenswrapper[4631]: I1128 13:36:39.664833 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/72e434fb-dcb5-485f-9236-7e23d77f84fc-additional-scripts\") pod \"ovn-controller-7jgdf-config-gv897\" (UID: \"72e434fb-dcb5-485f-9236-7e23d77f84fc\") " pod="openstack/ovn-controller-7jgdf-config-gv897" Nov 28 13:36:39 crc kubenswrapper[4631]: I1128 13:36:39.666417 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/72e434fb-dcb5-485f-9236-7e23d77f84fc-scripts\") pod \"ovn-controller-7jgdf-config-gv897\" (UID: \"72e434fb-dcb5-485f-9236-7e23d77f84fc\") " pod="openstack/ovn-controller-7jgdf-config-gv897" Nov 28 13:36:39 crc kubenswrapper[4631]: I1128 13:36:39.715092 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sbf72\" (UniqueName: \"kubernetes.io/projected/72e434fb-dcb5-485f-9236-7e23d77f84fc-kube-api-access-sbf72\") pod \"ovn-controller-7jgdf-config-gv897\" (UID: \"72e434fb-dcb5-485f-9236-7e23d77f84fc\") " pod="openstack/ovn-controller-7jgdf-config-gv897" Nov 28 13:36:39 crc kubenswrapper[4631]: I1128 13:36:39.763789 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-7jgdf-config-gv897" Nov 28 13:36:40 crc kubenswrapper[4631]: I1128 13:36:40.324888 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-7jgdf-config-gv897"] Nov 28 13:36:40 crc kubenswrapper[4631]: I1128 13:36:40.683344 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-7jgdf-config-gv897" event={"ID":"72e434fb-dcb5-485f-9236-7e23d77f84fc","Type":"ContainerStarted","Data":"82a190b758af94141a2fd7a1f3593be555d457e8214d51ba9ff702989f342a84"} Nov 28 13:36:41 crc kubenswrapper[4631]: I1128 13:36:41.801143 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/c3e1f6fb-1845-4a39-8eca-1262da1ff281-etc-swift\") pod \"swift-storage-0\" (UID: \"c3e1f6fb-1845-4a39-8eca-1262da1ff281\") " pod="openstack/swift-storage-0" Nov 28 13:36:41 crc kubenswrapper[4631]: I1128 13:36:41.812402 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/c3e1f6fb-1845-4a39-8eca-1262da1ff281-etc-swift\") pod \"swift-storage-0\" (UID: \"c3e1f6fb-1845-4a39-8eca-1262da1ff281\") " pod="openstack/swift-storage-0" Nov 28 13:36:41 crc kubenswrapper[4631]: I1128 13:36:41.930028 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Nov 28 13:36:42 crc kubenswrapper[4631]: I1128 13:36:42.756438 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Nov 28 13:36:42 crc kubenswrapper[4631]: W1128 13:36:42.768626 4631 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc3e1f6fb_1845_4a39_8eca_1262da1ff281.slice/crio-8110b1a55bcc7c32c286951a050f165c9bc54f0343949898998396f1ba4cf998 WatchSource:0}: Error finding container 8110b1a55bcc7c32c286951a050f165c9bc54f0343949898998396f1ba4cf998: Status 404 returned error can't find the container with id 8110b1a55bcc7c32c286951a050f165c9bc54f0343949898998396f1ba4cf998 Nov 28 13:36:43 crc kubenswrapper[4631]: I1128 13:36:43.714358 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-7jgdf-config-gv897" event={"ID":"72e434fb-dcb5-485f-9236-7e23d77f84fc","Type":"ContainerStarted","Data":"12413999ef935ad5f2fe03a148a79c915c22824ce8578cc03526415f1f365612"} Nov 28 13:36:43 crc kubenswrapper[4631]: I1128 13:36:43.717145 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"c3e1f6fb-1845-4a39-8eca-1262da1ff281","Type":"ContainerStarted","Data":"8110b1a55bcc7c32c286951a050f165c9bc54f0343949898998396f1ba4cf998"} Nov 28 13:36:43 crc kubenswrapper[4631]: I1128 13:36:43.745822 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-7jgdf-config-gv897" podStartSLOduration=4.745802556 podStartE2EDuration="4.745802556s" podCreationTimestamp="2025-11-28 13:36:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 13:36:43.735981681 +0000 UTC m=+960.543285025" watchObservedRunningTime="2025-11-28 13:36:43.745802556 +0000 UTC m=+960.553105900" Nov 28 13:36:43 crc kubenswrapper[4631]: I1128 13:36:43.992599 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-7jgdf" Nov 28 13:36:44 crc kubenswrapper[4631]: I1128 13:36:44.731483 4631 generic.go:334] "Generic (PLEG): container finished" podID="72e434fb-dcb5-485f-9236-7e23d77f84fc" containerID="12413999ef935ad5f2fe03a148a79c915c22824ce8578cc03526415f1f365612" exitCode=0 Nov 28 13:36:44 crc kubenswrapper[4631]: I1128 13:36:44.731563 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-7jgdf-config-gv897" event={"ID":"72e434fb-dcb5-485f-9236-7e23d77f84fc","Type":"ContainerDied","Data":"12413999ef935ad5f2fe03a148a79c915c22824ce8578cc03526415f1f365612"} Nov 28 13:36:45 crc kubenswrapper[4631]: I1128 13:36:45.745138 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"c3e1f6fb-1845-4a39-8eca-1262da1ff281","Type":"ContainerStarted","Data":"b781aca45814ff27f8c42af2ec6957bbf2381679c05da70b65f5722c7d668543"} Nov 28 13:36:45 crc kubenswrapper[4631]: I1128 13:36:45.745856 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"c3e1f6fb-1845-4a39-8eca-1262da1ff281","Type":"ContainerStarted","Data":"dd5f6acb92345e1becc3995f556c4ef30b7a2392af851a961b05ad967ddc33d1"} Nov 28 13:36:45 crc kubenswrapper[4631]: I1128 13:36:45.748958 4631 generic.go:334] "Generic (PLEG): container finished" podID="387c22fd-1a37-4031-a53c-1a5d9e322bed" containerID="354420ec5de1c2a4fa0ce766623d0472250845914103d054c4b82bac7368edb8" exitCode=0 Nov 28 13:36:45 crc kubenswrapper[4631]: I1128 13:36:45.749154 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-g7926" event={"ID":"387c22fd-1a37-4031-a53c-1a5d9e322bed","Type":"ContainerDied","Data":"354420ec5de1c2a4fa0ce766623d0472250845914103d054c4b82bac7368edb8"} Nov 28 13:36:46 crc kubenswrapper[4631]: I1128 13:36:46.646343 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-7jgdf-config-gv897" Nov 28 13:36:46 crc kubenswrapper[4631]: I1128 13:36:46.759815 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-7jgdf-config-gv897" event={"ID":"72e434fb-dcb5-485f-9236-7e23d77f84fc","Type":"ContainerDied","Data":"82a190b758af94141a2fd7a1f3593be555d457e8214d51ba9ff702989f342a84"} Nov 28 13:36:46 crc kubenswrapper[4631]: I1128 13:36:46.759859 4631 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="82a190b758af94141a2fd7a1f3593be555d457e8214d51ba9ff702989f342a84" Nov 28 13:36:46 crc kubenswrapper[4631]: I1128 13:36:46.759940 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-7jgdf-config-gv897" Nov 28 13:36:46 crc kubenswrapper[4631]: I1128 13:36:46.762232 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"c3e1f6fb-1845-4a39-8eca-1262da1ff281","Type":"ContainerStarted","Data":"89b1a32f794b966f7f89d0d2fc78436aed6c2d99a83c2fdcd694bc3c76db234a"} Nov 28 13:36:46 crc kubenswrapper[4631]: I1128 13:36:46.762273 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"c3e1f6fb-1845-4a39-8eca-1262da1ff281","Type":"ContainerStarted","Data":"35d17b9c20d144f3a3e8681f5bfa1cddf9cbb8c3d617a71e845162fd48ad8a52"} Nov 28 13:36:46 crc kubenswrapper[4631]: I1128 13:36:46.836793 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/72e434fb-dcb5-485f-9236-7e23d77f84fc-var-log-ovn\") pod \"72e434fb-dcb5-485f-9236-7e23d77f84fc\" (UID: \"72e434fb-dcb5-485f-9236-7e23d77f84fc\") " Nov 28 13:36:46 crc kubenswrapper[4631]: I1128 13:36:46.836876 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/72e434fb-dcb5-485f-9236-7e23d77f84fc-var-run-ovn\") pod \"72e434fb-dcb5-485f-9236-7e23d77f84fc\" (UID: \"72e434fb-dcb5-485f-9236-7e23d77f84fc\") " Nov 28 13:36:46 crc kubenswrapper[4631]: I1128 13:36:46.837017 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/72e434fb-dcb5-485f-9236-7e23d77f84fc-var-run\") pod \"72e434fb-dcb5-485f-9236-7e23d77f84fc\" (UID: \"72e434fb-dcb5-485f-9236-7e23d77f84fc\") " Nov 28 13:36:46 crc kubenswrapper[4631]: I1128 13:36:46.837056 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sbf72\" (UniqueName: \"kubernetes.io/projected/72e434fb-dcb5-485f-9236-7e23d77f84fc-kube-api-access-sbf72\") pod \"72e434fb-dcb5-485f-9236-7e23d77f84fc\" (UID: \"72e434fb-dcb5-485f-9236-7e23d77f84fc\") " Nov 28 13:36:46 crc kubenswrapper[4631]: I1128 13:36:46.837080 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/72e434fb-dcb5-485f-9236-7e23d77f84fc-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "72e434fb-dcb5-485f-9236-7e23d77f84fc" (UID: "72e434fb-dcb5-485f-9236-7e23d77f84fc"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 13:36:46 crc kubenswrapper[4631]: I1128 13:36:46.837093 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/72e434fb-dcb5-485f-9236-7e23d77f84fc-scripts\") pod \"72e434fb-dcb5-485f-9236-7e23d77f84fc\" (UID: \"72e434fb-dcb5-485f-9236-7e23d77f84fc\") " Nov 28 13:36:46 crc kubenswrapper[4631]: I1128 13:36:46.837080 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/72e434fb-dcb5-485f-9236-7e23d77f84fc-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "72e434fb-dcb5-485f-9236-7e23d77f84fc" (UID: "72e434fb-dcb5-485f-9236-7e23d77f84fc"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 13:36:46 crc kubenswrapper[4631]: I1128 13:36:46.837108 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/72e434fb-dcb5-485f-9236-7e23d77f84fc-var-run" (OuterVolumeSpecName: "var-run") pod "72e434fb-dcb5-485f-9236-7e23d77f84fc" (UID: "72e434fb-dcb5-485f-9236-7e23d77f84fc"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 13:36:46 crc kubenswrapper[4631]: I1128 13:36:46.837183 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/72e434fb-dcb5-485f-9236-7e23d77f84fc-additional-scripts\") pod \"72e434fb-dcb5-485f-9236-7e23d77f84fc\" (UID: \"72e434fb-dcb5-485f-9236-7e23d77f84fc\") " Nov 28 13:36:46 crc kubenswrapper[4631]: I1128 13:36:46.837696 4631 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/72e434fb-dcb5-485f-9236-7e23d77f84fc-var-run\") on node \"crc\" DevicePath \"\"" Nov 28 13:36:46 crc kubenswrapper[4631]: I1128 13:36:46.837721 4631 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/72e434fb-dcb5-485f-9236-7e23d77f84fc-var-log-ovn\") on node \"crc\" DevicePath \"\"" Nov 28 13:36:46 crc kubenswrapper[4631]: I1128 13:36:46.837736 4631 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/72e434fb-dcb5-485f-9236-7e23d77f84fc-var-run-ovn\") on node \"crc\" DevicePath \"\"" Nov 28 13:36:46 crc kubenswrapper[4631]: I1128 13:36:46.837948 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/72e434fb-dcb5-485f-9236-7e23d77f84fc-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "72e434fb-dcb5-485f-9236-7e23d77f84fc" (UID: "72e434fb-dcb5-485f-9236-7e23d77f84fc"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 13:36:46 crc kubenswrapper[4631]: I1128 13:36:46.838181 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/72e434fb-dcb5-485f-9236-7e23d77f84fc-scripts" (OuterVolumeSpecName: "scripts") pod "72e434fb-dcb5-485f-9236-7e23d77f84fc" (UID: "72e434fb-dcb5-485f-9236-7e23d77f84fc"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 13:36:46 crc kubenswrapper[4631]: I1128 13:36:46.854594 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/72e434fb-dcb5-485f-9236-7e23d77f84fc-kube-api-access-sbf72" (OuterVolumeSpecName: "kube-api-access-sbf72") pod "72e434fb-dcb5-485f-9236-7e23d77f84fc" (UID: "72e434fb-dcb5-485f-9236-7e23d77f84fc"). InnerVolumeSpecName "kube-api-access-sbf72". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:36:46 crc kubenswrapper[4631]: I1128 13:36:46.939147 4631 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/72e434fb-dcb5-485f-9236-7e23d77f84fc-additional-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 13:36:46 crc kubenswrapper[4631]: I1128 13:36:46.939590 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sbf72\" (UniqueName: \"kubernetes.io/projected/72e434fb-dcb5-485f-9236-7e23d77f84fc-kube-api-access-sbf72\") on node \"crc\" DevicePath \"\"" Nov 28 13:36:46 crc kubenswrapper[4631]: I1128 13:36:46.939610 4631 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/72e434fb-dcb5-485f-9236-7e23d77f84fc-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 13:36:47 crc kubenswrapper[4631]: I1128 13:36:47.167686 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-g7926" Nov 28 13:36:47 crc kubenswrapper[4631]: I1128 13:36:47.347319 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/387c22fd-1a37-4031-a53c-1a5d9e322bed-config-data\") pod \"387c22fd-1a37-4031-a53c-1a5d9e322bed\" (UID: \"387c22fd-1a37-4031-a53c-1a5d9e322bed\") " Nov 28 13:36:47 crc kubenswrapper[4631]: I1128 13:36:47.347419 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/387c22fd-1a37-4031-a53c-1a5d9e322bed-db-sync-config-data\") pod \"387c22fd-1a37-4031-a53c-1a5d9e322bed\" (UID: \"387c22fd-1a37-4031-a53c-1a5d9e322bed\") " Nov 28 13:36:47 crc kubenswrapper[4631]: I1128 13:36:47.347512 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/387c22fd-1a37-4031-a53c-1a5d9e322bed-combined-ca-bundle\") pod \"387c22fd-1a37-4031-a53c-1a5d9e322bed\" (UID: \"387c22fd-1a37-4031-a53c-1a5d9e322bed\") " Nov 28 13:36:47 crc kubenswrapper[4631]: I1128 13:36:47.347642 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9g5xx\" (UniqueName: \"kubernetes.io/projected/387c22fd-1a37-4031-a53c-1a5d9e322bed-kube-api-access-9g5xx\") pod \"387c22fd-1a37-4031-a53c-1a5d9e322bed\" (UID: \"387c22fd-1a37-4031-a53c-1a5d9e322bed\") " Nov 28 13:36:47 crc kubenswrapper[4631]: I1128 13:36:47.352773 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/387c22fd-1a37-4031-a53c-1a5d9e322bed-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "387c22fd-1a37-4031-a53c-1a5d9e322bed" (UID: "387c22fd-1a37-4031-a53c-1a5d9e322bed"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:36:47 crc kubenswrapper[4631]: I1128 13:36:47.353325 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/387c22fd-1a37-4031-a53c-1a5d9e322bed-kube-api-access-9g5xx" (OuterVolumeSpecName: "kube-api-access-9g5xx") pod "387c22fd-1a37-4031-a53c-1a5d9e322bed" (UID: "387c22fd-1a37-4031-a53c-1a5d9e322bed"). InnerVolumeSpecName "kube-api-access-9g5xx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:36:47 crc kubenswrapper[4631]: I1128 13:36:47.384191 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/387c22fd-1a37-4031-a53c-1a5d9e322bed-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "387c22fd-1a37-4031-a53c-1a5d9e322bed" (UID: "387c22fd-1a37-4031-a53c-1a5d9e322bed"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:36:47 crc kubenswrapper[4631]: I1128 13:36:47.389322 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/387c22fd-1a37-4031-a53c-1a5d9e322bed-config-data" (OuterVolumeSpecName: "config-data") pod "387c22fd-1a37-4031-a53c-1a5d9e322bed" (UID: "387c22fd-1a37-4031-a53c-1a5d9e322bed"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:36:47 crc kubenswrapper[4631]: I1128 13:36:47.450971 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9g5xx\" (UniqueName: \"kubernetes.io/projected/387c22fd-1a37-4031-a53c-1a5d9e322bed-kube-api-access-9g5xx\") on node \"crc\" DevicePath \"\"" Nov 28 13:36:47 crc kubenswrapper[4631]: I1128 13:36:47.451018 4631 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/387c22fd-1a37-4031-a53c-1a5d9e322bed-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 13:36:47 crc kubenswrapper[4631]: I1128 13:36:47.451030 4631 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/387c22fd-1a37-4031-a53c-1a5d9e322bed-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 13:36:47 crc kubenswrapper[4631]: I1128 13:36:47.451041 4631 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/387c22fd-1a37-4031-a53c-1a5d9e322bed-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 13:36:47 crc kubenswrapper[4631]: I1128 13:36:47.772966 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-g7926" event={"ID":"387c22fd-1a37-4031-a53c-1a5d9e322bed","Type":"ContainerDied","Data":"0c145e40542d56fae99ece808ecb6f3fe1b858b045c8401bf0b58516d854fb81"} Nov 28 13:36:47 crc kubenswrapper[4631]: I1128 13:36:47.773025 4631 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0c145e40542d56fae99ece808ecb6f3fe1b858b045c8401bf0b58516d854fb81" Nov 28 13:36:47 crc kubenswrapper[4631]: I1128 13:36:47.773069 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-g7926" Nov 28 13:36:47 crc kubenswrapper[4631]: I1128 13:36:47.783194 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-7jgdf-config-gv897"] Nov 28 13:36:47 crc kubenswrapper[4631]: I1128 13:36:47.793185 4631 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-7jgdf-config-gv897"] Nov 28 13:36:48 crc kubenswrapper[4631]: I1128 13:36:48.217016 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5b946c75cc-rnp88"] Nov 28 13:36:48 crc kubenswrapper[4631]: E1128 13:36:48.218048 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="387c22fd-1a37-4031-a53c-1a5d9e322bed" containerName="glance-db-sync" Nov 28 13:36:48 crc kubenswrapper[4631]: I1128 13:36:48.218075 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="387c22fd-1a37-4031-a53c-1a5d9e322bed" containerName="glance-db-sync" Nov 28 13:36:48 crc kubenswrapper[4631]: E1128 13:36:48.218107 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="72e434fb-dcb5-485f-9236-7e23d77f84fc" containerName="ovn-config" Nov 28 13:36:48 crc kubenswrapper[4631]: I1128 13:36:48.218116 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="72e434fb-dcb5-485f-9236-7e23d77f84fc" containerName="ovn-config" Nov 28 13:36:48 crc kubenswrapper[4631]: I1128 13:36:48.218687 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="72e434fb-dcb5-485f-9236-7e23d77f84fc" containerName="ovn-config" Nov 28 13:36:48 crc kubenswrapper[4631]: I1128 13:36:48.218733 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="387c22fd-1a37-4031-a53c-1a5d9e322bed" containerName="glance-db-sync" Nov 28 13:36:48 crc kubenswrapper[4631]: I1128 13:36:48.220226 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b946c75cc-rnp88" Nov 28 13:36:48 crc kubenswrapper[4631]: I1128 13:36:48.265554 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5b946c75cc-rnp88"] Nov 28 13:36:48 crc kubenswrapper[4631]: I1128 13:36:48.272613 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2sksh\" (UniqueName: \"kubernetes.io/projected/684ed85e-280f-4cf7-b528-45284096b772-kube-api-access-2sksh\") pod \"dnsmasq-dns-5b946c75cc-rnp88\" (UID: \"684ed85e-280f-4cf7-b528-45284096b772\") " pod="openstack/dnsmasq-dns-5b946c75cc-rnp88" Nov 28 13:36:48 crc kubenswrapper[4631]: I1128 13:36:48.272686 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/684ed85e-280f-4cf7-b528-45284096b772-dns-svc\") pod \"dnsmasq-dns-5b946c75cc-rnp88\" (UID: \"684ed85e-280f-4cf7-b528-45284096b772\") " pod="openstack/dnsmasq-dns-5b946c75cc-rnp88" Nov 28 13:36:48 crc kubenswrapper[4631]: I1128 13:36:48.272727 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/684ed85e-280f-4cf7-b528-45284096b772-ovsdbserver-sb\") pod \"dnsmasq-dns-5b946c75cc-rnp88\" (UID: \"684ed85e-280f-4cf7-b528-45284096b772\") " pod="openstack/dnsmasq-dns-5b946c75cc-rnp88" Nov 28 13:36:48 crc kubenswrapper[4631]: I1128 13:36:48.272750 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/684ed85e-280f-4cf7-b528-45284096b772-ovsdbserver-nb\") pod \"dnsmasq-dns-5b946c75cc-rnp88\" (UID: \"684ed85e-280f-4cf7-b528-45284096b772\") " pod="openstack/dnsmasq-dns-5b946c75cc-rnp88" Nov 28 13:36:48 crc kubenswrapper[4631]: I1128 13:36:48.272774 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/684ed85e-280f-4cf7-b528-45284096b772-config\") pod \"dnsmasq-dns-5b946c75cc-rnp88\" (UID: \"684ed85e-280f-4cf7-b528-45284096b772\") " pod="openstack/dnsmasq-dns-5b946c75cc-rnp88" Nov 28 13:36:48 crc kubenswrapper[4631]: I1128 13:36:48.373848 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/684ed85e-280f-4cf7-b528-45284096b772-ovsdbserver-nb\") pod \"dnsmasq-dns-5b946c75cc-rnp88\" (UID: \"684ed85e-280f-4cf7-b528-45284096b772\") " pod="openstack/dnsmasq-dns-5b946c75cc-rnp88" Nov 28 13:36:48 crc kubenswrapper[4631]: I1128 13:36:48.373909 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/684ed85e-280f-4cf7-b528-45284096b772-config\") pod \"dnsmasq-dns-5b946c75cc-rnp88\" (UID: \"684ed85e-280f-4cf7-b528-45284096b772\") " pod="openstack/dnsmasq-dns-5b946c75cc-rnp88" Nov 28 13:36:48 crc kubenswrapper[4631]: I1128 13:36:48.374008 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2sksh\" (UniqueName: \"kubernetes.io/projected/684ed85e-280f-4cf7-b528-45284096b772-kube-api-access-2sksh\") pod \"dnsmasq-dns-5b946c75cc-rnp88\" (UID: \"684ed85e-280f-4cf7-b528-45284096b772\") " pod="openstack/dnsmasq-dns-5b946c75cc-rnp88" Nov 28 13:36:48 crc kubenswrapper[4631]: I1128 13:36:48.374422 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/684ed85e-280f-4cf7-b528-45284096b772-dns-svc\") pod \"dnsmasq-dns-5b946c75cc-rnp88\" (UID: \"684ed85e-280f-4cf7-b528-45284096b772\") " pod="openstack/dnsmasq-dns-5b946c75cc-rnp88" Nov 28 13:36:48 crc kubenswrapper[4631]: I1128 13:36:48.374964 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/684ed85e-280f-4cf7-b528-45284096b772-ovsdbserver-nb\") pod \"dnsmasq-dns-5b946c75cc-rnp88\" (UID: \"684ed85e-280f-4cf7-b528-45284096b772\") " pod="openstack/dnsmasq-dns-5b946c75cc-rnp88" Nov 28 13:36:48 crc kubenswrapper[4631]: I1128 13:36:48.375043 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/684ed85e-280f-4cf7-b528-45284096b772-config\") pod \"dnsmasq-dns-5b946c75cc-rnp88\" (UID: \"684ed85e-280f-4cf7-b528-45284096b772\") " pod="openstack/dnsmasq-dns-5b946c75cc-rnp88" Nov 28 13:36:48 crc kubenswrapper[4631]: I1128 13:36:48.375234 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/684ed85e-280f-4cf7-b528-45284096b772-dns-svc\") pod \"dnsmasq-dns-5b946c75cc-rnp88\" (UID: \"684ed85e-280f-4cf7-b528-45284096b772\") " pod="openstack/dnsmasq-dns-5b946c75cc-rnp88" Nov 28 13:36:48 crc kubenswrapper[4631]: I1128 13:36:48.375804 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/684ed85e-280f-4cf7-b528-45284096b772-ovsdbserver-sb\") pod \"dnsmasq-dns-5b946c75cc-rnp88\" (UID: \"684ed85e-280f-4cf7-b528-45284096b772\") " pod="openstack/dnsmasq-dns-5b946c75cc-rnp88" Nov 28 13:36:48 crc kubenswrapper[4631]: I1128 13:36:48.376274 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/684ed85e-280f-4cf7-b528-45284096b772-ovsdbserver-sb\") pod \"dnsmasq-dns-5b946c75cc-rnp88\" (UID: \"684ed85e-280f-4cf7-b528-45284096b772\") " pod="openstack/dnsmasq-dns-5b946c75cc-rnp88" Nov 28 13:36:48 crc kubenswrapper[4631]: I1128 13:36:48.399562 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2sksh\" (UniqueName: \"kubernetes.io/projected/684ed85e-280f-4cf7-b528-45284096b772-kube-api-access-2sksh\") pod \"dnsmasq-dns-5b946c75cc-rnp88\" (UID: \"684ed85e-280f-4cf7-b528-45284096b772\") " pod="openstack/dnsmasq-dns-5b946c75cc-rnp88" Nov 28 13:36:48 crc kubenswrapper[4631]: I1128 13:36:48.549328 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b946c75cc-rnp88" Nov 28 13:36:49 crc kubenswrapper[4631]: I1128 13:36:49.139035 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5b946c75cc-rnp88"] Nov 28 13:36:49 crc kubenswrapper[4631]: I1128 13:36:49.525727 4631 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="72e434fb-dcb5-485f-9236-7e23d77f84fc" path="/var/lib/kubelet/pods/72e434fb-dcb5-485f-9236-7e23d77f84fc/volumes" Nov 28 13:36:49 crc kubenswrapper[4631]: I1128 13:36:49.635826 4631 patch_prober.go:28] interesting pod/machine-config-daemon-47sxw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 28 13:36:49 crc kubenswrapper[4631]: I1128 13:36:49.636151 4631 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" podUID="8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 28 13:36:49 crc kubenswrapper[4631]: I1128 13:36:49.803857 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"c3e1f6fb-1845-4a39-8eca-1262da1ff281","Type":"ContainerStarted","Data":"9a516b6a68b36ab7ad0f195bbad4cc4d77f6ea135fd2fdd5fc0b8ac1bd8327d8"} Nov 28 13:36:49 crc kubenswrapper[4631]: I1128 13:36:49.804188 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"c3e1f6fb-1845-4a39-8eca-1262da1ff281","Type":"ContainerStarted","Data":"fb67fd72d6c0086a0bf3b2688825cbe2ee21b1e37a5080028d14dd951fc13126"} Nov 28 13:36:49 crc kubenswrapper[4631]: I1128 13:36:49.804261 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"c3e1f6fb-1845-4a39-8eca-1262da1ff281","Type":"ContainerStarted","Data":"53793f98ea03e8745e93fb9e646c195e810ca925e916378a90730d507024382e"} Nov 28 13:36:49 crc kubenswrapper[4631]: I1128 13:36:49.804341 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"c3e1f6fb-1845-4a39-8eca-1262da1ff281","Type":"ContainerStarted","Data":"e01506fb0045196e976fca2652da56e1c95c9c78e9016a22ba7d5ad5b7cff13d"} Nov 28 13:36:49 crc kubenswrapper[4631]: I1128 13:36:49.805885 4631 generic.go:334] "Generic (PLEG): container finished" podID="684ed85e-280f-4cf7-b528-45284096b772" containerID="e88aa3e2ea1bd4be9bab880af18e21ed1a97e0e89a7e7d8d890346fcb7edd576" exitCode=0 Nov 28 13:36:49 crc kubenswrapper[4631]: I1128 13:36:49.805973 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b946c75cc-rnp88" event={"ID":"684ed85e-280f-4cf7-b528-45284096b772","Type":"ContainerDied","Data":"e88aa3e2ea1bd4be9bab880af18e21ed1a97e0e89a7e7d8d890346fcb7edd576"} Nov 28 13:36:49 crc kubenswrapper[4631]: I1128 13:36:49.806037 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b946c75cc-rnp88" event={"ID":"684ed85e-280f-4cf7-b528-45284096b772","Type":"ContainerStarted","Data":"923378e21e4a6ca1879e802a4ec08326506801a1c2f51b1d8526285e110919b8"} Nov 28 13:36:50 crc kubenswrapper[4631]: I1128 13:36:50.821410 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b946c75cc-rnp88" event={"ID":"684ed85e-280f-4cf7-b528-45284096b772","Type":"ContainerStarted","Data":"d2f07cceea9a52630a5bf19886f3266f82037c6d0807528ab6e1a449f221a276"} Nov 28 13:36:50 crc kubenswrapper[4631]: I1128 13:36:50.821858 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5b946c75cc-rnp88" Nov 28 13:36:51 crc kubenswrapper[4631]: I1128 13:36:51.832312 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"c3e1f6fb-1845-4a39-8eca-1262da1ff281","Type":"ContainerStarted","Data":"42937d0c93a457b9d0a7ec9e91028751831476669587bdbf889391c92e969ab2"} Nov 28 13:36:52 crc kubenswrapper[4631]: I1128 13:36:52.398750 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Nov 28 13:36:52 crc kubenswrapper[4631]: I1128 13:36:52.424684 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5b946c75cc-rnp88" podStartSLOduration=4.42465766 podStartE2EDuration="4.42465766s" podCreationTimestamp="2025-11-28 13:36:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 13:36:50.850585807 +0000 UTC m=+967.657889161" watchObservedRunningTime="2025-11-28 13:36:52.42465766 +0000 UTC m=+969.231961004" Nov 28 13:36:52 crc kubenswrapper[4631]: I1128 13:36:52.853468 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Nov 28 13:36:52 crc kubenswrapper[4631]: I1128 13:36:52.862042 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"c3e1f6fb-1845-4a39-8eca-1262da1ff281","Type":"ContainerStarted","Data":"2639cdccfbf87135b538229bdd613b54187a650b3bc647636e435ee890a4f7ca"} Nov 28 13:36:52 crc kubenswrapper[4631]: I1128 13:36:52.862078 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"c3e1f6fb-1845-4a39-8eca-1262da1ff281","Type":"ContainerStarted","Data":"866833bf0884031a6ffb88e24abed9e1f8742b10caa79bf4ab35f07761aa0423"} Nov 28 13:36:52 crc kubenswrapper[4631]: I1128 13:36:52.874140 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-wrd26"] Nov 28 13:36:52 crc kubenswrapper[4631]: I1128 13:36:52.875327 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-wrd26" Nov 28 13:36:52 crc kubenswrapper[4631]: I1128 13:36:52.975174 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/98564bcf-8a3f-4324-af81-bafdc2ed2934-operator-scripts\") pod \"cinder-db-create-wrd26\" (UID: \"98564bcf-8a3f-4324-af81-bafdc2ed2934\") " pod="openstack/cinder-db-create-wrd26" Nov 28 13:36:52 crc kubenswrapper[4631]: I1128 13:36:52.975319 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wvns5\" (UniqueName: \"kubernetes.io/projected/98564bcf-8a3f-4324-af81-bafdc2ed2934-kube-api-access-wvns5\") pod \"cinder-db-create-wrd26\" (UID: \"98564bcf-8a3f-4324-af81-bafdc2ed2934\") " pod="openstack/cinder-db-create-wrd26" Nov 28 13:36:52 crc kubenswrapper[4631]: I1128 13:36:52.979324 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-wrd26"] Nov 28 13:36:53 crc kubenswrapper[4631]: I1128 13:36:53.077379 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wvns5\" (UniqueName: \"kubernetes.io/projected/98564bcf-8a3f-4324-af81-bafdc2ed2934-kube-api-access-wvns5\") pod \"cinder-db-create-wrd26\" (UID: \"98564bcf-8a3f-4324-af81-bafdc2ed2934\") " pod="openstack/cinder-db-create-wrd26" Nov 28 13:36:53 crc kubenswrapper[4631]: I1128 13:36:53.077515 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/98564bcf-8a3f-4324-af81-bafdc2ed2934-operator-scripts\") pod \"cinder-db-create-wrd26\" (UID: \"98564bcf-8a3f-4324-af81-bafdc2ed2934\") " pod="openstack/cinder-db-create-wrd26" Nov 28 13:36:53 crc kubenswrapper[4631]: I1128 13:36:53.078226 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/98564bcf-8a3f-4324-af81-bafdc2ed2934-operator-scripts\") pod \"cinder-db-create-wrd26\" (UID: \"98564bcf-8a3f-4324-af81-bafdc2ed2934\") " pod="openstack/cinder-db-create-wrd26" Nov 28 13:36:53 crc kubenswrapper[4631]: I1128 13:36:53.087488 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-m4zmz"] Nov 28 13:36:53 crc kubenswrapper[4631]: I1128 13:36:53.088861 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-m4zmz" Nov 28 13:36:53 crc kubenswrapper[4631]: I1128 13:36:53.128394 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wvns5\" (UniqueName: \"kubernetes.io/projected/98564bcf-8a3f-4324-af81-bafdc2ed2934-kube-api-access-wvns5\") pod \"cinder-db-create-wrd26\" (UID: \"98564bcf-8a3f-4324-af81-bafdc2ed2934\") " pod="openstack/cinder-db-create-wrd26" Nov 28 13:36:53 crc kubenswrapper[4631]: I1128 13:36:53.146599 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-m4zmz"] Nov 28 13:36:53 crc kubenswrapper[4631]: I1128 13:36:53.158975 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-8a8c-account-create-update-lskqg"] Nov 28 13:36:53 crc kubenswrapper[4631]: I1128 13:36:53.170343 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-8a8c-account-create-update-lskqg" Nov 28 13:36:53 crc kubenswrapper[4631]: I1128 13:36:53.179094 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ss6q5\" (UniqueName: \"kubernetes.io/projected/95d067ec-b8ae-4286-a35f-eb431d025ac9-kube-api-access-ss6q5\") pod \"barbican-db-create-m4zmz\" (UID: \"95d067ec-b8ae-4286-a35f-eb431d025ac9\") " pod="openstack/barbican-db-create-m4zmz" Nov 28 13:36:53 crc kubenswrapper[4631]: I1128 13:36:53.179256 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/95d067ec-b8ae-4286-a35f-eb431d025ac9-operator-scripts\") pod \"barbican-db-create-m4zmz\" (UID: \"95d067ec-b8ae-4286-a35f-eb431d025ac9\") " pod="openstack/barbican-db-create-m4zmz" Nov 28 13:36:53 crc kubenswrapper[4631]: I1128 13:36:53.204738 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Nov 28 13:36:53 crc kubenswrapper[4631]: I1128 13:36:53.205079 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-wrd26" Nov 28 13:36:53 crc kubenswrapper[4631]: I1128 13:36:53.239730 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-8a8c-account-create-update-lskqg"] Nov 28 13:36:53 crc kubenswrapper[4631]: I1128 13:36:53.287012 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/127100c3-8f48-4ddf-a289-f08f3cdf0d8a-operator-scripts\") pod \"barbican-8a8c-account-create-update-lskqg\" (UID: \"127100c3-8f48-4ddf-a289-f08f3cdf0d8a\") " pod="openstack/barbican-8a8c-account-create-update-lskqg" Nov 28 13:36:53 crc kubenswrapper[4631]: I1128 13:36:53.287140 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ss6q5\" (UniqueName: \"kubernetes.io/projected/95d067ec-b8ae-4286-a35f-eb431d025ac9-kube-api-access-ss6q5\") pod \"barbican-db-create-m4zmz\" (UID: \"95d067ec-b8ae-4286-a35f-eb431d025ac9\") " pod="openstack/barbican-db-create-m4zmz" Nov 28 13:36:53 crc kubenswrapper[4631]: I1128 13:36:53.287232 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/95d067ec-b8ae-4286-a35f-eb431d025ac9-operator-scripts\") pod \"barbican-db-create-m4zmz\" (UID: \"95d067ec-b8ae-4286-a35f-eb431d025ac9\") " pod="openstack/barbican-db-create-m4zmz" Nov 28 13:36:53 crc kubenswrapper[4631]: I1128 13:36:53.287277 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h7nsj\" (UniqueName: \"kubernetes.io/projected/127100c3-8f48-4ddf-a289-f08f3cdf0d8a-kube-api-access-h7nsj\") pod \"barbican-8a8c-account-create-update-lskqg\" (UID: \"127100c3-8f48-4ddf-a289-f08f3cdf0d8a\") " pod="openstack/barbican-8a8c-account-create-update-lskqg" Nov 28 13:36:53 crc kubenswrapper[4631]: I1128 13:36:53.290502 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/95d067ec-b8ae-4286-a35f-eb431d025ac9-operator-scripts\") pod \"barbican-db-create-m4zmz\" (UID: \"95d067ec-b8ae-4286-a35f-eb431d025ac9\") " pod="openstack/barbican-db-create-m4zmz" Nov 28 13:36:53 crc kubenswrapper[4631]: I1128 13:36:53.342084 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ss6q5\" (UniqueName: \"kubernetes.io/projected/95d067ec-b8ae-4286-a35f-eb431d025ac9-kube-api-access-ss6q5\") pod \"barbican-db-create-m4zmz\" (UID: \"95d067ec-b8ae-4286-a35f-eb431d025ac9\") " pod="openstack/barbican-db-create-m4zmz" Nov 28 13:36:53 crc kubenswrapper[4631]: I1128 13:36:53.360187 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-c506-account-create-update-nr5s5"] Nov 28 13:36:53 crc kubenswrapper[4631]: I1128 13:36:53.364502 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-c506-account-create-update-nr5s5" Nov 28 13:36:53 crc kubenswrapper[4631]: I1128 13:36:53.371163 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Nov 28 13:36:53 crc kubenswrapper[4631]: I1128 13:36:53.386872 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-c506-account-create-update-nr5s5"] Nov 28 13:36:53 crc kubenswrapper[4631]: I1128 13:36:53.390239 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/127100c3-8f48-4ddf-a289-f08f3cdf0d8a-operator-scripts\") pod \"barbican-8a8c-account-create-update-lskqg\" (UID: \"127100c3-8f48-4ddf-a289-f08f3cdf0d8a\") " pod="openstack/barbican-8a8c-account-create-update-lskqg" Nov 28 13:36:53 crc kubenswrapper[4631]: I1128 13:36:53.390339 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h7nsj\" (UniqueName: \"kubernetes.io/projected/127100c3-8f48-4ddf-a289-f08f3cdf0d8a-kube-api-access-h7nsj\") pod \"barbican-8a8c-account-create-update-lskqg\" (UID: \"127100c3-8f48-4ddf-a289-f08f3cdf0d8a\") " pod="openstack/barbican-8a8c-account-create-update-lskqg" Nov 28 13:36:53 crc kubenswrapper[4631]: I1128 13:36:53.391235 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/127100c3-8f48-4ddf-a289-f08f3cdf0d8a-operator-scripts\") pod \"barbican-8a8c-account-create-update-lskqg\" (UID: \"127100c3-8f48-4ddf-a289-f08f3cdf0d8a\") " pod="openstack/barbican-8a8c-account-create-update-lskqg" Nov 28 13:36:53 crc kubenswrapper[4631]: I1128 13:36:53.418929 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h7nsj\" (UniqueName: \"kubernetes.io/projected/127100c3-8f48-4ddf-a289-f08f3cdf0d8a-kube-api-access-h7nsj\") pod \"barbican-8a8c-account-create-update-lskqg\" (UID: \"127100c3-8f48-4ddf-a289-f08f3cdf0d8a\") " pod="openstack/barbican-8a8c-account-create-update-lskqg" Nov 28 13:36:53 crc kubenswrapper[4631]: I1128 13:36:53.421862 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-m4zmz" Nov 28 13:36:53 crc kubenswrapper[4631]: I1128 13:36:53.497943 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-8a8c-account-create-update-lskqg" Nov 28 13:36:53 crc kubenswrapper[4631]: I1128 13:36:53.501263 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ecb43d0a-167c-42a9-9ea9-bc303f5890b8-operator-scripts\") pod \"cinder-c506-account-create-update-nr5s5\" (UID: \"ecb43d0a-167c-42a9-9ea9-bc303f5890b8\") " pod="openstack/cinder-c506-account-create-update-nr5s5" Nov 28 13:36:53 crc kubenswrapper[4631]: I1128 13:36:53.501684 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mcrq6\" (UniqueName: \"kubernetes.io/projected/ecb43d0a-167c-42a9-9ea9-bc303f5890b8-kube-api-access-mcrq6\") pod \"cinder-c506-account-create-update-nr5s5\" (UID: \"ecb43d0a-167c-42a9-9ea9-bc303f5890b8\") " pod="openstack/cinder-c506-account-create-update-nr5s5" Nov 28 13:36:53 crc kubenswrapper[4631]: I1128 13:36:53.599647 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-hplqp"] Nov 28 13:36:53 crc kubenswrapper[4631]: I1128 13:36:53.601142 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-hplqp"] Nov 28 13:36:53 crc kubenswrapper[4631]: I1128 13:36:53.604443 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-hplqp" Nov 28 13:36:53 crc kubenswrapper[4631]: I1128 13:36:53.607706 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ecb43d0a-167c-42a9-9ea9-bc303f5890b8-operator-scripts\") pod \"cinder-c506-account-create-update-nr5s5\" (UID: \"ecb43d0a-167c-42a9-9ea9-bc303f5890b8\") " pod="openstack/cinder-c506-account-create-update-nr5s5" Nov 28 13:36:53 crc kubenswrapper[4631]: I1128 13:36:53.607762 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mcrq6\" (UniqueName: \"kubernetes.io/projected/ecb43d0a-167c-42a9-9ea9-bc303f5890b8-kube-api-access-mcrq6\") pod \"cinder-c506-account-create-update-nr5s5\" (UID: \"ecb43d0a-167c-42a9-9ea9-bc303f5890b8\") " pod="openstack/cinder-c506-account-create-update-nr5s5" Nov 28 13:36:53 crc kubenswrapper[4631]: I1128 13:36:53.613002 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ecb43d0a-167c-42a9-9ea9-bc303f5890b8-operator-scripts\") pod \"cinder-c506-account-create-update-nr5s5\" (UID: \"ecb43d0a-167c-42a9-9ea9-bc303f5890b8\") " pod="openstack/cinder-c506-account-create-update-nr5s5" Nov 28 13:36:53 crc kubenswrapper[4631]: I1128 13:36:53.664228 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mcrq6\" (UniqueName: \"kubernetes.io/projected/ecb43d0a-167c-42a9-9ea9-bc303f5890b8-kube-api-access-mcrq6\") pod \"cinder-c506-account-create-update-nr5s5\" (UID: \"ecb43d0a-167c-42a9-9ea9-bc303f5890b8\") " pod="openstack/cinder-c506-account-create-update-nr5s5" Nov 28 13:36:53 crc kubenswrapper[4631]: I1128 13:36:53.682026 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-99b5-account-create-update-67zx8"] Nov 28 13:36:53 crc kubenswrapper[4631]: I1128 13:36:53.683462 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-99b5-account-create-update-67zx8" Nov 28 13:36:53 crc kubenswrapper[4631]: I1128 13:36:53.688006 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-c506-account-create-update-nr5s5" Nov 28 13:36:53 crc kubenswrapper[4631]: I1128 13:36:53.709883 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gl8xp\" (UniqueName: \"kubernetes.io/projected/e8fd4db1-5308-414f-9eac-ddef2c41867a-kube-api-access-gl8xp\") pod \"neutron-db-create-hplqp\" (UID: \"e8fd4db1-5308-414f-9eac-ddef2c41867a\") " pod="openstack/neutron-db-create-hplqp" Nov 28 13:36:53 crc kubenswrapper[4631]: I1128 13:36:53.709961 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e8fd4db1-5308-414f-9eac-ddef2c41867a-operator-scripts\") pod \"neutron-db-create-hplqp\" (UID: \"e8fd4db1-5308-414f-9eac-ddef2c41867a\") " pod="openstack/neutron-db-create-hplqp" Nov 28 13:36:53 crc kubenswrapper[4631]: I1128 13:36:53.734123 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Nov 28 13:36:53 crc kubenswrapper[4631]: I1128 13:36:53.790226 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-99b5-account-create-update-67zx8"] Nov 28 13:36:53 crc kubenswrapper[4631]: I1128 13:36:53.811305 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lvvjh\" (UniqueName: \"kubernetes.io/projected/cee396c9-93ab-45dd-82d0-9641e985644a-kube-api-access-lvvjh\") pod \"neutron-99b5-account-create-update-67zx8\" (UID: \"cee396c9-93ab-45dd-82d0-9641e985644a\") " pod="openstack/neutron-99b5-account-create-update-67zx8" Nov 28 13:36:53 crc kubenswrapper[4631]: I1128 13:36:53.811377 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gl8xp\" (UniqueName: \"kubernetes.io/projected/e8fd4db1-5308-414f-9eac-ddef2c41867a-kube-api-access-gl8xp\") pod \"neutron-db-create-hplqp\" (UID: \"e8fd4db1-5308-414f-9eac-ddef2c41867a\") " pod="openstack/neutron-db-create-hplqp" Nov 28 13:36:53 crc kubenswrapper[4631]: I1128 13:36:53.811440 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e8fd4db1-5308-414f-9eac-ddef2c41867a-operator-scripts\") pod \"neutron-db-create-hplqp\" (UID: \"e8fd4db1-5308-414f-9eac-ddef2c41867a\") " pod="openstack/neutron-db-create-hplqp" Nov 28 13:36:53 crc kubenswrapper[4631]: I1128 13:36:53.811506 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cee396c9-93ab-45dd-82d0-9641e985644a-operator-scripts\") pod \"neutron-99b5-account-create-update-67zx8\" (UID: \"cee396c9-93ab-45dd-82d0-9641e985644a\") " pod="openstack/neutron-99b5-account-create-update-67zx8" Nov 28 13:36:53 crc kubenswrapper[4631]: I1128 13:36:53.812603 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e8fd4db1-5308-414f-9eac-ddef2c41867a-operator-scripts\") pod \"neutron-db-create-hplqp\" (UID: \"e8fd4db1-5308-414f-9eac-ddef2c41867a\") " pod="openstack/neutron-db-create-hplqp" Nov 28 13:36:53 crc kubenswrapper[4631]: I1128 13:36:53.860393 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-xs457"] Nov 28 13:36:53 crc kubenswrapper[4631]: I1128 13:36:53.861538 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gl8xp\" (UniqueName: \"kubernetes.io/projected/e8fd4db1-5308-414f-9eac-ddef2c41867a-kube-api-access-gl8xp\") pod \"neutron-db-create-hplqp\" (UID: \"e8fd4db1-5308-414f-9eac-ddef2c41867a\") " pod="openstack/neutron-db-create-hplqp" Nov 28 13:36:53 crc kubenswrapper[4631]: I1128 13:36:53.861844 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-xs457" Nov 28 13:36:53 crc kubenswrapper[4631]: I1128 13:36:53.874920 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Nov 28 13:36:53 crc kubenswrapper[4631]: I1128 13:36:53.875338 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Nov 28 13:36:53 crc kubenswrapper[4631]: I1128 13:36:53.875640 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-szfjn" Nov 28 13:36:53 crc kubenswrapper[4631]: I1128 13:36:53.875758 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Nov 28 13:36:53 crc kubenswrapper[4631]: I1128 13:36:53.913327 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lvvjh\" (UniqueName: \"kubernetes.io/projected/cee396c9-93ab-45dd-82d0-9641e985644a-kube-api-access-lvvjh\") pod \"neutron-99b5-account-create-update-67zx8\" (UID: \"cee396c9-93ab-45dd-82d0-9641e985644a\") " pod="openstack/neutron-99b5-account-create-update-67zx8" Nov 28 13:36:53 crc kubenswrapper[4631]: I1128 13:36:53.913650 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b6844d09-9bb8-44a7-bf9a-eca18ce5c1c8-config-data\") pod \"keystone-db-sync-xs457\" (UID: \"b6844d09-9bb8-44a7-bf9a-eca18ce5c1c8\") " pod="openstack/keystone-db-sync-xs457" Nov 28 13:36:53 crc kubenswrapper[4631]: I1128 13:36:53.913816 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wnwzq\" (UniqueName: \"kubernetes.io/projected/b6844d09-9bb8-44a7-bf9a-eca18ce5c1c8-kube-api-access-wnwzq\") pod \"keystone-db-sync-xs457\" (UID: \"b6844d09-9bb8-44a7-bf9a-eca18ce5c1c8\") " pod="openstack/keystone-db-sync-xs457" Nov 28 13:36:53 crc kubenswrapper[4631]: I1128 13:36:53.913995 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b6844d09-9bb8-44a7-bf9a-eca18ce5c1c8-combined-ca-bundle\") pod \"keystone-db-sync-xs457\" (UID: \"b6844d09-9bb8-44a7-bf9a-eca18ce5c1c8\") " pod="openstack/keystone-db-sync-xs457" Nov 28 13:36:53 crc kubenswrapper[4631]: I1128 13:36:53.914147 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cee396c9-93ab-45dd-82d0-9641e985644a-operator-scripts\") pod \"neutron-99b5-account-create-update-67zx8\" (UID: \"cee396c9-93ab-45dd-82d0-9641e985644a\") " pod="openstack/neutron-99b5-account-create-update-67zx8" Nov 28 13:36:53 crc kubenswrapper[4631]: I1128 13:36:53.914923 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cee396c9-93ab-45dd-82d0-9641e985644a-operator-scripts\") pod \"neutron-99b5-account-create-update-67zx8\" (UID: \"cee396c9-93ab-45dd-82d0-9641e985644a\") " pod="openstack/neutron-99b5-account-create-update-67zx8" Nov 28 13:36:53 crc kubenswrapper[4631]: I1128 13:36:53.921355 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"c3e1f6fb-1845-4a39-8eca-1262da1ff281","Type":"ContainerStarted","Data":"439a857a9e8d924d5c4f137c1e901e3a751ad458f100376acb2d01e227f69100"} Nov 28 13:36:53 crc kubenswrapper[4631]: I1128 13:36:53.930611 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-xs457"] Nov 28 13:36:53 crc kubenswrapper[4631]: I1128 13:36:53.983376 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lvvjh\" (UniqueName: \"kubernetes.io/projected/cee396c9-93ab-45dd-82d0-9641e985644a-kube-api-access-lvvjh\") pod \"neutron-99b5-account-create-update-67zx8\" (UID: \"cee396c9-93ab-45dd-82d0-9641e985644a\") " pod="openstack/neutron-99b5-account-create-update-67zx8" Nov 28 13:36:54 crc kubenswrapper[4631]: I1128 13:36:54.022171 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b6844d09-9bb8-44a7-bf9a-eca18ce5c1c8-config-data\") pod \"keystone-db-sync-xs457\" (UID: \"b6844d09-9bb8-44a7-bf9a-eca18ce5c1c8\") " pod="openstack/keystone-db-sync-xs457" Nov 28 13:36:54 crc kubenswrapper[4631]: I1128 13:36:54.022240 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wnwzq\" (UniqueName: \"kubernetes.io/projected/b6844d09-9bb8-44a7-bf9a-eca18ce5c1c8-kube-api-access-wnwzq\") pod \"keystone-db-sync-xs457\" (UID: \"b6844d09-9bb8-44a7-bf9a-eca18ce5c1c8\") " pod="openstack/keystone-db-sync-xs457" Nov 28 13:36:54 crc kubenswrapper[4631]: I1128 13:36:54.022519 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b6844d09-9bb8-44a7-bf9a-eca18ce5c1c8-combined-ca-bundle\") pod \"keystone-db-sync-xs457\" (UID: \"b6844d09-9bb8-44a7-bf9a-eca18ce5c1c8\") " pod="openstack/keystone-db-sync-xs457" Nov 28 13:36:54 crc kubenswrapper[4631]: I1128 13:36:54.030997 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-hplqp" Nov 28 13:36:54 crc kubenswrapper[4631]: I1128 13:36:54.080549 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b6844d09-9bb8-44a7-bf9a-eca18ce5c1c8-config-data\") pod \"keystone-db-sync-xs457\" (UID: \"b6844d09-9bb8-44a7-bf9a-eca18ce5c1c8\") " pod="openstack/keystone-db-sync-xs457" Nov 28 13:36:54 crc kubenswrapper[4631]: I1128 13:36:54.094214 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b6844d09-9bb8-44a7-bf9a-eca18ce5c1c8-combined-ca-bundle\") pod \"keystone-db-sync-xs457\" (UID: \"b6844d09-9bb8-44a7-bf9a-eca18ce5c1c8\") " pod="openstack/keystone-db-sync-xs457" Nov 28 13:36:54 crc kubenswrapper[4631]: I1128 13:36:54.110795 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wnwzq\" (UniqueName: \"kubernetes.io/projected/b6844d09-9bb8-44a7-bf9a-eca18ce5c1c8-kube-api-access-wnwzq\") pod \"keystone-db-sync-xs457\" (UID: \"b6844d09-9bb8-44a7-bf9a-eca18ce5c1c8\") " pod="openstack/keystone-db-sync-xs457" Nov 28 13:36:54 crc kubenswrapper[4631]: I1128 13:36:54.133720 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-wrd26"] Nov 28 13:36:54 crc kubenswrapper[4631]: W1128 13:36:54.332531 4631 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod98564bcf_8a3f_4324_af81_bafdc2ed2934.slice/crio-083b9088b5d4d41c287b02703a1c8fbfa46aa8576b48f0e0cfc972d7ab77d79a WatchSource:0}: Error finding container 083b9088b5d4d41c287b02703a1c8fbfa46aa8576b48f0e0cfc972d7ab77d79a: Status 404 returned error can't find the container with id 083b9088b5d4d41c287b02703a1c8fbfa46aa8576b48f0e0cfc972d7ab77d79a Nov 28 13:36:54 crc kubenswrapper[4631]: I1128 13:36:54.337275 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-99b5-account-create-update-67zx8" Nov 28 13:36:54 crc kubenswrapper[4631]: I1128 13:36:54.376638 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-xs457" Nov 28 13:36:54 crc kubenswrapper[4631]: I1128 13:36:54.762939 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-8a8c-account-create-update-lskqg"] Nov 28 13:36:54 crc kubenswrapper[4631]: I1128 13:36:54.782199 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-m4zmz"] Nov 28 13:36:55 crc kubenswrapper[4631]: I1128 13:36:55.002059 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"c3e1f6fb-1845-4a39-8eca-1262da1ff281","Type":"ContainerStarted","Data":"764fdf6d258709c9dde8bee57a40787fc6fd5caa954e2ebaf089e4f5955f18d8"} Nov 28 13:36:55 crc kubenswrapper[4631]: I1128 13:36:55.010170 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-m4zmz" event={"ID":"95d067ec-b8ae-4286-a35f-eb431d025ac9","Type":"ContainerStarted","Data":"cbac8b9ba48d9ebd1b5b5ce43698c8fdd1f016df114dc385990c5ec216ac8ee8"} Nov 28 13:36:55 crc kubenswrapper[4631]: I1128 13:36:55.040364 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-wrd26" event={"ID":"98564bcf-8a3f-4324-af81-bafdc2ed2934","Type":"ContainerStarted","Data":"083b9088b5d4d41c287b02703a1c8fbfa46aa8576b48f0e0cfc972d7ab77d79a"} Nov 28 13:36:55 crc kubenswrapper[4631]: I1128 13:36:55.045907 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-8a8c-account-create-update-lskqg" event={"ID":"127100c3-8f48-4ddf-a289-f08f3cdf0d8a","Type":"ContainerStarted","Data":"bdcf0f0360d41edb285cdf11f52b9614dd00c3037eac91c5749e15a7f568846b"} Nov 28 13:36:55 crc kubenswrapper[4631]: I1128 13:36:55.051391 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-c506-account-create-update-nr5s5"] Nov 28 13:36:57 crc kubenswrapper[4631]: I1128 13:36:55.265606 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-99b5-account-create-update-67zx8"] Nov 28 13:36:57 crc kubenswrapper[4631]: I1128 13:36:55.369034 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-hplqp"] Nov 28 13:36:57 crc kubenswrapper[4631]: I1128 13:36:55.505624 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-xs457"] Nov 28 13:36:57 crc kubenswrapper[4631]: I1128 13:36:56.062098 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-c506-account-create-update-nr5s5" event={"ID":"ecb43d0a-167c-42a9-9ea9-bc303f5890b8","Type":"ContainerStarted","Data":"de1e465e275bc8f34db5322875dd82f091e44b7cb0f3f75e8cca3eed50170f6c"} Nov 28 13:36:57 crc kubenswrapper[4631]: I1128 13:36:56.068151 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-99b5-account-create-update-67zx8" event={"ID":"cee396c9-93ab-45dd-82d0-9641e985644a","Type":"ContainerStarted","Data":"d125961efe689f993c939f0a9bdf44e9f34f1579ab7afe15f1e85717edc50656"} Nov 28 13:36:57 crc kubenswrapper[4631]: I1128 13:36:56.070851 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-hplqp" event={"ID":"e8fd4db1-5308-414f-9eac-ddef2c41867a","Type":"ContainerStarted","Data":"bc8aedf9f691384db0116de1386d6b2839d9cd4da28760539c96719ba11b9cda"} Nov 28 13:36:57 crc kubenswrapper[4631]: I1128 13:36:56.079379 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-xs457" event={"ID":"b6844d09-9bb8-44a7-bf9a-eca18ce5c1c8","Type":"ContainerStarted","Data":"f4adc0bdd4d8461a300e92778ea0211b15567cec11c29f5da558ca7ac66064d8"} Nov 28 13:36:57 crc kubenswrapper[4631]: I1128 13:36:57.099226 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-m4zmz" event={"ID":"95d067ec-b8ae-4286-a35f-eb431d025ac9","Type":"ContainerStarted","Data":"a3f65d289623bf199c36739c4c862db70137638bad3d4278065fec2e5206cf1d"} Nov 28 13:36:57 crc kubenswrapper[4631]: I1128 13:36:57.117165 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-wrd26" event={"ID":"98564bcf-8a3f-4324-af81-bafdc2ed2934","Type":"ContainerStarted","Data":"fc8711f77c49ff38e295ea35860dbf498a2f8e950091ebc9ba461288c475526d"} Nov 28 13:36:57 crc kubenswrapper[4631]: I1128 13:36:57.120378 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-8a8c-account-create-update-lskqg" event={"ID":"127100c3-8f48-4ddf-a289-f08f3cdf0d8a","Type":"ContainerStarted","Data":"482ce3c6826106b00b097fc029b492fb977fb23e7ade8a08dfaa1ece091c188e"} Nov 28 13:36:57 crc kubenswrapper[4631]: I1128 13:36:57.123353 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-c506-account-create-update-nr5s5" event={"ID":"ecb43d0a-167c-42a9-9ea9-bc303f5890b8","Type":"ContainerStarted","Data":"b03a3cbbf0292dae3632dfc0115e24e09e6ef8227d086f5bfbb211d9d09c3046"} Nov 28 13:36:57 crc kubenswrapper[4631]: I1128 13:36:57.134538 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-create-m4zmz" podStartSLOduration=4.134492592 podStartE2EDuration="4.134492592s" podCreationTimestamp="2025-11-28 13:36:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 13:36:57.124509537 +0000 UTC m=+973.931812881" watchObservedRunningTime="2025-11-28 13:36:57.134492592 +0000 UTC m=+973.941795936" Nov 28 13:36:57 crc kubenswrapper[4631]: I1128 13:36:57.150465 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-99b5-account-create-update-67zx8" event={"ID":"cee396c9-93ab-45dd-82d0-9641e985644a","Type":"ContainerStarted","Data":"bf7fdb2038d321e6b67de7b7a577f31a3431fe807da206dd011898c5be98301b"} Nov 28 13:36:57 crc kubenswrapper[4631]: I1128 13:36:57.166093 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-8a8c-account-create-update-lskqg" podStartSLOduration=4.166065257 podStartE2EDuration="4.166065257s" podCreationTimestamp="2025-11-28 13:36:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 13:36:57.160120372 +0000 UTC m=+973.967423716" watchObservedRunningTime="2025-11-28 13:36:57.166065257 +0000 UTC m=+973.973368591" Nov 28 13:36:57 crc kubenswrapper[4631]: I1128 13:36:57.171701 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-hplqp" event={"ID":"e8fd4db1-5308-414f-9eac-ddef2c41867a","Type":"ContainerStarted","Data":"459e0a9725c8f0753466ab2f643456b6345e1e52cf742bce27b3e8c960230948"} Nov 28 13:36:57 crc kubenswrapper[4631]: I1128 13:36:57.204143 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"c3e1f6fb-1845-4a39-8eca-1262da1ff281","Type":"ContainerStarted","Data":"7f25b77138eb277c3218bd075547ac24013abe29ca13b47883e3e3718e1ec0e3"} Nov 28 13:36:57 crc kubenswrapper[4631]: I1128 13:36:57.308170 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-create-wrd26" podStartSLOduration=5.308139459 podStartE2EDuration="5.308139459s" podCreationTimestamp="2025-11-28 13:36:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 13:36:57.198809273 +0000 UTC m=+974.006112617" watchObservedRunningTime="2025-11-28 13:36:57.308139459 +0000 UTC m=+974.115442803" Nov 28 13:36:57 crc kubenswrapper[4631]: I1128 13:36:57.309600 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-c506-account-create-update-nr5s5" podStartSLOduration=4.309563574 podStartE2EDuration="4.309563574s" podCreationTimestamp="2025-11-28 13:36:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 13:36:57.230253635 +0000 UTC m=+974.037556989" watchObservedRunningTime="2025-11-28 13:36:57.309563574 +0000 UTC m=+974.116866918" Nov 28 13:36:57 crc kubenswrapper[4631]: I1128 13:36:57.329482 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-create-hplqp" podStartSLOduration=4.3294487329999995 podStartE2EDuration="4.329448733s" podCreationTimestamp="2025-11-28 13:36:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 13:36:57.258391806 +0000 UTC m=+974.065695170" watchObservedRunningTime="2025-11-28 13:36:57.329448733 +0000 UTC m=+974.136752077" Nov 28 13:36:57 crc kubenswrapper[4631]: I1128 13:36:57.338446 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-99b5-account-create-update-67zx8" podStartSLOduration=4.338419543 podStartE2EDuration="4.338419543s" podCreationTimestamp="2025-11-28 13:36:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 13:36:57.295177181 +0000 UTC m=+974.102480545" watchObservedRunningTime="2025-11-28 13:36:57.338419543 +0000 UTC m=+974.145722897" Nov 28 13:36:58 crc kubenswrapper[4631]: I1128 13:36:58.221025 4631 generic.go:334] "Generic (PLEG): container finished" podID="95d067ec-b8ae-4286-a35f-eb431d025ac9" containerID="a3f65d289623bf199c36739c4c862db70137638bad3d4278065fec2e5206cf1d" exitCode=0 Nov 28 13:36:58 crc kubenswrapper[4631]: I1128 13:36:58.221127 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-m4zmz" event={"ID":"95d067ec-b8ae-4286-a35f-eb431d025ac9","Type":"ContainerDied","Data":"a3f65d289623bf199c36739c4c862db70137638bad3d4278065fec2e5206cf1d"} Nov 28 13:36:58 crc kubenswrapper[4631]: I1128 13:36:58.226357 4631 generic.go:334] "Generic (PLEG): container finished" podID="98564bcf-8a3f-4324-af81-bafdc2ed2934" containerID="fc8711f77c49ff38e295ea35860dbf498a2f8e950091ebc9ba461288c475526d" exitCode=0 Nov 28 13:36:58 crc kubenswrapper[4631]: I1128 13:36:58.226427 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-wrd26" event={"ID":"98564bcf-8a3f-4324-af81-bafdc2ed2934","Type":"ContainerDied","Data":"fc8711f77c49ff38e295ea35860dbf498a2f8e950091ebc9ba461288c475526d"} Nov 28 13:36:58 crc kubenswrapper[4631]: I1128 13:36:58.228489 4631 generic.go:334] "Generic (PLEG): container finished" podID="127100c3-8f48-4ddf-a289-f08f3cdf0d8a" containerID="482ce3c6826106b00b097fc029b492fb977fb23e7ade8a08dfaa1ece091c188e" exitCode=0 Nov 28 13:36:58 crc kubenswrapper[4631]: I1128 13:36:58.228602 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-8a8c-account-create-update-lskqg" event={"ID":"127100c3-8f48-4ddf-a289-f08f3cdf0d8a","Type":"ContainerDied","Data":"482ce3c6826106b00b097fc029b492fb977fb23e7ade8a08dfaa1ece091c188e"} Nov 28 13:36:58 crc kubenswrapper[4631]: I1128 13:36:58.237848 4631 generic.go:334] "Generic (PLEG): container finished" podID="ecb43d0a-167c-42a9-9ea9-bc303f5890b8" containerID="b03a3cbbf0292dae3632dfc0115e24e09e6ef8227d086f5bfbb211d9d09c3046" exitCode=0 Nov 28 13:36:58 crc kubenswrapper[4631]: I1128 13:36:58.237948 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-c506-account-create-update-nr5s5" event={"ID":"ecb43d0a-167c-42a9-9ea9-bc303f5890b8","Type":"ContainerDied","Data":"b03a3cbbf0292dae3632dfc0115e24e09e6ef8227d086f5bfbb211d9d09c3046"} Nov 28 13:36:58 crc kubenswrapper[4631]: I1128 13:36:58.240156 4631 generic.go:334] "Generic (PLEG): container finished" podID="cee396c9-93ab-45dd-82d0-9641e985644a" containerID="bf7fdb2038d321e6b67de7b7a577f31a3431fe807da206dd011898c5be98301b" exitCode=0 Nov 28 13:36:58 crc kubenswrapper[4631]: I1128 13:36:58.240346 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-99b5-account-create-update-67zx8" event={"ID":"cee396c9-93ab-45dd-82d0-9641e985644a","Type":"ContainerDied","Data":"bf7fdb2038d321e6b67de7b7a577f31a3431fe807da206dd011898c5be98301b"} Nov 28 13:36:58 crc kubenswrapper[4631]: I1128 13:36:58.242796 4631 generic.go:334] "Generic (PLEG): container finished" podID="e8fd4db1-5308-414f-9eac-ddef2c41867a" containerID="459e0a9725c8f0753466ab2f643456b6345e1e52cf742bce27b3e8c960230948" exitCode=0 Nov 28 13:36:58 crc kubenswrapper[4631]: I1128 13:36:58.242896 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-hplqp" event={"ID":"e8fd4db1-5308-414f-9eac-ddef2c41867a","Type":"ContainerDied","Data":"459e0a9725c8f0753466ab2f643456b6345e1e52cf742bce27b3e8c960230948"} Nov 28 13:36:58 crc kubenswrapper[4631]: I1128 13:36:58.274333 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"c3e1f6fb-1845-4a39-8eca-1262da1ff281","Type":"ContainerStarted","Data":"12f647585fbebbfa89381ebebb911858dac72218c12c68c224533b0b90c094a2"} Nov 28 13:36:58 crc kubenswrapper[4631]: I1128 13:36:58.458071 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-storage-0" podStartSLOduration=42.258707285 podStartE2EDuration="50.458047775s" podCreationTimestamp="2025-11-28 13:36:08 +0000 UTC" firstStartedPulling="2025-11-28 13:36:42.771369342 +0000 UTC m=+959.578672676" lastFinishedPulling="2025-11-28 13:36:50.970709812 +0000 UTC m=+967.778013166" observedRunningTime="2025-11-28 13:36:58.455421571 +0000 UTC m=+975.262724925" watchObservedRunningTime="2025-11-28 13:36:58.458047775 +0000 UTC m=+975.265351109" Nov 28 13:36:58 crc kubenswrapper[4631]: I1128 13:36:58.552484 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5b946c75cc-rnp88" Nov 28 13:36:58 crc kubenswrapper[4631]: I1128 13:36:58.671404 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-698758b865-m2zrc"] Nov 28 13:36:58 crc kubenswrapper[4631]: I1128 13:36:58.671663 4631 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-698758b865-m2zrc" podUID="ed04b83d-c180-4bf2-b6b9-b112a536d9da" containerName="dnsmasq-dns" containerID="cri-o://6abb47410b6f4f8e4c483aea263ea839b52160f49e3c028e22b32ec972d55780" gracePeriod=10 Nov 28 13:36:58 crc kubenswrapper[4631]: I1128 13:36:58.942710 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-74f6bcbc87-r8g9l"] Nov 28 13:36:58 crc kubenswrapper[4631]: I1128 13:36:58.944555 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74f6bcbc87-r8g9l" Nov 28 13:36:58 crc kubenswrapper[4631]: I1128 13:36:58.954419 4631 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-698758b865-m2zrc" podUID="ed04b83d-c180-4bf2-b6b9-b112a536d9da" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.116:5353: connect: connection refused" Nov 28 13:36:58 crc kubenswrapper[4631]: I1128 13:36:58.960174 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-swift-storage-0" Nov 28 13:36:58 crc kubenswrapper[4631]: I1128 13:36:58.981481 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-74f6bcbc87-r8g9l"] Nov 28 13:36:59 crc kubenswrapper[4631]: I1128 13:36:59.061998 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/166f0f43-e5cb-42ad-a2f5-1b207accc673-dns-swift-storage-0\") pod \"dnsmasq-dns-74f6bcbc87-r8g9l\" (UID: \"166f0f43-e5cb-42ad-a2f5-1b207accc673\") " pod="openstack/dnsmasq-dns-74f6bcbc87-r8g9l" Nov 28 13:36:59 crc kubenswrapper[4631]: I1128 13:36:59.062126 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-brqml\" (UniqueName: \"kubernetes.io/projected/166f0f43-e5cb-42ad-a2f5-1b207accc673-kube-api-access-brqml\") pod \"dnsmasq-dns-74f6bcbc87-r8g9l\" (UID: \"166f0f43-e5cb-42ad-a2f5-1b207accc673\") " pod="openstack/dnsmasq-dns-74f6bcbc87-r8g9l" Nov 28 13:36:59 crc kubenswrapper[4631]: I1128 13:36:59.062170 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/166f0f43-e5cb-42ad-a2f5-1b207accc673-config\") pod \"dnsmasq-dns-74f6bcbc87-r8g9l\" (UID: \"166f0f43-e5cb-42ad-a2f5-1b207accc673\") " pod="openstack/dnsmasq-dns-74f6bcbc87-r8g9l" Nov 28 13:36:59 crc kubenswrapper[4631]: I1128 13:36:59.062237 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/166f0f43-e5cb-42ad-a2f5-1b207accc673-ovsdbserver-nb\") pod \"dnsmasq-dns-74f6bcbc87-r8g9l\" (UID: \"166f0f43-e5cb-42ad-a2f5-1b207accc673\") " pod="openstack/dnsmasq-dns-74f6bcbc87-r8g9l" Nov 28 13:36:59 crc kubenswrapper[4631]: I1128 13:36:59.062260 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/166f0f43-e5cb-42ad-a2f5-1b207accc673-ovsdbserver-sb\") pod \"dnsmasq-dns-74f6bcbc87-r8g9l\" (UID: \"166f0f43-e5cb-42ad-a2f5-1b207accc673\") " pod="openstack/dnsmasq-dns-74f6bcbc87-r8g9l" Nov 28 13:36:59 crc kubenswrapper[4631]: I1128 13:36:59.062371 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/166f0f43-e5cb-42ad-a2f5-1b207accc673-dns-svc\") pod \"dnsmasq-dns-74f6bcbc87-r8g9l\" (UID: \"166f0f43-e5cb-42ad-a2f5-1b207accc673\") " pod="openstack/dnsmasq-dns-74f6bcbc87-r8g9l" Nov 28 13:36:59 crc kubenswrapper[4631]: I1128 13:36:59.164377 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/166f0f43-e5cb-42ad-a2f5-1b207accc673-dns-svc\") pod \"dnsmasq-dns-74f6bcbc87-r8g9l\" (UID: \"166f0f43-e5cb-42ad-a2f5-1b207accc673\") " pod="openstack/dnsmasq-dns-74f6bcbc87-r8g9l" Nov 28 13:36:59 crc kubenswrapper[4631]: I1128 13:36:59.164485 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/166f0f43-e5cb-42ad-a2f5-1b207accc673-dns-swift-storage-0\") pod \"dnsmasq-dns-74f6bcbc87-r8g9l\" (UID: \"166f0f43-e5cb-42ad-a2f5-1b207accc673\") " pod="openstack/dnsmasq-dns-74f6bcbc87-r8g9l" Nov 28 13:36:59 crc kubenswrapper[4631]: I1128 13:36:59.164537 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-brqml\" (UniqueName: \"kubernetes.io/projected/166f0f43-e5cb-42ad-a2f5-1b207accc673-kube-api-access-brqml\") pod \"dnsmasq-dns-74f6bcbc87-r8g9l\" (UID: \"166f0f43-e5cb-42ad-a2f5-1b207accc673\") " pod="openstack/dnsmasq-dns-74f6bcbc87-r8g9l" Nov 28 13:36:59 crc kubenswrapper[4631]: I1128 13:36:59.164567 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/166f0f43-e5cb-42ad-a2f5-1b207accc673-config\") pod \"dnsmasq-dns-74f6bcbc87-r8g9l\" (UID: \"166f0f43-e5cb-42ad-a2f5-1b207accc673\") " pod="openstack/dnsmasq-dns-74f6bcbc87-r8g9l" Nov 28 13:36:59 crc kubenswrapper[4631]: I1128 13:36:59.164633 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/166f0f43-e5cb-42ad-a2f5-1b207accc673-ovsdbserver-nb\") pod \"dnsmasq-dns-74f6bcbc87-r8g9l\" (UID: \"166f0f43-e5cb-42ad-a2f5-1b207accc673\") " pod="openstack/dnsmasq-dns-74f6bcbc87-r8g9l" Nov 28 13:36:59 crc kubenswrapper[4631]: I1128 13:36:59.164653 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/166f0f43-e5cb-42ad-a2f5-1b207accc673-ovsdbserver-sb\") pod \"dnsmasq-dns-74f6bcbc87-r8g9l\" (UID: \"166f0f43-e5cb-42ad-a2f5-1b207accc673\") " pod="openstack/dnsmasq-dns-74f6bcbc87-r8g9l" Nov 28 13:36:59 crc kubenswrapper[4631]: I1128 13:36:59.165588 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/166f0f43-e5cb-42ad-a2f5-1b207accc673-ovsdbserver-sb\") pod \"dnsmasq-dns-74f6bcbc87-r8g9l\" (UID: \"166f0f43-e5cb-42ad-a2f5-1b207accc673\") " pod="openstack/dnsmasq-dns-74f6bcbc87-r8g9l" Nov 28 13:36:59 crc kubenswrapper[4631]: I1128 13:36:59.166090 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/166f0f43-e5cb-42ad-a2f5-1b207accc673-dns-svc\") pod \"dnsmasq-dns-74f6bcbc87-r8g9l\" (UID: \"166f0f43-e5cb-42ad-a2f5-1b207accc673\") " pod="openstack/dnsmasq-dns-74f6bcbc87-r8g9l" Nov 28 13:36:59 crc kubenswrapper[4631]: I1128 13:36:59.166649 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/166f0f43-e5cb-42ad-a2f5-1b207accc673-dns-swift-storage-0\") pod \"dnsmasq-dns-74f6bcbc87-r8g9l\" (UID: \"166f0f43-e5cb-42ad-a2f5-1b207accc673\") " pod="openstack/dnsmasq-dns-74f6bcbc87-r8g9l" Nov 28 13:36:59 crc kubenswrapper[4631]: I1128 13:36:59.167994 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/166f0f43-e5cb-42ad-a2f5-1b207accc673-config\") pod \"dnsmasq-dns-74f6bcbc87-r8g9l\" (UID: \"166f0f43-e5cb-42ad-a2f5-1b207accc673\") " pod="openstack/dnsmasq-dns-74f6bcbc87-r8g9l" Nov 28 13:36:59 crc kubenswrapper[4631]: I1128 13:36:59.168604 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/166f0f43-e5cb-42ad-a2f5-1b207accc673-ovsdbserver-nb\") pod \"dnsmasq-dns-74f6bcbc87-r8g9l\" (UID: \"166f0f43-e5cb-42ad-a2f5-1b207accc673\") " pod="openstack/dnsmasq-dns-74f6bcbc87-r8g9l" Nov 28 13:36:59 crc kubenswrapper[4631]: I1128 13:36:59.217674 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-brqml\" (UniqueName: \"kubernetes.io/projected/166f0f43-e5cb-42ad-a2f5-1b207accc673-kube-api-access-brqml\") pod \"dnsmasq-dns-74f6bcbc87-r8g9l\" (UID: \"166f0f43-e5cb-42ad-a2f5-1b207accc673\") " pod="openstack/dnsmasq-dns-74f6bcbc87-r8g9l" Nov 28 13:36:59 crc kubenswrapper[4631]: I1128 13:36:59.266672 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74f6bcbc87-r8g9l" Nov 28 13:36:59 crc kubenswrapper[4631]: I1128 13:36:59.286891 4631 generic.go:334] "Generic (PLEG): container finished" podID="ed04b83d-c180-4bf2-b6b9-b112a536d9da" containerID="6abb47410b6f4f8e4c483aea263ea839b52160f49e3c028e22b32ec972d55780" exitCode=0 Nov 28 13:36:59 crc kubenswrapper[4631]: I1128 13:36:59.287134 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-m2zrc" event={"ID":"ed04b83d-c180-4bf2-b6b9-b112a536d9da","Type":"ContainerDied","Data":"6abb47410b6f4f8e4c483aea263ea839b52160f49e3c028e22b32ec972d55780"} Nov 28 13:36:59 crc kubenswrapper[4631]: I1128 13:36:59.451096 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-m2zrc" Nov 28 13:36:59 crc kubenswrapper[4631]: I1128 13:36:59.570806 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ed04b83d-c180-4bf2-b6b9-b112a536d9da-dns-svc\") pod \"ed04b83d-c180-4bf2-b6b9-b112a536d9da\" (UID: \"ed04b83d-c180-4bf2-b6b9-b112a536d9da\") " Nov 28 13:36:59 crc kubenswrapper[4631]: I1128 13:36:59.571025 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ed04b83d-c180-4bf2-b6b9-b112a536d9da-ovsdbserver-sb\") pod \"ed04b83d-c180-4bf2-b6b9-b112a536d9da\" (UID: \"ed04b83d-c180-4bf2-b6b9-b112a536d9da\") " Nov 28 13:36:59 crc kubenswrapper[4631]: I1128 13:36:59.571194 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t8w2f\" (UniqueName: \"kubernetes.io/projected/ed04b83d-c180-4bf2-b6b9-b112a536d9da-kube-api-access-t8w2f\") pod \"ed04b83d-c180-4bf2-b6b9-b112a536d9da\" (UID: \"ed04b83d-c180-4bf2-b6b9-b112a536d9da\") " Nov 28 13:36:59 crc kubenswrapper[4631]: I1128 13:36:59.571312 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ed04b83d-c180-4bf2-b6b9-b112a536d9da-config\") pod \"ed04b83d-c180-4bf2-b6b9-b112a536d9da\" (UID: \"ed04b83d-c180-4bf2-b6b9-b112a536d9da\") " Nov 28 13:36:59 crc kubenswrapper[4631]: I1128 13:36:59.571423 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ed04b83d-c180-4bf2-b6b9-b112a536d9da-ovsdbserver-nb\") pod \"ed04b83d-c180-4bf2-b6b9-b112a536d9da\" (UID: \"ed04b83d-c180-4bf2-b6b9-b112a536d9da\") " Nov 28 13:36:59 crc kubenswrapper[4631]: I1128 13:36:59.580756 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ed04b83d-c180-4bf2-b6b9-b112a536d9da-kube-api-access-t8w2f" (OuterVolumeSpecName: "kube-api-access-t8w2f") pod "ed04b83d-c180-4bf2-b6b9-b112a536d9da" (UID: "ed04b83d-c180-4bf2-b6b9-b112a536d9da"). InnerVolumeSpecName "kube-api-access-t8w2f". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:36:59 crc kubenswrapper[4631]: I1128 13:36:59.658221 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ed04b83d-c180-4bf2-b6b9-b112a536d9da-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "ed04b83d-c180-4bf2-b6b9-b112a536d9da" (UID: "ed04b83d-c180-4bf2-b6b9-b112a536d9da"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 13:36:59 crc kubenswrapper[4631]: I1128 13:36:59.691232 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t8w2f\" (UniqueName: \"kubernetes.io/projected/ed04b83d-c180-4bf2-b6b9-b112a536d9da-kube-api-access-t8w2f\") on node \"crc\" DevicePath \"\"" Nov 28 13:36:59 crc kubenswrapper[4631]: I1128 13:36:59.692115 4631 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ed04b83d-c180-4bf2-b6b9-b112a536d9da-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 28 13:36:59 crc kubenswrapper[4631]: I1128 13:36:59.712596 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ed04b83d-c180-4bf2-b6b9-b112a536d9da-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "ed04b83d-c180-4bf2-b6b9-b112a536d9da" (UID: "ed04b83d-c180-4bf2-b6b9-b112a536d9da"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 13:36:59 crc kubenswrapper[4631]: I1128 13:36:59.739011 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ed04b83d-c180-4bf2-b6b9-b112a536d9da-config" (OuterVolumeSpecName: "config") pod "ed04b83d-c180-4bf2-b6b9-b112a536d9da" (UID: "ed04b83d-c180-4bf2-b6b9-b112a536d9da"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 13:36:59 crc kubenswrapper[4631]: I1128 13:36:59.751547 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ed04b83d-c180-4bf2-b6b9-b112a536d9da-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "ed04b83d-c180-4bf2-b6b9-b112a536d9da" (UID: "ed04b83d-c180-4bf2-b6b9-b112a536d9da"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 13:36:59 crc kubenswrapper[4631]: I1128 13:36:59.794352 4631 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ed04b83d-c180-4bf2-b6b9-b112a536d9da-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 28 13:36:59 crc kubenswrapper[4631]: I1128 13:36:59.794412 4631 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ed04b83d-c180-4bf2-b6b9-b112a536d9da-config\") on node \"crc\" DevicePath \"\"" Nov 28 13:36:59 crc kubenswrapper[4631]: I1128 13:36:59.794425 4631 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ed04b83d-c180-4bf2-b6b9-b112a536d9da-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 28 13:36:59 crc kubenswrapper[4631]: I1128 13:36:59.860640 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-hplqp" Nov 28 13:37:00 crc kubenswrapper[4631]: I1128 13:37:00.025913 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e8fd4db1-5308-414f-9eac-ddef2c41867a-operator-scripts\") pod \"e8fd4db1-5308-414f-9eac-ddef2c41867a\" (UID: \"e8fd4db1-5308-414f-9eac-ddef2c41867a\") " Nov 28 13:37:00 crc kubenswrapper[4631]: I1128 13:37:00.026017 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gl8xp\" (UniqueName: \"kubernetes.io/projected/e8fd4db1-5308-414f-9eac-ddef2c41867a-kube-api-access-gl8xp\") pod \"e8fd4db1-5308-414f-9eac-ddef2c41867a\" (UID: \"e8fd4db1-5308-414f-9eac-ddef2c41867a\") " Nov 28 13:37:00 crc kubenswrapper[4631]: I1128 13:37:00.028865 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e8fd4db1-5308-414f-9eac-ddef2c41867a-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "e8fd4db1-5308-414f-9eac-ddef2c41867a" (UID: "e8fd4db1-5308-414f-9eac-ddef2c41867a"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 13:37:00 crc kubenswrapper[4631]: I1128 13:37:00.049249 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e8fd4db1-5308-414f-9eac-ddef2c41867a-kube-api-access-gl8xp" (OuterVolumeSpecName: "kube-api-access-gl8xp") pod "e8fd4db1-5308-414f-9eac-ddef2c41867a" (UID: "e8fd4db1-5308-414f-9eac-ddef2c41867a"). InnerVolumeSpecName "kube-api-access-gl8xp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:37:00 crc kubenswrapper[4631]: I1128 13:37:00.141000 4631 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e8fd4db1-5308-414f-9eac-ddef2c41867a-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 13:37:00 crc kubenswrapper[4631]: I1128 13:37:00.141034 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gl8xp\" (UniqueName: \"kubernetes.io/projected/e8fd4db1-5308-414f-9eac-ddef2c41867a-kube-api-access-gl8xp\") on node \"crc\" DevicePath \"\"" Nov 28 13:37:00 crc kubenswrapper[4631]: I1128 13:37:00.296720 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-wrd26" Nov 28 13:37:00 crc kubenswrapper[4631]: I1128 13:37:00.309138 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-99b5-account-create-update-67zx8" Nov 28 13:37:00 crc kubenswrapper[4631]: I1128 13:37:00.314483 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-m2zrc" event={"ID":"ed04b83d-c180-4bf2-b6b9-b112a536d9da","Type":"ContainerDied","Data":"6c9354e875a4d160ec866ea5f7f4933f80b648e00763d7c48e9e07ea5fac8369"} Nov 28 13:37:00 crc kubenswrapper[4631]: I1128 13:37:00.314568 4631 scope.go:117] "RemoveContainer" containerID="6abb47410b6f4f8e4c483aea263ea839b52160f49e3c028e22b32ec972d55780" Nov 28 13:37:00 crc kubenswrapper[4631]: I1128 13:37:00.314804 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-m2zrc" Nov 28 13:37:00 crc kubenswrapper[4631]: I1128 13:37:00.341828 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-99b5-account-create-update-67zx8" event={"ID":"cee396c9-93ab-45dd-82d0-9641e985644a","Type":"ContainerDied","Data":"d125961efe689f993c939f0a9bdf44e9f34f1579ab7afe15f1e85717edc50656"} Nov 28 13:37:00 crc kubenswrapper[4631]: I1128 13:37:00.341914 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-99b5-account-create-update-67zx8" Nov 28 13:37:00 crc kubenswrapper[4631]: I1128 13:37:00.342776 4631 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d125961efe689f993c939f0a9bdf44e9f34f1579ab7afe15f1e85717edc50656" Nov 28 13:37:00 crc kubenswrapper[4631]: I1128 13:37:00.349515 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-8a8c-account-create-update-lskqg" Nov 28 13:37:00 crc kubenswrapper[4631]: I1128 13:37:00.359853 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-hplqp" event={"ID":"e8fd4db1-5308-414f-9eac-ddef2c41867a","Type":"ContainerDied","Data":"bc8aedf9f691384db0116de1386d6b2839d9cd4da28760539c96719ba11b9cda"} Nov 28 13:37:00 crc kubenswrapper[4631]: I1128 13:37:00.359895 4631 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bc8aedf9f691384db0116de1386d6b2839d9cd4da28760539c96719ba11b9cda" Nov 28 13:37:00 crc kubenswrapper[4631]: I1128 13:37:00.359955 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-hplqp" Nov 28 13:37:00 crc kubenswrapper[4631]: I1128 13:37:00.361090 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-c506-account-create-update-nr5s5" Nov 28 13:37:00 crc kubenswrapper[4631]: I1128 13:37:00.363209 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-m4zmz" event={"ID":"95d067ec-b8ae-4286-a35f-eb431d025ac9","Type":"ContainerDied","Data":"cbac8b9ba48d9ebd1b5b5ce43698c8fdd1f016df114dc385990c5ec216ac8ee8"} Nov 28 13:37:00 crc kubenswrapper[4631]: I1128 13:37:00.363261 4631 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cbac8b9ba48d9ebd1b5b5ce43698c8fdd1f016df114dc385990c5ec216ac8ee8" Nov 28 13:37:00 crc kubenswrapper[4631]: I1128 13:37:00.377560 4631 scope.go:117] "RemoveContainer" containerID="eec174b25ab3c2391bf3c823680cdaa945d25dc0520c28fc082487d5af87f766" Nov 28 13:37:00 crc kubenswrapper[4631]: I1128 13:37:00.391202 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-wrd26" event={"ID":"98564bcf-8a3f-4324-af81-bafdc2ed2934","Type":"ContainerDied","Data":"083b9088b5d4d41c287b02703a1c8fbfa46aa8576b48f0e0cfc972d7ab77d79a"} Nov 28 13:37:00 crc kubenswrapper[4631]: I1128 13:37:00.391248 4631 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="083b9088b5d4d41c287b02703a1c8fbfa46aa8576b48f0e0cfc972d7ab77d79a" Nov 28 13:37:00 crc kubenswrapper[4631]: I1128 13:37:00.391555 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-wrd26" Nov 28 13:37:00 crc kubenswrapper[4631]: I1128 13:37:00.397345 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-m4zmz" Nov 28 13:37:00 crc kubenswrapper[4631]: I1128 13:37:00.400985 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-8a8c-account-create-update-lskqg" event={"ID":"127100c3-8f48-4ddf-a289-f08f3cdf0d8a","Type":"ContainerDied","Data":"bdcf0f0360d41edb285cdf11f52b9614dd00c3037eac91c5749e15a7f568846b"} Nov 28 13:37:00 crc kubenswrapper[4631]: I1128 13:37:00.401018 4631 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bdcf0f0360d41edb285cdf11f52b9614dd00c3037eac91c5749e15a7f568846b" Nov 28 13:37:00 crc kubenswrapper[4631]: I1128 13:37:00.401070 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-8a8c-account-create-update-lskqg" Nov 28 13:37:00 crc kubenswrapper[4631]: I1128 13:37:00.412941 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-c506-account-create-update-nr5s5" event={"ID":"ecb43d0a-167c-42a9-9ea9-bc303f5890b8","Type":"ContainerDied","Data":"de1e465e275bc8f34db5322875dd82f091e44b7cb0f3f75e8cca3eed50170f6c"} Nov 28 13:37:00 crc kubenswrapper[4631]: I1128 13:37:00.413006 4631 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="de1e465e275bc8f34db5322875dd82f091e44b7cb0f3f75e8cca3eed50170f6c" Nov 28 13:37:00 crc kubenswrapper[4631]: I1128 13:37:00.414274 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-c506-account-create-update-nr5s5" Nov 28 13:37:00 crc kubenswrapper[4631]: I1128 13:37:00.465025 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lvvjh\" (UniqueName: \"kubernetes.io/projected/cee396c9-93ab-45dd-82d0-9641e985644a-kube-api-access-lvvjh\") pod \"cee396c9-93ab-45dd-82d0-9641e985644a\" (UID: \"cee396c9-93ab-45dd-82d0-9641e985644a\") " Nov 28 13:37:00 crc kubenswrapper[4631]: I1128 13:37:00.465108 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ecb43d0a-167c-42a9-9ea9-bc303f5890b8-operator-scripts\") pod \"ecb43d0a-167c-42a9-9ea9-bc303f5890b8\" (UID: \"ecb43d0a-167c-42a9-9ea9-bc303f5890b8\") " Nov 28 13:37:00 crc kubenswrapper[4631]: I1128 13:37:00.465180 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/98564bcf-8a3f-4324-af81-bafdc2ed2934-operator-scripts\") pod \"98564bcf-8a3f-4324-af81-bafdc2ed2934\" (UID: \"98564bcf-8a3f-4324-af81-bafdc2ed2934\") " Nov 28 13:37:00 crc kubenswrapper[4631]: I1128 13:37:00.465267 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mcrq6\" (UniqueName: \"kubernetes.io/projected/ecb43d0a-167c-42a9-9ea9-bc303f5890b8-kube-api-access-mcrq6\") pod \"ecb43d0a-167c-42a9-9ea9-bc303f5890b8\" (UID: \"ecb43d0a-167c-42a9-9ea9-bc303f5890b8\") " Nov 28 13:37:00 crc kubenswrapper[4631]: I1128 13:37:00.465344 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/127100c3-8f48-4ddf-a289-f08f3cdf0d8a-operator-scripts\") pod \"127100c3-8f48-4ddf-a289-f08f3cdf0d8a\" (UID: \"127100c3-8f48-4ddf-a289-f08f3cdf0d8a\") " Nov 28 13:37:00 crc kubenswrapper[4631]: I1128 13:37:00.465375 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wvns5\" (UniqueName: \"kubernetes.io/projected/98564bcf-8a3f-4324-af81-bafdc2ed2934-kube-api-access-wvns5\") pod \"98564bcf-8a3f-4324-af81-bafdc2ed2934\" (UID: \"98564bcf-8a3f-4324-af81-bafdc2ed2934\") " Nov 28 13:37:00 crc kubenswrapper[4631]: I1128 13:37:00.465417 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h7nsj\" (UniqueName: \"kubernetes.io/projected/127100c3-8f48-4ddf-a289-f08f3cdf0d8a-kube-api-access-h7nsj\") pod \"127100c3-8f48-4ddf-a289-f08f3cdf0d8a\" (UID: \"127100c3-8f48-4ddf-a289-f08f3cdf0d8a\") " Nov 28 13:37:00 crc kubenswrapper[4631]: I1128 13:37:00.465476 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cee396c9-93ab-45dd-82d0-9641e985644a-operator-scripts\") pod \"cee396c9-93ab-45dd-82d0-9641e985644a\" (UID: \"cee396c9-93ab-45dd-82d0-9641e985644a\") " Nov 28 13:37:00 crc kubenswrapper[4631]: I1128 13:37:00.466461 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cee396c9-93ab-45dd-82d0-9641e985644a-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "cee396c9-93ab-45dd-82d0-9641e985644a" (UID: "cee396c9-93ab-45dd-82d0-9641e985644a"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 13:37:00 crc kubenswrapper[4631]: I1128 13:37:00.470907 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ecb43d0a-167c-42a9-9ea9-bc303f5890b8-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "ecb43d0a-167c-42a9-9ea9-bc303f5890b8" (UID: "ecb43d0a-167c-42a9-9ea9-bc303f5890b8"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 13:37:00 crc kubenswrapper[4631]: I1128 13:37:00.471203 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/98564bcf-8a3f-4324-af81-bafdc2ed2934-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "98564bcf-8a3f-4324-af81-bafdc2ed2934" (UID: "98564bcf-8a3f-4324-af81-bafdc2ed2934"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 13:37:00 crc kubenswrapper[4631]: I1128 13:37:00.471653 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/127100c3-8f48-4ddf-a289-f08f3cdf0d8a-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "127100c3-8f48-4ddf-a289-f08f3cdf0d8a" (UID: "127100c3-8f48-4ddf-a289-f08f3cdf0d8a"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 13:37:00 crc kubenswrapper[4631]: I1128 13:37:00.476210 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/98564bcf-8a3f-4324-af81-bafdc2ed2934-kube-api-access-wvns5" (OuterVolumeSpecName: "kube-api-access-wvns5") pod "98564bcf-8a3f-4324-af81-bafdc2ed2934" (UID: "98564bcf-8a3f-4324-af81-bafdc2ed2934"). InnerVolumeSpecName "kube-api-access-wvns5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:37:00 crc kubenswrapper[4631]: I1128 13:37:00.479610 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cee396c9-93ab-45dd-82d0-9641e985644a-kube-api-access-lvvjh" (OuterVolumeSpecName: "kube-api-access-lvvjh") pod "cee396c9-93ab-45dd-82d0-9641e985644a" (UID: "cee396c9-93ab-45dd-82d0-9641e985644a"). InnerVolumeSpecName "kube-api-access-lvvjh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:37:00 crc kubenswrapper[4631]: I1128 13:37:00.479686 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ecb43d0a-167c-42a9-9ea9-bc303f5890b8-kube-api-access-mcrq6" (OuterVolumeSpecName: "kube-api-access-mcrq6") pod "ecb43d0a-167c-42a9-9ea9-bc303f5890b8" (UID: "ecb43d0a-167c-42a9-9ea9-bc303f5890b8"). InnerVolumeSpecName "kube-api-access-mcrq6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:37:00 crc kubenswrapper[4631]: I1128 13:37:00.481019 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/127100c3-8f48-4ddf-a289-f08f3cdf0d8a-kube-api-access-h7nsj" (OuterVolumeSpecName: "kube-api-access-h7nsj") pod "127100c3-8f48-4ddf-a289-f08f3cdf0d8a" (UID: "127100c3-8f48-4ddf-a289-f08f3cdf0d8a"). InnerVolumeSpecName "kube-api-access-h7nsj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:37:00 crc kubenswrapper[4631]: I1128 13:37:00.483588 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-698758b865-m2zrc"] Nov 28 13:37:00 crc kubenswrapper[4631]: I1128 13:37:00.498713 4631 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-698758b865-m2zrc"] Nov 28 13:37:00 crc kubenswrapper[4631]: W1128 13:37:00.541583 4631 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod166f0f43_e5cb_42ad_a2f5_1b207accc673.slice/crio-53ac8e91a456a4174a50d67629061bd3506db6839b116a69454647eb2c4a7266 WatchSource:0}: Error finding container 53ac8e91a456a4174a50d67629061bd3506db6839b116a69454647eb2c4a7266: Status 404 returned error can't find the container with id 53ac8e91a456a4174a50d67629061bd3506db6839b116a69454647eb2c4a7266 Nov 28 13:37:00 crc kubenswrapper[4631]: I1128 13:37:00.552129 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-74f6bcbc87-r8g9l"] Nov 28 13:37:00 crc kubenswrapper[4631]: E1128 13:37:00.561662 4631 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode8fd4db1_5308_414f_9eac_ddef2c41867a.slice/crio-bc8aedf9f691384db0116de1386d6b2839d9cd4da28760539c96719ba11b9cda\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poded04b83d_c180_4bf2_b6b9_b112a536d9da.slice/crio-6c9354e875a4d160ec866ea5f7f4933f80b648e00763d7c48e9e07ea5fac8369\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poded04b83d_c180_4bf2_b6b9_b112a536d9da.slice\": RecentStats: unable to find data in memory cache]" Nov 28 13:37:00 crc kubenswrapper[4631]: I1128 13:37:00.566985 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/95d067ec-b8ae-4286-a35f-eb431d025ac9-operator-scripts\") pod \"95d067ec-b8ae-4286-a35f-eb431d025ac9\" (UID: \"95d067ec-b8ae-4286-a35f-eb431d025ac9\") " Nov 28 13:37:00 crc kubenswrapper[4631]: I1128 13:37:00.567693 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/95d067ec-b8ae-4286-a35f-eb431d025ac9-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "95d067ec-b8ae-4286-a35f-eb431d025ac9" (UID: "95d067ec-b8ae-4286-a35f-eb431d025ac9"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 13:37:00 crc kubenswrapper[4631]: I1128 13:37:00.567990 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ss6q5\" (UniqueName: \"kubernetes.io/projected/95d067ec-b8ae-4286-a35f-eb431d025ac9-kube-api-access-ss6q5\") pod \"95d067ec-b8ae-4286-a35f-eb431d025ac9\" (UID: \"95d067ec-b8ae-4286-a35f-eb431d025ac9\") " Nov 28 13:37:00 crc kubenswrapper[4631]: I1128 13:37:00.568764 4631 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/127100c3-8f48-4ddf-a289-f08f3cdf0d8a-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 13:37:00 crc kubenswrapper[4631]: I1128 13:37:00.568896 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wvns5\" (UniqueName: \"kubernetes.io/projected/98564bcf-8a3f-4324-af81-bafdc2ed2934-kube-api-access-wvns5\") on node \"crc\" DevicePath \"\"" Nov 28 13:37:00 crc kubenswrapper[4631]: I1128 13:37:00.568976 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h7nsj\" (UniqueName: \"kubernetes.io/projected/127100c3-8f48-4ddf-a289-f08f3cdf0d8a-kube-api-access-h7nsj\") on node \"crc\" DevicePath \"\"" Nov 28 13:37:00 crc kubenswrapper[4631]: I1128 13:37:00.569049 4631 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cee396c9-93ab-45dd-82d0-9641e985644a-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 13:37:00 crc kubenswrapper[4631]: I1128 13:37:00.569110 4631 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/95d067ec-b8ae-4286-a35f-eb431d025ac9-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 13:37:00 crc kubenswrapper[4631]: I1128 13:37:00.569178 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lvvjh\" (UniqueName: \"kubernetes.io/projected/cee396c9-93ab-45dd-82d0-9641e985644a-kube-api-access-lvvjh\") on node \"crc\" DevicePath \"\"" Nov 28 13:37:00 crc kubenswrapper[4631]: I1128 13:37:00.569235 4631 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ecb43d0a-167c-42a9-9ea9-bc303f5890b8-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 13:37:00 crc kubenswrapper[4631]: I1128 13:37:00.569322 4631 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/98564bcf-8a3f-4324-af81-bafdc2ed2934-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 13:37:00 crc kubenswrapper[4631]: I1128 13:37:00.569382 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mcrq6\" (UniqueName: \"kubernetes.io/projected/ecb43d0a-167c-42a9-9ea9-bc303f5890b8-kube-api-access-mcrq6\") on node \"crc\" DevicePath \"\"" Nov 28 13:37:00 crc kubenswrapper[4631]: I1128 13:37:00.574349 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/95d067ec-b8ae-4286-a35f-eb431d025ac9-kube-api-access-ss6q5" (OuterVolumeSpecName: "kube-api-access-ss6q5") pod "95d067ec-b8ae-4286-a35f-eb431d025ac9" (UID: "95d067ec-b8ae-4286-a35f-eb431d025ac9"). InnerVolumeSpecName "kube-api-access-ss6q5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:37:00 crc kubenswrapper[4631]: I1128 13:37:00.671082 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ss6q5\" (UniqueName: \"kubernetes.io/projected/95d067ec-b8ae-4286-a35f-eb431d025ac9-kube-api-access-ss6q5\") on node \"crc\" DevicePath \"\"" Nov 28 13:37:01 crc kubenswrapper[4631]: I1128 13:37:01.446248 4631 generic.go:334] "Generic (PLEG): container finished" podID="166f0f43-e5cb-42ad-a2f5-1b207accc673" containerID="d35b76cd7932fe54552de2a63b3892d588152fa871b777d24b6c7c5406ac5703" exitCode=0 Nov 28 13:37:01 crc kubenswrapper[4631]: I1128 13:37:01.446829 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74f6bcbc87-r8g9l" event={"ID":"166f0f43-e5cb-42ad-a2f5-1b207accc673","Type":"ContainerDied","Data":"d35b76cd7932fe54552de2a63b3892d588152fa871b777d24b6c7c5406ac5703"} Nov 28 13:37:01 crc kubenswrapper[4631]: I1128 13:37:01.446881 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74f6bcbc87-r8g9l" event={"ID":"166f0f43-e5cb-42ad-a2f5-1b207accc673","Type":"ContainerStarted","Data":"53ac8e91a456a4174a50d67629061bd3506db6839b116a69454647eb2c4a7266"} Nov 28 13:37:01 crc kubenswrapper[4631]: I1128 13:37:01.448499 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-m4zmz" Nov 28 13:37:01 crc kubenswrapper[4631]: I1128 13:37:01.530551 4631 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ed04b83d-c180-4bf2-b6b9-b112a536d9da" path="/var/lib/kubelet/pods/ed04b83d-c180-4bf2-b6b9-b112a536d9da/volumes" Nov 28 13:37:05 crc kubenswrapper[4631]: I1128 13:37:05.495779 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-xs457" event={"ID":"b6844d09-9bb8-44a7-bf9a-eca18ce5c1c8","Type":"ContainerStarted","Data":"8b91851c1160b6f6360af453ef82165da62b7b53a25c4fb8390d0871f3df8f25"} Nov 28 13:37:05 crc kubenswrapper[4631]: I1128 13:37:05.499996 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74f6bcbc87-r8g9l" event={"ID":"166f0f43-e5cb-42ad-a2f5-1b207accc673","Type":"ContainerStarted","Data":"e9a7a7a5cf57eda39ae2aac9a52f8fe72136a44113973538ac6ec5d2590a54ef"} Nov 28 13:37:05 crc kubenswrapper[4631]: I1128 13:37:05.500621 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-74f6bcbc87-r8g9l" Nov 28 13:37:05 crc kubenswrapper[4631]: I1128 13:37:05.545369 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-xs457" podStartSLOduration=3.606609591 podStartE2EDuration="12.54534202s" podCreationTimestamp="2025-11-28 13:36:53 +0000 UTC" firstStartedPulling="2025-11-28 13:36:55.550466619 +0000 UTC m=+972.357769953" lastFinishedPulling="2025-11-28 13:37:04.489199038 +0000 UTC m=+981.296502382" observedRunningTime="2025-11-28 13:37:05.524266282 +0000 UTC m=+982.331569626" watchObservedRunningTime="2025-11-28 13:37:05.54534202 +0000 UTC m=+982.352645364" Nov 28 13:37:09 crc kubenswrapper[4631]: I1128 13:37:09.269580 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-74f6bcbc87-r8g9l" Nov 28 13:37:09 crc kubenswrapper[4631]: I1128 13:37:09.300089 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-74f6bcbc87-r8g9l" podStartSLOduration=11.300065165 podStartE2EDuration="11.300065165s" podCreationTimestamp="2025-11-28 13:36:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 13:37:05.561553929 +0000 UTC m=+982.368857273" watchObservedRunningTime="2025-11-28 13:37:09.300065165 +0000 UTC m=+986.107368509" Nov 28 13:37:09 crc kubenswrapper[4631]: I1128 13:37:09.379042 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5b946c75cc-rnp88"] Nov 28 13:37:09 crc kubenswrapper[4631]: I1128 13:37:09.379318 4631 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5b946c75cc-rnp88" podUID="684ed85e-280f-4cf7-b528-45284096b772" containerName="dnsmasq-dns" containerID="cri-o://d2f07cceea9a52630a5bf19886f3266f82037c6d0807528ab6e1a449f221a276" gracePeriod=10 Nov 28 13:37:09 crc kubenswrapper[4631]: I1128 13:37:09.592475 4631 generic.go:334] "Generic (PLEG): container finished" podID="684ed85e-280f-4cf7-b528-45284096b772" containerID="d2f07cceea9a52630a5bf19886f3266f82037c6d0807528ab6e1a449f221a276" exitCode=0 Nov 28 13:37:09 crc kubenswrapper[4631]: I1128 13:37:09.592567 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b946c75cc-rnp88" event={"ID":"684ed85e-280f-4cf7-b528-45284096b772","Type":"ContainerDied","Data":"d2f07cceea9a52630a5bf19886f3266f82037c6d0807528ab6e1a449f221a276"} Nov 28 13:37:09 crc kubenswrapper[4631]: I1128 13:37:09.981968 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b946c75cc-rnp88" Nov 28 13:37:10 crc kubenswrapper[4631]: I1128 13:37:10.000787 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/684ed85e-280f-4cf7-b528-45284096b772-ovsdbserver-sb\") pod \"684ed85e-280f-4cf7-b528-45284096b772\" (UID: \"684ed85e-280f-4cf7-b528-45284096b772\") " Nov 28 13:37:10 crc kubenswrapper[4631]: I1128 13:37:10.000904 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2sksh\" (UniqueName: \"kubernetes.io/projected/684ed85e-280f-4cf7-b528-45284096b772-kube-api-access-2sksh\") pod \"684ed85e-280f-4cf7-b528-45284096b772\" (UID: \"684ed85e-280f-4cf7-b528-45284096b772\") " Nov 28 13:37:10 crc kubenswrapper[4631]: I1128 13:37:10.000962 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/684ed85e-280f-4cf7-b528-45284096b772-config\") pod \"684ed85e-280f-4cf7-b528-45284096b772\" (UID: \"684ed85e-280f-4cf7-b528-45284096b772\") " Nov 28 13:37:10 crc kubenswrapper[4631]: I1128 13:37:10.001050 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/684ed85e-280f-4cf7-b528-45284096b772-dns-svc\") pod \"684ed85e-280f-4cf7-b528-45284096b772\" (UID: \"684ed85e-280f-4cf7-b528-45284096b772\") " Nov 28 13:37:10 crc kubenswrapper[4631]: I1128 13:37:10.001100 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/684ed85e-280f-4cf7-b528-45284096b772-ovsdbserver-nb\") pod \"684ed85e-280f-4cf7-b528-45284096b772\" (UID: \"684ed85e-280f-4cf7-b528-45284096b772\") " Nov 28 13:37:10 crc kubenswrapper[4631]: I1128 13:37:10.031471 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/684ed85e-280f-4cf7-b528-45284096b772-kube-api-access-2sksh" (OuterVolumeSpecName: "kube-api-access-2sksh") pod "684ed85e-280f-4cf7-b528-45284096b772" (UID: "684ed85e-280f-4cf7-b528-45284096b772"). InnerVolumeSpecName "kube-api-access-2sksh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:37:10 crc kubenswrapper[4631]: I1128 13:37:10.060189 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/684ed85e-280f-4cf7-b528-45284096b772-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "684ed85e-280f-4cf7-b528-45284096b772" (UID: "684ed85e-280f-4cf7-b528-45284096b772"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 13:37:10 crc kubenswrapper[4631]: I1128 13:37:10.080587 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/684ed85e-280f-4cf7-b528-45284096b772-config" (OuterVolumeSpecName: "config") pod "684ed85e-280f-4cf7-b528-45284096b772" (UID: "684ed85e-280f-4cf7-b528-45284096b772"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 13:37:10 crc kubenswrapper[4631]: I1128 13:37:10.091569 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/684ed85e-280f-4cf7-b528-45284096b772-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "684ed85e-280f-4cf7-b528-45284096b772" (UID: "684ed85e-280f-4cf7-b528-45284096b772"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 13:37:10 crc kubenswrapper[4631]: I1128 13:37:10.104309 4631 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/684ed85e-280f-4cf7-b528-45284096b772-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 28 13:37:10 crc kubenswrapper[4631]: I1128 13:37:10.104418 4631 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/684ed85e-280f-4cf7-b528-45284096b772-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 28 13:37:10 crc kubenswrapper[4631]: I1128 13:37:10.104440 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2sksh\" (UniqueName: \"kubernetes.io/projected/684ed85e-280f-4cf7-b528-45284096b772-kube-api-access-2sksh\") on node \"crc\" DevicePath \"\"" Nov 28 13:37:10 crc kubenswrapper[4631]: I1128 13:37:10.104695 4631 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/684ed85e-280f-4cf7-b528-45284096b772-config\") on node \"crc\" DevicePath \"\"" Nov 28 13:37:10 crc kubenswrapper[4631]: I1128 13:37:10.108173 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/684ed85e-280f-4cf7-b528-45284096b772-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "684ed85e-280f-4cf7-b528-45284096b772" (UID: "684ed85e-280f-4cf7-b528-45284096b772"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 13:37:10 crc kubenswrapper[4631]: I1128 13:37:10.205899 4631 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/684ed85e-280f-4cf7-b528-45284096b772-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 28 13:37:10 crc kubenswrapper[4631]: I1128 13:37:10.604114 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b946c75cc-rnp88" event={"ID":"684ed85e-280f-4cf7-b528-45284096b772","Type":"ContainerDied","Data":"923378e21e4a6ca1879e802a4ec08326506801a1c2f51b1d8526285e110919b8"} Nov 28 13:37:10 crc kubenswrapper[4631]: I1128 13:37:10.604221 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b946c75cc-rnp88" Nov 28 13:37:10 crc kubenswrapper[4631]: I1128 13:37:10.604679 4631 scope.go:117] "RemoveContainer" containerID="d2f07cceea9a52630a5bf19886f3266f82037c6d0807528ab6e1a449f221a276" Nov 28 13:37:10 crc kubenswrapper[4631]: I1128 13:37:10.657663 4631 scope.go:117] "RemoveContainer" containerID="e88aa3e2ea1bd4be9bab880af18e21ed1a97e0e89a7e7d8d890346fcb7edd576" Nov 28 13:37:10 crc kubenswrapper[4631]: I1128 13:37:10.672331 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5b946c75cc-rnp88"] Nov 28 13:37:10 crc kubenswrapper[4631]: I1128 13:37:10.681380 4631 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5b946c75cc-rnp88"] Nov 28 13:37:10 crc kubenswrapper[4631]: E1128 13:37:10.787631 4631 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod684ed85e_280f_4cf7_b528_45284096b772.slice/crio-923378e21e4a6ca1879e802a4ec08326506801a1c2f51b1d8526285e110919b8\": RecentStats: unable to find data in memory cache]" Nov 28 13:37:11 crc kubenswrapper[4631]: I1128 13:37:11.521390 4631 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="684ed85e-280f-4cf7-b528-45284096b772" path="/var/lib/kubelet/pods/684ed85e-280f-4cf7-b528-45284096b772/volumes" Nov 28 13:37:11 crc kubenswrapper[4631]: I1128 13:37:11.614933 4631 generic.go:334] "Generic (PLEG): container finished" podID="b6844d09-9bb8-44a7-bf9a-eca18ce5c1c8" containerID="8b91851c1160b6f6360af453ef82165da62b7b53a25c4fb8390d0871f3df8f25" exitCode=0 Nov 28 13:37:11 crc kubenswrapper[4631]: I1128 13:37:11.615012 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-xs457" event={"ID":"b6844d09-9bb8-44a7-bf9a-eca18ce5c1c8","Type":"ContainerDied","Data":"8b91851c1160b6f6360af453ef82165da62b7b53a25c4fb8390d0871f3df8f25"} Nov 28 13:37:12 crc kubenswrapper[4631]: I1128 13:37:12.966798 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-xs457" Nov 28 13:37:13 crc kubenswrapper[4631]: I1128 13:37:13.048920 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b6844d09-9bb8-44a7-bf9a-eca18ce5c1c8-config-data\") pod \"b6844d09-9bb8-44a7-bf9a-eca18ce5c1c8\" (UID: \"b6844d09-9bb8-44a7-bf9a-eca18ce5c1c8\") " Nov 28 13:37:13 crc kubenswrapper[4631]: I1128 13:37:13.049258 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b6844d09-9bb8-44a7-bf9a-eca18ce5c1c8-combined-ca-bundle\") pod \"b6844d09-9bb8-44a7-bf9a-eca18ce5c1c8\" (UID: \"b6844d09-9bb8-44a7-bf9a-eca18ce5c1c8\") " Nov 28 13:37:13 crc kubenswrapper[4631]: I1128 13:37:13.049389 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wnwzq\" (UniqueName: \"kubernetes.io/projected/b6844d09-9bb8-44a7-bf9a-eca18ce5c1c8-kube-api-access-wnwzq\") pod \"b6844d09-9bb8-44a7-bf9a-eca18ce5c1c8\" (UID: \"b6844d09-9bb8-44a7-bf9a-eca18ce5c1c8\") " Nov 28 13:37:13 crc kubenswrapper[4631]: I1128 13:37:13.088367 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6844d09-9bb8-44a7-bf9a-eca18ce5c1c8-kube-api-access-wnwzq" (OuterVolumeSpecName: "kube-api-access-wnwzq") pod "b6844d09-9bb8-44a7-bf9a-eca18ce5c1c8" (UID: "b6844d09-9bb8-44a7-bf9a-eca18ce5c1c8"). InnerVolumeSpecName "kube-api-access-wnwzq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:37:13 crc kubenswrapper[4631]: I1128 13:37:13.111825 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6844d09-9bb8-44a7-bf9a-eca18ce5c1c8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b6844d09-9bb8-44a7-bf9a-eca18ce5c1c8" (UID: "b6844d09-9bb8-44a7-bf9a-eca18ce5c1c8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:37:13 crc kubenswrapper[4631]: I1128 13:37:13.120481 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6844d09-9bb8-44a7-bf9a-eca18ce5c1c8-config-data" (OuterVolumeSpecName: "config-data") pod "b6844d09-9bb8-44a7-bf9a-eca18ce5c1c8" (UID: "b6844d09-9bb8-44a7-bf9a-eca18ce5c1c8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:37:13 crc kubenswrapper[4631]: I1128 13:37:13.151675 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wnwzq\" (UniqueName: \"kubernetes.io/projected/b6844d09-9bb8-44a7-bf9a-eca18ce5c1c8-kube-api-access-wnwzq\") on node \"crc\" DevicePath \"\"" Nov 28 13:37:13 crc kubenswrapper[4631]: I1128 13:37:13.151719 4631 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b6844d09-9bb8-44a7-bf9a-eca18ce5c1c8-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 13:37:13 crc kubenswrapper[4631]: I1128 13:37:13.151730 4631 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b6844d09-9bb8-44a7-bf9a-eca18ce5c1c8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 13:37:13 crc kubenswrapper[4631]: I1128 13:37:13.635829 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-xs457" event={"ID":"b6844d09-9bb8-44a7-bf9a-eca18ce5c1c8","Type":"ContainerDied","Data":"f4adc0bdd4d8461a300e92778ea0211b15567cec11c29f5da558ca7ac66064d8"} Nov 28 13:37:13 crc kubenswrapper[4631]: I1128 13:37:13.635884 4631 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f4adc0bdd4d8461a300e92778ea0211b15567cec11c29f5da558ca7ac66064d8" Nov 28 13:37:13 crc kubenswrapper[4631]: I1128 13:37:13.635975 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-xs457" Nov 28 13:37:13 crc kubenswrapper[4631]: I1128 13:37:13.990263 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-847c4cc679-mjczg"] Nov 28 13:37:14 crc kubenswrapper[4631]: E1128 13:37:14.001956 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="127100c3-8f48-4ddf-a289-f08f3cdf0d8a" containerName="mariadb-account-create-update" Nov 28 13:37:14 crc kubenswrapper[4631]: I1128 13:37:14.002017 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="127100c3-8f48-4ddf-a289-f08f3cdf0d8a" containerName="mariadb-account-create-update" Nov 28 13:37:14 crc kubenswrapper[4631]: E1128 13:37:14.002043 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b6844d09-9bb8-44a7-bf9a-eca18ce5c1c8" containerName="keystone-db-sync" Nov 28 13:37:14 crc kubenswrapper[4631]: I1128 13:37:14.002052 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="b6844d09-9bb8-44a7-bf9a-eca18ce5c1c8" containerName="keystone-db-sync" Nov 28 13:37:14 crc kubenswrapper[4631]: E1128 13:37:14.002072 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="98564bcf-8a3f-4324-af81-bafdc2ed2934" containerName="mariadb-database-create" Nov 28 13:37:14 crc kubenswrapper[4631]: I1128 13:37:14.002081 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="98564bcf-8a3f-4324-af81-bafdc2ed2934" containerName="mariadb-database-create" Nov 28 13:37:14 crc kubenswrapper[4631]: E1128 13:37:14.002114 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ed04b83d-c180-4bf2-b6b9-b112a536d9da" containerName="init" Nov 28 13:37:14 crc kubenswrapper[4631]: I1128 13:37:14.002135 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="ed04b83d-c180-4bf2-b6b9-b112a536d9da" containerName="init" Nov 28 13:37:14 crc kubenswrapper[4631]: E1128 13:37:14.002149 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e8fd4db1-5308-414f-9eac-ddef2c41867a" containerName="mariadb-database-create" Nov 28 13:37:14 crc kubenswrapper[4631]: I1128 13:37:14.002157 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="e8fd4db1-5308-414f-9eac-ddef2c41867a" containerName="mariadb-database-create" Nov 28 13:37:14 crc kubenswrapper[4631]: E1128 13:37:14.002175 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ed04b83d-c180-4bf2-b6b9-b112a536d9da" containerName="dnsmasq-dns" Nov 28 13:37:14 crc kubenswrapper[4631]: I1128 13:37:14.002182 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="ed04b83d-c180-4bf2-b6b9-b112a536d9da" containerName="dnsmasq-dns" Nov 28 13:37:14 crc kubenswrapper[4631]: E1128 13:37:14.002199 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ecb43d0a-167c-42a9-9ea9-bc303f5890b8" containerName="mariadb-account-create-update" Nov 28 13:37:14 crc kubenswrapper[4631]: I1128 13:37:14.002217 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="ecb43d0a-167c-42a9-9ea9-bc303f5890b8" containerName="mariadb-account-create-update" Nov 28 13:37:14 crc kubenswrapper[4631]: E1128 13:37:14.002229 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="684ed85e-280f-4cf7-b528-45284096b772" containerName="init" Nov 28 13:37:14 crc kubenswrapper[4631]: I1128 13:37:14.002237 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="684ed85e-280f-4cf7-b528-45284096b772" containerName="init" Nov 28 13:37:14 crc kubenswrapper[4631]: E1128 13:37:14.002259 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cee396c9-93ab-45dd-82d0-9641e985644a" containerName="mariadb-account-create-update" Nov 28 13:37:14 crc kubenswrapper[4631]: I1128 13:37:14.002266 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="cee396c9-93ab-45dd-82d0-9641e985644a" containerName="mariadb-account-create-update" Nov 28 13:37:14 crc kubenswrapper[4631]: E1128 13:37:14.002276 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95d067ec-b8ae-4286-a35f-eb431d025ac9" containerName="mariadb-database-create" Nov 28 13:37:14 crc kubenswrapper[4631]: I1128 13:37:14.002302 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="95d067ec-b8ae-4286-a35f-eb431d025ac9" containerName="mariadb-database-create" Nov 28 13:37:14 crc kubenswrapper[4631]: E1128 13:37:14.002321 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="684ed85e-280f-4cf7-b528-45284096b772" containerName="dnsmasq-dns" Nov 28 13:37:14 crc kubenswrapper[4631]: I1128 13:37:14.002329 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="684ed85e-280f-4cf7-b528-45284096b772" containerName="dnsmasq-dns" Nov 28 13:37:14 crc kubenswrapper[4631]: I1128 13:37:14.002567 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="cee396c9-93ab-45dd-82d0-9641e985644a" containerName="mariadb-account-create-update" Nov 28 13:37:14 crc kubenswrapper[4631]: I1128 13:37:14.002594 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="e8fd4db1-5308-414f-9eac-ddef2c41867a" containerName="mariadb-database-create" Nov 28 13:37:14 crc kubenswrapper[4631]: I1128 13:37:14.002623 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="b6844d09-9bb8-44a7-bf9a-eca18ce5c1c8" containerName="keystone-db-sync" Nov 28 13:37:14 crc kubenswrapper[4631]: I1128 13:37:14.002639 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="98564bcf-8a3f-4324-af81-bafdc2ed2934" containerName="mariadb-database-create" Nov 28 13:37:14 crc kubenswrapper[4631]: I1128 13:37:14.002656 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="ecb43d0a-167c-42a9-9ea9-bc303f5890b8" containerName="mariadb-account-create-update" Nov 28 13:37:14 crc kubenswrapper[4631]: I1128 13:37:14.002678 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="ed04b83d-c180-4bf2-b6b9-b112a536d9da" containerName="dnsmasq-dns" Nov 28 13:37:14 crc kubenswrapper[4631]: I1128 13:37:14.002690 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="127100c3-8f48-4ddf-a289-f08f3cdf0d8a" containerName="mariadb-account-create-update" Nov 28 13:37:14 crc kubenswrapper[4631]: I1128 13:37:14.002704 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="684ed85e-280f-4cf7-b528-45284096b772" containerName="dnsmasq-dns" Nov 28 13:37:14 crc kubenswrapper[4631]: I1128 13:37:14.002719 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="95d067ec-b8ae-4286-a35f-eb431d025ac9" containerName="mariadb-database-create" Nov 28 13:37:14 crc kubenswrapper[4631]: I1128 13:37:14.004059 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-847c4cc679-mjczg" Nov 28 13:37:14 crc kubenswrapper[4631]: I1128 13:37:14.015636 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-65dxp"] Nov 28 13:37:14 crc kubenswrapper[4631]: I1128 13:37:14.024495 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-65dxp" Nov 28 13:37:14 crc kubenswrapper[4631]: I1128 13:37:14.034105 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-szfjn" Nov 28 13:37:14 crc kubenswrapper[4631]: I1128 13:37:14.039085 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Nov 28 13:37:14 crc kubenswrapper[4631]: I1128 13:37:14.039332 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Nov 28 13:37:14 crc kubenswrapper[4631]: I1128 13:37:14.039357 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Nov 28 13:37:14 crc kubenswrapper[4631]: I1128 13:37:14.039529 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Nov 28 13:37:14 crc kubenswrapper[4631]: I1128 13:37:14.055649 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-847c4cc679-mjczg"] Nov 28 13:37:14 crc kubenswrapper[4631]: I1128 13:37:14.094237 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-65dxp"] Nov 28 13:37:14 crc kubenswrapper[4631]: I1128 13:37:14.168552 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ccdbbd8a-4aeb-4294-b047-754cac601f3d-ovsdbserver-nb\") pod \"dnsmasq-dns-847c4cc679-mjczg\" (UID: \"ccdbbd8a-4aeb-4294-b047-754cac601f3d\") " pod="openstack/dnsmasq-dns-847c4cc679-mjczg" Nov 28 13:37:14 crc kubenswrapper[4631]: I1128 13:37:14.168614 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/edf3a98f-0cd6-4eb7-8bf9-0320ae2186c7-config-data\") pod \"keystone-bootstrap-65dxp\" (UID: \"edf3a98f-0cd6-4eb7-8bf9-0320ae2186c7\") " pod="openstack/keystone-bootstrap-65dxp" Nov 28 13:37:14 crc kubenswrapper[4631]: I1128 13:37:14.168661 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ccdbbd8a-4aeb-4294-b047-754cac601f3d-config\") pod \"dnsmasq-dns-847c4cc679-mjczg\" (UID: \"ccdbbd8a-4aeb-4294-b047-754cac601f3d\") " pod="openstack/dnsmasq-dns-847c4cc679-mjczg" Nov 28 13:37:14 crc kubenswrapper[4631]: I1128 13:37:14.168688 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/edf3a98f-0cd6-4eb7-8bf9-0320ae2186c7-credential-keys\") pod \"keystone-bootstrap-65dxp\" (UID: \"edf3a98f-0cd6-4eb7-8bf9-0320ae2186c7\") " pod="openstack/keystone-bootstrap-65dxp" Nov 28 13:37:14 crc kubenswrapper[4631]: I1128 13:37:14.168729 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ccdbbd8a-4aeb-4294-b047-754cac601f3d-dns-svc\") pod \"dnsmasq-dns-847c4cc679-mjczg\" (UID: \"ccdbbd8a-4aeb-4294-b047-754cac601f3d\") " pod="openstack/dnsmasq-dns-847c4cc679-mjczg" Nov 28 13:37:14 crc kubenswrapper[4631]: I1128 13:37:14.168750 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qbmh5\" (UniqueName: \"kubernetes.io/projected/ccdbbd8a-4aeb-4294-b047-754cac601f3d-kube-api-access-qbmh5\") pod \"dnsmasq-dns-847c4cc679-mjczg\" (UID: \"ccdbbd8a-4aeb-4294-b047-754cac601f3d\") " pod="openstack/dnsmasq-dns-847c4cc679-mjczg" Nov 28 13:37:14 crc kubenswrapper[4631]: I1128 13:37:14.168791 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/edf3a98f-0cd6-4eb7-8bf9-0320ae2186c7-scripts\") pod \"keystone-bootstrap-65dxp\" (UID: \"edf3a98f-0cd6-4eb7-8bf9-0320ae2186c7\") " pod="openstack/keystone-bootstrap-65dxp" Nov 28 13:37:14 crc kubenswrapper[4631]: I1128 13:37:14.168821 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dfmp6\" (UniqueName: \"kubernetes.io/projected/edf3a98f-0cd6-4eb7-8bf9-0320ae2186c7-kube-api-access-dfmp6\") pod \"keystone-bootstrap-65dxp\" (UID: \"edf3a98f-0cd6-4eb7-8bf9-0320ae2186c7\") " pod="openstack/keystone-bootstrap-65dxp" Nov 28 13:37:14 crc kubenswrapper[4631]: I1128 13:37:14.168838 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/edf3a98f-0cd6-4eb7-8bf9-0320ae2186c7-fernet-keys\") pod \"keystone-bootstrap-65dxp\" (UID: \"edf3a98f-0cd6-4eb7-8bf9-0320ae2186c7\") " pod="openstack/keystone-bootstrap-65dxp" Nov 28 13:37:14 crc kubenswrapper[4631]: I1128 13:37:14.168866 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ccdbbd8a-4aeb-4294-b047-754cac601f3d-ovsdbserver-sb\") pod \"dnsmasq-dns-847c4cc679-mjczg\" (UID: \"ccdbbd8a-4aeb-4294-b047-754cac601f3d\") " pod="openstack/dnsmasq-dns-847c4cc679-mjczg" Nov 28 13:37:14 crc kubenswrapper[4631]: I1128 13:37:14.168890 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/edf3a98f-0cd6-4eb7-8bf9-0320ae2186c7-combined-ca-bundle\") pod \"keystone-bootstrap-65dxp\" (UID: \"edf3a98f-0cd6-4eb7-8bf9-0320ae2186c7\") " pod="openstack/keystone-bootstrap-65dxp" Nov 28 13:37:14 crc kubenswrapper[4631]: I1128 13:37:14.168919 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ccdbbd8a-4aeb-4294-b047-754cac601f3d-dns-swift-storage-0\") pod \"dnsmasq-dns-847c4cc679-mjczg\" (UID: \"ccdbbd8a-4aeb-4294-b047-754cac601f3d\") " pod="openstack/dnsmasq-dns-847c4cc679-mjczg" Nov 28 13:37:14 crc kubenswrapper[4631]: I1128 13:37:14.272529 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ccdbbd8a-4aeb-4294-b047-754cac601f3d-ovsdbserver-nb\") pod \"dnsmasq-dns-847c4cc679-mjczg\" (UID: \"ccdbbd8a-4aeb-4294-b047-754cac601f3d\") " pod="openstack/dnsmasq-dns-847c4cc679-mjczg" Nov 28 13:37:14 crc kubenswrapper[4631]: I1128 13:37:14.272587 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/edf3a98f-0cd6-4eb7-8bf9-0320ae2186c7-config-data\") pod \"keystone-bootstrap-65dxp\" (UID: \"edf3a98f-0cd6-4eb7-8bf9-0320ae2186c7\") " pod="openstack/keystone-bootstrap-65dxp" Nov 28 13:37:14 crc kubenswrapper[4631]: I1128 13:37:14.272626 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ccdbbd8a-4aeb-4294-b047-754cac601f3d-config\") pod \"dnsmasq-dns-847c4cc679-mjczg\" (UID: \"ccdbbd8a-4aeb-4294-b047-754cac601f3d\") " pod="openstack/dnsmasq-dns-847c4cc679-mjczg" Nov 28 13:37:14 crc kubenswrapper[4631]: I1128 13:37:14.272649 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/edf3a98f-0cd6-4eb7-8bf9-0320ae2186c7-credential-keys\") pod \"keystone-bootstrap-65dxp\" (UID: \"edf3a98f-0cd6-4eb7-8bf9-0320ae2186c7\") " pod="openstack/keystone-bootstrap-65dxp" Nov 28 13:37:14 crc kubenswrapper[4631]: I1128 13:37:14.272696 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ccdbbd8a-4aeb-4294-b047-754cac601f3d-dns-svc\") pod \"dnsmasq-dns-847c4cc679-mjczg\" (UID: \"ccdbbd8a-4aeb-4294-b047-754cac601f3d\") " pod="openstack/dnsmasq-dns-847c4cc679-mjczg" Nov 28 13:37:14 crc kubenswrapper[4631]: I1128 13:37:14.272721 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qbmh5\" (UniqueName: \"kubernetes.io/projected/ccdbbd8a-4aeb-4294-b047-754cac601f3d-kube-api-access-qbmh5\") pod \"dnsmasq-dns-847c4cc679-mjczg\" (UID: \"ccdbbd8a-4aeb-4294-b047-754cac601f3d\") " pod="openstack/dnsmasq-dns-847c4cc679-mjczg" Nov 28 13:37:14 crc kubenswrapper[4631]: I1128 13:37:14.272761 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/edf3a98f-0cd6-4eb7-8bf9-0320ae2186c7-scripts\") pod \"keystone-bootstrap-65dxp\" (UID: \"edf3a98f-0cd6-4eb7-8bf9-0320ae2186c7\") " pod="openstack/keystone-bootstrap-65dxp" Nov 28 13:37:14 crc kubenswrapper[4631]: I1128 13:37:14.272803 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dfmp6\" (UniqueName: \"kubernetes.io/projected/edf3a98f-0cd6-4eb7-8bf9-0320ae2186c7-kube-api-access-dfmp6\") pod \"keystone-bootstrap-65dxp\" (UID: \"edf3a98f-0cd6-4eb7-8bf9-0320ae2186c7\") " pod="openstack/keystone-bootstrap-65dxp" Nov 28 13:37:14 crc kubenswrapper[4631]: I1128 13:37:14.272828 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/edf3a98f-0cd6-4eb7-8bf9-0320ae2186c7-fernet-keys\") pod \"keystone-bootstrap-65dxp\" (UID: \"edf3a98f-0cd6-4eb7-8bf9-0320ae2186c7\") " pod="openstack/keystone-bootstrap-65dxp" Nov 28 13:37:14 crc kubenswrapper[4631]: I1128 13:37:14.272864 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ccdbbd8a-4aeb-4294-b047-754cac601f3d-ovsdbserver-sb\") pod \"dnsmasq-dns-847c4cc679-mjczg\" (UID: \"ccdbbd8a-4aeb-4294-b047-754cac601f3d\") " pod="openstack/dnsmasq-dns-847c4cc679-mjczg" Nov 28 13:37:14 crc kubenswrapper[4631]: I1128 13:37:14.272893 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/edf3a98f-0cd6-4eb7-8bf9-0320ae2186c7-combined-ca-bundle\") pod \"keystone-bootstrap-65dxp\" (UID: \"edf3a98f-0cd6-4eb7-8bf9-0320ae2186c7\") " pod="openstack/keystone-bootstrap-65dxp" Nov 28 13:37:14 crc kubenswrapper[4631]: I1128 13:37:14.272927 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ccdbbd8a-4aeb-4294-b047-754cac601f3d-dns-swift-storage-0\") pod \"dnsmasq-dns-847c4cc679-mjczg\" (UID: \"ccdbbd8a-4aeb-4294-b047-754cac601f3d\") " pod="openstack/dnsmasq-dns-847c4cc679-mjczg" Nov 28 13:37:14 crc kubenswrapper[4631]: I1128 13:37:14.274059 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ccdbbd8a-4aeb-4294-b047-754cac601f3d-dns-swift-storage-0\") pod \"dnsmasq-dns-847c4cc679-mjczg\" (UID: \"ccdbbd8a-4aeb-4294-b047-754cac601f3d\") " pod="openstack/dnsmasq-dns-847c4cc679-mjczg" Nov 28 13:37:14 crc kubenswrapper[4631]: I1128 13:37:14.274755 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ccdbbd8a-4aeb-4294-b047-754cac601f3d-ovsdbserver-nb\") pod \"dnsmasq-dns-847c4cc679-mjczg\" (UID: \"ccdbbd8a-4aeb-4294-b047-754cac601f3d\") " pod="openstack/dnsmasq-dns-847c4cc679-mjczg" Nov 28 13:37:14 crc kubenswrapper[4631]: I1128 13:37:14.274055 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ccdbbd8a-4aeb-4294-b047-754cac601f3d-dns-svc\") pod \"dnsmasq-dns-847c4cc679-mjczg\" (UID: \"ccdbbd8a-4aeb-4294-b047-754cac601f3d\") " pod="openstack/dnsmasq-dns-847c4cc679-mjczg" Nov 28 13:37:14 crc kubenswrapper[4631]: I1128 13:37:14.276276 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ccdbbd8a-4aeb-4294-b047-754cac601f3d-ovsdbserver-sb\") pod \"dnsmasq-dns-847c4cc679-mjczg\" (UID: \"ccdbbd8a-4aeb-4294-b047-754cac601f3d\") " pod="openstack/dnsmasq-dns-847c4cc679-mjczg" Nov 28 13:37:14 crc kubenswrapper[4631]: I1128 13:37:14.284118 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/edf3a98f-0cd6-4eb7-8bf9-0320ae2186c7-credential-keys\") pod \"keystone-bootstrap-65dxp\" (UID: \"edf3a98f-0cd6-4eb7-8bf9-0320ae2186c7\") " pod="openstack/keystone-bootstrap-65dxp" Nov 28 13:37:14 crc kubenswrapper[4631]: I1128 13:37:14.285497 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/edf3a98f-0cd6-4eb7-8bf9-0320ae2186c7-fernet-keys\") pod \"keystone-bootstrap-65dxp\" (UID: \"edf3a98f-0cd6-4eb7-8bf9-0320ae2186c7\") " pod="openstack/keystone-bootstrap-65dxp" Nov 28 13:37:14 crc kubenswrapper[4631]: I1128 13:37:14.293675 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/edf3a98f-0cd6-4eb7-8bf9-0320ae2186c7-scripts\") pod \"keystone-bootstrap-65dxp\" (UID: \"edf3a98f-0cd6-4eb7-8bf9-0320ae2186c7\") " pod="openstack/keystone-bootstrap-65dxp" Nov 28 13:37:14 crc kubenswrapper[4631]: I1128 13:37:14.304891 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ccdbbd8a-4aeb-4294-b047-754cac601f3d-config\") pod \"dnsmasq-dns-847c4cc679-mjczg\" (UID: \"ccdbbd8a-4aeb-4294-b047-754cac601f3d\") " pod="openstack/dnsmasq-dns-847c4cc679-mjczg" Nov 28 13:37:14 crc kubenswrapper[4631]: I1128 13:37:14.316806 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/edf3a98f-0cd6-4eb7-8bf9-0320ae2186c7-combined-ca-bundle\") pod \"keystone-bootstrap-65dxp\" (UID: \"edf3a98f-0cd6-4eb7-8bf9-0320ae2186c7\") " pod="openstack/keystone-bootstrap-65dxp" Nov 28 13:37:14 crc kubenswrapper[4631]: I1128 13:37:14.317767 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qbmh5\" (UniqueName: \"kubernetes.io/projected/ccdbbd8a-4aeb-4294-b047-754cac601f3d-kube-api-access-qbmh5\") pod \"dnsmasq-dns-847c4cc679-mjczg\" (UID: \"ccdbbd8a-4aeb-4294-b047-754cac601f3d\") " pod="openstack/dnsmasq-dns-847c4cc679-mjczg" Nov 28 13:37:14 crc kubenswrapper[4631]: I1128 13:37:14.331602 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-847c4cc679-mjczg" Nov 28 13:37:14 crc kubenswrapper[4631]: I1128 13:37:14.333460 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/edf3a98f-0cd6-4eb7-8bf9-0320ae2186c7-config-data\") pod \"keystone-bootstrap-65dxp\" (UID: \"edf3a98f-0cd6-4eb7-8bf9-0320ae2186c7\") " pod="openstack/keystone-bootstrap-65dxp" Nov 28 13:37:14 crc kubenswrapper[4631]: I1128 13:37:14.359464 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dfmp6\" (UniqueName: \"kubernetes.io/projected/edf3a98f-0cd6-4eb7-8bf9-0320ae2186c7-kube-api-access-dfmp6\") pod \"keystone-bootstrap-65dxp\" (UID: \"edf3a98f-0cd6-4eb7-8bf9-0320ae2186c7\") " pod="openstack/keystone-bootstrap-65dxp" Nov 28 13:37:14 crc kubenswrapper[4631]: I1128 13:37:14.406235 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-gprj7"] Nov 28 13:37:14 crc kubenswrapper[4631]: I1128 13:37:14.408037 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-gprj7" Nov 28 13:37:14 crc kubenswrapper[4631]: I1128 13:37:14.414390 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Nov 28 13:37:14 crc kubenswrapper[4631]: I1128 13:37:14.418650 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Nov 28 13:37:14 crc kubenswrapper[4631]: I1128 13:37:14.419413 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-t2s6l" Nov 28 13:37:14 crc kubenswrapper[4631]: I1128 13:37:14.454394 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-6d58995c-nzh9k"] Nov 28 13:37:14 crc kubenswrapper[4631]: I1128 13:37:14.455943 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6d58995c-nzh9k" Nov 28 13:37:14 crc kubenswrapper[4631]: I1128 13:37:14.463430 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-config-data" Nov 28 13:37:14 crc kubenswrapper[4631]: I1128 13:37:14.463967 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon-horizon-dockercfg-pvkbc" Nov 28 13:37:14 crc kubenswrapper[4631]: I1128 13:37:14.481790 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-gprj7"] Nov 28 13:37:14 crc kubenswrapper[4631]: I1128 13:37:14.486548 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-scripts" Nov 28 13:37:14 crc kubenswrapper[4631]: I1128 13:37:14.486831 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon" Nov 28 13:37:14 crc kubenswrapper[4631]: I1128 13:37:14.584301 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/af357d67-237c-468a-81ac-990964e8e8b2-db-sync-config-data\") pod \"cinder-db-sync-gprj7\" (UID: \"af357d67-237c-468a-81ac-990964e8e8b2\") " pod="openstack/cinder-db-sync-gprj7" Nov 28 13:37:14 crc kubenswrapper[4631]: I1128 13:37:14.584371 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/af357d67-237c-468a-81ac-990964e8e8b2-scripts\") pod \"cinder-db-sync-gprj7\" (UID: \"af357d67-237c-468a-81ac-990964e8e8b2\") " pod="openstack/cinder-db-sync-gprj7" Nov 28 13:37:14 crc kubenswrapper[4631]: I1128 13:37:14.584433 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/af357d67-237c-468a-81ac-990964e8e8b2-combined-ca-bundle\") pod \"cinder-db-sync-gprj7\" (UID: \"af357d67-237c-468a-81ac-990964e8e8b2\") " pod="openstack/cinder-db-sync-gprj7" Nov 28 13:37:14 crc kubenswrapper[4631]: I1128 13:37:14.584458 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/af357d67-237c-468a-81ac-990964e8e8b2-config-data\") pod \"cinder-db-sync-gprj7\" (UID: \"af357d67-237c-468a-81ac-990964e8e8b2\") " pod="openstack/cinder-db-sync-gprj7" Nov 28 13:37:14 crc kubenswrapper[4631]: I1128 13:37:14.584500 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c3bcfdda-21af-4504-806c-50e6bc0f8770-scripts\") pod \"horizon-6d58995c-nzh9k\" (UID: \"c3bcfdda-21af-4504-806c-50e6bc0f8770\") " pod="openstack/horizon-6d58995c-nzh9k" Nov 28 13:37:14 crc kubenswrapper[4631]: I1128 13:37:14.584524 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c3bcfdda-21af-4504-806c-50e6bc0f8770-config-data\") pod \"horizon-6d58995c-nzh9k\" (UID: \"c3bcfdda-21af-4504-806c-50e6bc0f8770\") " pod="openstack/horizon-6d58995c-nzh9k" Nov 28 13:37:14 crc kubenswrapper[4631]: I1128 13:37:14.584549 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/af357d67-237c-468a-81ac-990964e8e8b2-etc-machine-id\") pod \"cinder-db-sync-gprj7\" (UID: \"af357d67-237c-468a-81ac-990964e8e8b2\") " pod="openstack/cinder-db-sync-gprj7" Nov 28 13:37:14 crc kubenswrapper[4631]: I1128 13:37:14.584582 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c3bcfdda-21af-4504-806c-50e6bc0f8770-logs\") pod \"horizon-6d58995c-nzh9k\" (UID: \"c3bcfdda-21af-4504-806c-50e6bc0f8770\") " pod="openstack/horizon-6d58995c-nzh9k" Nov 28 13:37:14 crc kubenswrapper[4631]: I1128 13:37:14.584606 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lrt64\" (UniqueName: \"kubernetes.io/projected/af357d67-237c-468a-81ac-990964e8e8b2-kube-api-access-lrt64\") pod \"cinder-db-sync-gprj7\" (UID: \"af357d67-237c-468a-81ac-990964e8e8b2\") " pod="openstack/cinder-db-sync-gprj7" Nov 28 13:37:14 crc kubenswrapper[4631]: I1128 13:37:14.584638 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/c3bcfdda-21af-4504-806c-50e6bc0f8770-horizon-secret-key\") pod \"horizon-6d58995c-nzh9k\" (UID: \"c3bcfdda-21af-4504-806c-50e6bc0f8770\") " pod="openstack/horizon-6d58995c-nzh9k" Nov 28 13:37:14 crc kubenswrapper[4631]: I1128 13:37:14.584664 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ndw5r\" (UniqueName: \"kubernetes.io/projected/c3bcfdda-21af-4504-806c-50e6bc0f8770-kube-api-access-ndw5r\") pod \"horizon-6d58995c-nzh9k\" (UID: \"c3bcfdda-21af-4504-806c-50e6bc0f8770\") " pod="openstack/horizon-6d58995c-nzh9k" Nov 28 13:37:14 crc kubenswrapper[4631]: I1128 13:37:14.619587 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-6lwcj"] Nov 28 13:37:14 crc kubenswrapper[4631]: I1128 13:37:14.620741 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-6lwcj" Nov 28 13:37:14 crc kubenswrapper[4631]: I1128 13:37:14.634827 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Nov 28 13:37:14 crc kubenswrapper[4631]: I1128 13:37:14.635470 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Nov 28 13:37:14 crc kubenswrapper[4631]: I1128 13:37:14.635726 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-cfjnw" Nov 28 13:37:14 crc kubenswrapper[4631]: I1128 13:37:14.657766 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-65dxp" Nov 28 13:37:14 crc kubenswrapper[4631]: I1128 13:37:14.691910 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/af357d67-237c-468a-81ac-990964e8e8b2-config-data\") pod \"cinder-db-sync-gprj7\" (UID: \"af357d67-237c-468a-81ac-990964e8e8b2\") " pod="openstack/cinder-db-sync-gprj7" Nov 28 13:37:14 crc kubenswrapper[4631]: I1128 13:37:14.691977 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c3bcfdda-21af-4504-806c-50e6bc0f8770-scripts\") pod \"horizon-6d58995c-nzh9k\" (UID: \"c3bcfdda-21af-4504-806c-50e6bc0f8770\") " pod="openstack/horizon-6d58995c-nzh9k" Nov 28 13:37:14 crc kubenswrapper[4631]: I1128 13:37:14.692004 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c3bcfdda-21af-4504-806c-50e6bc0f8770-config-data\") pod \"horizon-6d58995c-nzh9k\" (UID: \"c3bcfdda-21af-4504-806c-50e6bc0f8770\") " pod="openstack/horizon-6d58995c-nzh9k" Nov 28 13:37:14 crc kubenswrapper[4631]: I1128 13:37:14.692033 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/af357d67-237c-468a-81ac-990964e8e8b2-etc-machine-id\") pod \"cinder-db-sync-gprj7\" (UID: \"af357d67-237c-468a-81ac-990964e8e8b2\") " pod="openstack/cinder-db-sync-gprj7" Nov 28 13:37:14 crc kubenswrapper[4631]: I1128 13:37:14.692059 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c3bcfdda-21af-4504-806c-50e6bc0f8770-logs\") pod \"horizon-6d58995c-nzh9k\" (UID: \"c3bcfdda-21af-4504-806c-50e6bc0f8770\") " pod="openstack/horizon-6d58995c-nzh9k" Nov 28 13:37:14 crc kubenswrapper[4631]: I1128 13:37:14.692079 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lrt64\" (UniqueName: \"kubernetes.io/projected/af357d67-237c-468a-81ac-990964e8e8b2-kube-api-access-lrt64\") pod \"cinder-db-sync-gprj7\" (UID: \"af357d67-237c-468a-81ac-990964e8e8b2\") " pod="openstack/cinder-db-sync-gprj7" Nov 28 13:37:14 crc kubenswrapper[4631]: I1128 13:37:14.692116 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/c3bcfdda-21af-4504-806c-50e6bc0f8770-horizon-secret-key\") pod \"horizon-6d58995c-nzh9k\" (UID: \"c3bcfdda-21af-4504-806c-50e6bc0f8770\") " pod="openstack/horizon-6d58995c-nzh9k" Nov 28 13:37:14 crc kubenswrapper[4631]: I1128 13:37:14.692141 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ndw5r\" (UniqueName: \"kubernetes.io/projected/c3bcfdda-21af-4504-806c-50e6bc0f8770-kube-api-access-ndw5r\") pod \"horizon-6d58995c-nzh9k\" (UID: \"c3bcfdda-21af-4504-806c-50e6bc0f8770\") " pod="openstack/horizon-6d58995c-nzh9k" Nov 28 13:37:14 crc kubenswrapper[4631]: I1128 13:37:14.692186 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/af357d67-237c-468a-81ac-990964e8e8b2-db-sync-config-data\") pod \"cinder-db-sync-gprj7\" (UID: \"af357d67-237c-468a-81ac-990964e8e8b2\") " pod="openstack/cinder-db-sync-gprj7" Nov 28 13:37:14 crc kubenswrapper[4631]: I1128 13:37:14.692216 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/af357d67-237c-468a-81ac-990964e8e8b2-scripts\") pod \"cinder-db-sync-gprj7\" (UID: \"af357d67-237c-468a-81ac-990964e8e8b2\") " pod="openstack/cinder-db-sync-gprj7" Nov 28 13:37:14 crc kubenswrapper[4631]: I1128 13:37:14.692243 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/af357d67-237c-468a-81ac-990964e8e8b2-combined-ca-bundle\") pod \"cinder-db-sync-gprj7\" (UID: \"af357d67-237c-468a-81ac-990964e8e8b2\") " pod="openstack/cinder-db-sync-gprj7" Nov 28 13:37:14 crc kubenswrapper[4631]: I1128 13:37:14.712873 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c3bcfdda-21af-4504-806c-50e6bc0f8770-config-data\") pod \"horizon-6d58995c-nzh9k\" (UID: \"c3bcfdda-21af-4504-806c-50e6bc0f8770\") " pod="openstack/horizon-6d58995c-nzh9k" Nov 28 13:37:14 crc kubenswrapper[4631]: I1128 13:37:14.713672 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c3bcfdda-21af-4504-806c-50e6bc0f8770-scripts\") pod \"horizon-6d58995c-nzh9k\" (UID: \"c3bcfdda-21af-4504-806c-50e6bc0f8770\") " pod="openstack/horizon-6d58995c-nzh9k" Nov 28 13:37:14 crc kubenswrapper[4631]: I1128 13:37:14.719544 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/af357d67-237c-468a-81ac-990964e8e8b2-etc-machine-id\") pod \"cinder-db-sync-gprj7\" (UID: \"af357d67-237c-468a-81ac-990964e8e8b2\") " pod="openstack/cinder-db-sync-gprj7" Nov 28 13:37:14 crc kubenswrapper[4631]: I1128 13:37:14.727677 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c3bcfdda-21af-4504-806c-50e6bc0f8770-logs\") pod \"horizon-6d58995c-nzh9k\" (UID: \"c3bcfdda-21af-4504-806c-50e6bc0f8770\") " pod="openstack/horizon-6d58995c-nzh9k" Nov 28 13:37:14 crc kubenswrapper[4631]: I1128 13:37:14.742668 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/af357d67-237c-468a-81ac-990964e8e8b2-db-sync-config-data\") pod \"cinder-db-sync-gprj7\" (UID: \"af357d67-237c-468a-81ac-990964e8e8b2\") " pod="openstack/cinder-db-sync-gprj7" Nov 28 13:37:14 crc kubenswrapper[4631]: I1128 13:37:14.743248 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/af357d67-237c-468a-81ac-990964e8e8b2-scripts\") pod \"cinder-db-sync-gprj7\" (UID: \"af357d67-237c-468a-81ac-990964e8e8b2\") " pod="openstack/cinder-db-sync-gprj7" Nov 28 13:37:14 crc kubenswrapper[4631]: I1128 13:37:14.743792 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/af357d67-237c-468a-81ac-990964e8e8b2-combined-ca-bundle\") pod \"cinder-db-sync-gprj7\" (UID: \"af357d67-237c-468a-81ac-990964e8e8b2\") " pod="openstack/cinder-db-sync-gprj7" Nov 28 13:37:14 crc kubenswrapper[4631]: I1128 13:37:14.744032 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/af357d67-237c-468a-81ac-990964e8e8b2-config-data\") pod \"cinder-db-sync-gprj7\" (UID: \"af357d67-237c-468a-81ac-990964e8e8b2\") " pod="openstack/cinder-db-sync-gprj7" Nov 28 13:37:14 crc kubenswrapper[4631]: I1128 13:37:14.768408 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-6lwcj"] Nov 28 13:37:14 crc kubenswrapper[4631]: I1128 13:37:14.818689 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/c3bcfdda-21af-4504-806c-50e6bc0f8770-horizon-secret-key\") pod \"horizon-6d58995c-nzh9k\" (UID: \"c3bcfdda-21af-4504-806c-50e6bc0f8770\") " pod="openstack/horizon-6d58995c-nzh9k" Nov 28 13:37:14 crc kubenswrapper[4631]: I1128 13:37:14.841905 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/268feb7e-4677-4c47-b837-e04bab4b8010-combined-ca-bundle\") pod \"neutron-db-sync-6lwcj\" (UID: \"268feb7e-4677-4c47-b837-e04bab4b8010\") " pod="openstack/neutron-db-sync-6lwcj" Nov 28 13:37:14 crc kubenswrapper[4631]: I1128 13:37:14.842434 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/268feb7e-4677-4c47-b837-e04bab4b8010-config\") pod \"neutron-db-sync-6lwcj\" (UID: \"268feb7e-4677-4c47-b837-e04bab4b8010\") " pod="openstack/neutron-db-sync-6lwcj" Nov 28 13:37:14 crc kubenswrapper[4631]: I1128 13:37:14.842530 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-758jr\" (UniqueName: \"kubernetes.io/projected/268feb7e-4677-4c47-b837-e04bab4b8010-kube-api-access-758jr\") pod \"neutron-db-sync-6lwcj\" (UID: \"268feb7e-4677-4c47-b837-e04bab4b8010\") " pod="openstack/neutron-db-sync-6lwcj" Nov 28 13:37:14 crc kubenswrapper[4631]: I1128 13:37:14.843546 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-6d58995c-nzh9k"] Nov 28 13:37:14 crc kubenswrapper[4631]: I1128 13:37:14.844023 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ndw5r\" (UniqueName: \"kubernetes.io/projected/c3bcfdda-21af-4504-806c-50e6bc0f8770-kube-api-access-ndw5r\") pod \"horizon-6d58995c-nzh9k\" (UID: \"c3bcfdda-21af-4504-806c-50e6bc0f8770\") " pod="openstack/horizon-6d58995c-nzh9k" Nov 28 13:37:14 crc kubenswrapper[4631]: I1128 13:37:14.863571 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6d58995c-nzh9k" Nov 28 13:37:14 crc kubenswrapper[4631]: I1128 13:37:14.948470 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/268feb7e-4677-4c47-b837-e04bab4b8010-combined-ca-bundle\") pod \"neutron-db-sync-6lwcj\" (UID: \"268feb7e-4677-4c47-b837-e04bab4b8010\") " pod="openstack/neutron-db-sync-6lwcj" Nov 28 13:37:14 crc kubenswrapper[4631]: I1128 13:37:14.948626 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/268feb7e-4677-4c47-b837-e04bab4b8010-config\") pod \"neutron-db-sync-6lwcj\" (UID: \"268feb7e-4677-4c47-b837-e04bab4b8010\") " pod="openstack/neutron-db-sync-6lwcj" Nov 28 13:37:14 crc kubenswrapper[4631]: I1128 13:37:14.948669 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-758jr\" (UniqueName: \"kubernetes.io/projected/268feb7e-4677-4c47-b837-e04bab4b8010-kube-api-access-758jr\") pod \"neutron-db-sync-6lwcj\" (UID: \"268feb7e-4677-4c47-b837-e04bab4b8010\") " pod="openstack/neutron-db-sync-6lwcj" Nov 28 13:37:14 crc kubenswrapper[4631]: I1128 13:37:14.957384 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/268feb7e-4677-4c47-b837-e04bab4b8010-config\") pod \"neutron-db-sync-6lwcj\" (UID: \"268feb7e-4677-4c47-b837-e04bab4b8010\") " pod="openstack/neutron-db-sync-6lwcj" Nov 28 13:37:14 crc kubenswrapper[4631]: I1128 13:37:14.969381 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lrt64\" (UniqueName: \"kubernetes.io/projected/af357d67-237c-468a-81ac-990964e8e8b2-kube-api-access-lrt64\") pod \"cinder-db-sync-gprj7\" (UID: \"af357d67-237c-468a-81ac-990964e8e8b2\") " pod="openstack/cinder-db-sync-gprj7" Nov 28 13:37:14 crc kubenswrapper[4631]: I1128 13:37:14.975704 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/268feb7e-4677-4c47-b837-e04bab4b8010-combined-ca-bundle\") pod \"neutron-db-sync-6lwcj\" (UID: \"268feb7e-4677-4c47-b837-e04bab4b8010\") " pod="openstack/neutron-db-sync-6lwcj" Nov 28 13:37:15 crc kubenswrapper[4631]: I1128 13:37:15.044935 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-758jr\" (UniqueName: \"kubernetes.io/projected/268feb7e-4677-4c47-b837-e04bab4b8010-kube-api-access-758jr\") pod \"neutron-db-sync-6lwcj\" (UID: \"268feb7e-4677-4c47-b837-e04bab4b8010\") " pod="openstack/neutron-db-sync-6lwcj" Nov 28 13:37:15 crc kubenswrapper[4631]: I1128 13:37:15.058456 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 28 13:37:15 crc kubenswrapper[4631]: I1128 13:37:15.062566 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 28 13:37:15 crc kubenswrapper[4631]: I1128 13:37:15.107115 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 28 13:37:15 crc kubenswrapper[4631]: I1128 13:37:15.107412 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 28 13:37:15 crc kubenswrapper[4631]: I1128 13:37:15.114554 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 28 13:37:15 crc kubenswrapper[4631]: I1128 13:37:15.123088 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-gprj7" Nov 28 13:37:15 crc kubenswrapper[4631]: I1128 13:37:15.151407 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-z6tgb"] Nov 28 13:37:15 crc kubenswrapper[4631]: I1128 13:37:15.152806 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-z6tgb" Nov 28 13:37:15 crc kubenswrapper[4631]: I1128 13:37:15.163958 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Nov 28 13:37:15 crc kubenswrapper[4631]: I1128 13:37:15.164141 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-mxvzt" Nov 28 13:37:15 crc kubenswrapper[4631]: I1128 13:37:15.165654 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d95d3536-1392-4576-8ee5-390c09fe7a49-run-httpd\") pod \"ceilometer-0\" (UID: \"d95d3536-1392-4576-8ee5-390c09fe7a49\") " pod="openstack/ceilometer-0" Nov 28 13:37:15 crc kubenswrapper[4631]: I1128 13:37:15.165719 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d95d3536-1392-4576-8ee5-390c09fe7a49-log-httpd\") pod \"ceilometer-0\" (UID: \"d95d3536-1392-4576-8ee5-390c09fe7a49\") " pod="openstack/ceilometer-0" Nov 28 13:37:15 crc kubenswrapper[4631]: I1128 13:37:15.165750 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d95d3536-1392-4576-8ee5-390c09fe7a49-config-data\") pod \"ceilometer-0\" (UID: \"d95d3536-1392-4576-8ee5-390c09fe7a49\") " pod="openstack/ceilometer-0" Nov 28 13:37:15 crc kubenswrapper[4631]: I1128 13:37:15.165791 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d95d3536-1392-4576-8ee5-390c09fe7a49-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d95d3536-1392-4576-8ee5-390c09fe7a49\") " pod="openstack/ceilometer-0" Nov 28 13:37:15 crc kubenswrapper[4631]: I1128 13:37:15.165870 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d95d3536-1392-4576-8ee5-390c09fe7a49-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d95d3536-1392-4576-8ee5-390c09fe7a49\") " pod="openstack/ceilometer-0" Nov 28 13:37:15 crc kubenswrapper[4631]: I1128 13:37:15.165914 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d95d3536-1392-4576-8ee5-390c09fe7a49-scripts\") pod \"ceilometer-0\" (UID: \"d95d3536-1392-4576-8ee5-390c09fe7a49\") " pod="openstack/ceilometer-0" Nov 28 13:37:15 crc kubenswrapper[4631]: I1128 13:37:15.166036 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fd5sn\" (UniqueName: \"kubernetes.io/projected/d95d3536-1392-4576-8ee5-390c09fe7a49-kube-api-access-fd5sn\") pod \"ceilometer-0\" (UID: \"d95d3536-1392-4576-8ee5-390c09fe7a49\") " pod="openstack/ceilometer-0" Nov 28 13:37:15 crc kubenswrapper[4631]: I1128 13:37:15.219828 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-z6tgb"] Nov 28 13:37:15 crc kubenswrapper[4631]: I1128 13:37:15.244641 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-86566f7dbc-2zlk9"] Nov 28 13:37:15 crc kubenswrapper[4631]: I1128 13:37:15.246411 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-86566f7dbc-2zlk9" Nov 28 13:37:15 crc kubenswrapper[4631]: I1128 13:37:15.266056 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-86566f7dbc-2zlk9"] Nov 28 13:37:15 crc kubenswrapper[4631]: I1128 13:37:15.277206 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d95d3536-1392-4576-8ee5-390c09fe7a49-config-data\") pod \"ceilometer-0\" (UID: \"d95d3536-1392-4576-8ee5-390c09fe7a49\") " pod="openstack/ceilometer-0" Nov 28 13:37:15 crc kubenswrapper[4631]: I1128 13:37:15.277379 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d95d3536-1392-4576-8ee5-390c09fe7a49-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d95d3536-1392-4576-8ee5-390c09fe7a49\") " pod="openstack/ceilometer-0" Nov 28 13:37:15 crc kubenswrapper[4631]: I1128 13:37:15.277465 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/3050013d-2db4-4801-8626-1fb6398b714e-db-sync-config-data\") pod \"barbican-db-sync-z6tgb\" (UID: \"3050013d-2db4-4801-8626-1fb6398b714e\") " pod="openstack/barbican-db-sync-z6tgb" Nov 28 13:37:15 crc kubenswrapper[4631]: I1128 13:37:15.277641 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qjbrq\" (UniqueName: \"kubernetes.io/projected/3050013d-2db4-4801-8626-1fb6398b714e-kube-api-access-qjbrq\") pod \"barbican-db-sync-z6tgb\" (UID: \"3050013d-2db4-4801-8626-1fb6398b714e\") " pod="openstack/barbican-db-sync-z6tgb" Nov 28 13:37:15 crc kubenswrapper[4631]: I1128 13:37:15.277718 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d95d3536-1392-4576-8ee5-390c09fe7a49-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d95d3536-1392-4576-8ee5-390c09fe7a49\") " pod="openstack/ceilometer-0" Nov 28 13:37:15 crc kubenswrapper[4631]: I1128 13:37:15.277803 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d95d3536-1392-4576-8ee5-390c09fe7a49-scripts\") pod \"ceilometer-0\" (UID: \"d95d3536-1392-4576-8ee5-390c09fe7a49\") " pod="openstack/ceilometer-0" Nov 28 13:37:15 crc kubenswrapper[4631]: I1128 13:37:15.277836 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3050013d-2db4-4801-8626-1fb6398b714e-combined-ca-bundle\") pod \"barbican-db-sync-z6tgb\" (UID: \"3050013d-2db4-4801-8626-1fb6398b714e\") " pod="openstack/barbican-db-sync-z6tgb" Nov 28 13:37:15 crc kubenswrapper[4631]: I1128 13:37:15.277933 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fd5sn\" (UniqueName: \"kubernetes.io/projected/d95d3536-1392-4576-8ee5-390c09fe7a49-kube-api-access-fd5sn\") pod \"ceilometer-0\" (UID: \"d95d3536-1392-4576-8ee5-390c09fe7a49\") " pod="openstack/ceilometer-0" Nov 28 13:37:15 crc kubenswrapper[4631]: I1128 13:37:15.278022 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d95d3536-1392-4576-8ee5-390c09fe7a49-run-httpd\") pod \"ceilometer-0\" (UID: \"d95d3536-1392-4576-8ee5-390c09fe7a49\") " pod="openstack/ceilometer-0" Nov 28 13:37:15 crc kubenswrapper[4631]: I1128 13:37:15.278061 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d95d3536-1392-4576-8ee5-390c09fe7a49-log-httpd\") pod \"ceilometer-0\" (UID: \"d95d3536-1392-4576-8ee5-390c09fe7a49\") " pod="openstack/ceilometer-0" Nov 28 13:37:15 crc kubenswrapper[4631]: I1128 13:37:15.303794 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Nov 28 13:37:15 crc kubenswrapper[4631]: I1128 13:37:15.305876 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 28 13:37:15 crc kubenswrapper[4631]: I1128 13:37:15.315130 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Nov 28 13:37:15 crc kubenswrapper[4631]: I1128 13:37:15.315572 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-f22sn" Nov 28 13:37:15 crc kubenswrapper[4631]: I1128 13:37:15.315764 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Nov 28 13:37:15 crc kubenswrapper[4631]: I1128 13:37:15.315877 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Nov 28 13:37:15 crc kubenswrapper[4631]: I1128 13:37:15.316534 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-6lwcj" Nov 28 13:37:15 crc kubenswrapper[4631]: I1128 13:37:15.319649 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d95d3536-1392-4576-8ee5-390c09fe7a49-log-httpd\") pod \"ceilometer-0\" (UID: \"d95d3536-1392-4576-8ee5-390c09fe7a49\") " pod="openstack/ceilometer-0" Nov 28 13:37:15 crc kubenswrapper[4631]: I1128 13:37:15.323591 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d95d3536-1392-4576-8ee5-390c09fe7a49-run-httpd\") pod \"ceilometer-0\" (UID: \"d95d3536-1392-4576-8ee5-390c09fe7a49\") " pod="openstack/ceilometer-0" Nov 28 13:37:15 crc kubenswrapper[4631]: I1128 13:37:15.332728 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d95d3536-1392-4576-8ee5-390c09fe7a49-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d95d3536-1392-4576-8ee5-390c09fe7a49\") " pod="openstack/ceilometer-0" Nov 28 13:37:15 crc kubenswrapper[4631]: I1128 13:37:15.333437 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d95d3536-1392-4576-8ee5-390c09fe7a49-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d95d3536-1392-4576-8ee5-390c09fe7a49\") " pod="openstack/ceilometer-0" Nov 28 13:37:15 crc kubenswrapper[4631]: I1128 13:37:15.333525 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d95d3536-1392-4576-8ee5-390c09fe7a49-config-data\") pod \"ceilometer-0\" (UID: \"d95d3536-1392-4576-8ee5-390c09fe7a49\") " pod="openstack/ceilometer-0" Nov 28 13:37:15 crc kubenswrapper[4631]: I1128 13:37:15.343640 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fd5sn\" (UniqueName: \"kubernetes.io/projected/d95d3536-1392-4576-8ee5-390c09fe7a49-kube-api-access-fd5sn\") pod \"ceilometer-0\" (UID: \"d95d3536-1392-4576-8ee5-390c09fe7a49\") " pod="openstack/ceilometer-0" Nov 28 13:37:15 crc kubenswrapper[4631]: I1128 13:37:15.345337 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d95d3536-1392-4576-8ee5-390c09fe7a49-scripts\") pod \"ceilometer-0\" (UID: \"d95d3536-1392-4576-8ee5-390c09fe7a49\") " pod="openstack/ceilometer-0" Nov 28 13:37:15 crc kubenswrapper[4631]: I1128 13:37:15.379403 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-847c4cc679-mjczg"] Nov 28 13:37:15 crc kubenswrapper[4631]: I1128 13:37:15.379520 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3050013d-2db4-4801-8626-1fb6398b714e-combined-ca-bundle\") pod \"barbican-db-sync-z6tgb\" (UID: \"3050013d-2db4-4801-8626-1fb6398b714e\") " pod="openstack/barbican-db-sync-z6tgb" Nov 28 13:37:15 crc kubenswrapper[4631]: I1128 13:37:15.379587 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cqbq9\" (UniqueName: \"kubernetes.io/projected/e35e9173-ceb1-4f6a-8a76-2a8c869aa26b-kube-api-access-cqbq9\") pod \"glance-default-external-api-0\" (UID: \"e35e9173-ceb1-4f6a-8a76-2a8c869aa26b\") " pod="openstack/glance-default-external-api-0" Nov 28 13:37:15 crc kubenswrapper[4631]: I1128 13:37:15.379614 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/cec87f6b-b6fd-467c-adb9-f6c6ff099a18-horizon-secret-key\") pod \"horizon-86566f7dbc-2zlk9\" (UID: \"cec87f6b-b6fd-467c-adb9-f6c6ff099a18\") " pod="openstack/horizon-86566f7dbc-2zlk9" Nov 28 13:37:15 crc kubenswrapper[4631]: I1128 13:37:15.379636 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e35e9173-ceb1-4f6a-8a76-2a8c869aa26b-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"e35e9173-ceb1-4f6a-8a76-2a8c869aa26b\") " pod="openstack/glance-default-external-api-0" Nov 28 13:37:15 crc kubenswrapper[4631]: I1128 13:37:15.379656 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e35e9173-ceb1-4f6a-8a76-2a8c869aa26b-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"e35e9173-ceb1-4f6a-8a76-2a8c869aa26b\") " pod="openstack/glance-default-external-api-0" Nov 28 13:37:15 crc kubenswrapper[4631]: I1128 13:37:15.379675 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j9kq8\" (UniqueName: \"kubernetes.io/projected/cec87f6b-b6fd-467c-adb9-f6c6ff099a18-kube-api-access-j9kq8\") pod \"horizon-86566f7dbc-2zlk9\" (UID: \"cec87f6b-b6fd-467c-adb9-f6c6ff099a18\") " pod="openstack/horizon-86566f7dbc-2zlk9" Nov 28 13:37:15 crc kubenswrapper[4631]: I1128 13:37:15.379704 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cec87f6b-b6fd-467c-adb9-f6c6ff099a18-logs\") pod \"horizon-86566f7dbc-2zlk9\" (UID: \"cec87f6b-b6fd-467c-adb9-f6c6ff099a18\") " pod="openstack/horizon-86566f7dbc-2zlk9" Nov 28 13:37:15 crc kubenswrapper[4631]: I1128 13:37:15.379721 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e35e9173-ceb1-4f6a-8a76-2a8c869aa26b-config-data\") pod \"glance-default-external-api-0\" (UID: \"e35e9173-ceb1-4f6a-8a76-2a8c869aa26b\") " pod="openstack/glance-default-external-api-0" Nov 28 13:37:15 crc kubenswrapper[4631]: I1128 13:37:15.379738 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e35e9173-ceb1-4f6a-8a76-2a8c869aa26b-scripts\") pod \"glance-default-external-api-0\" (UID: \"e35e9173-ceb1-4f6a-8a76-2a8c869aa26b\") " pod="openstack/glance-default-external-api-0" Nov 28 13:37:15 crc kubenswrapper[4631]: I1128 13:37:15.379760 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/cec87f6b-b6fd-467c-adb9-f6c6ff099a18-config-data\") pod \"horizon-86566f7dbc-2zlk9\" (UID: \"cec87f6b-b6fd-467c-adb9-f6c6ff099a18\") " pod="openstack/horizon-86566f7dbc-2zlk9" Nov 28 13:37:15 crc kubenswrapper[4631]: I1128 13:37:15.379777 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/cec87f6b-b6fd-467c-adb9-f6c6ff099a18-scripts\") pod \"horizon-86566f7dbc-2zlk9\" (UID: \"cec87f6b-b6fd-467c-adb9-f6c6ff099a18\") " pod="openstack/horizon-86566f7dbc-2zlk9" Nov 28 13:37:15 crc kubenswrapper[4631]: I1128 13:37:15.379809 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e35e9173-ceb1-4f6a-8a76-2a8c869aa26b-logs\") pod \"glance-default-external-api-0\" (UID: \"e35e9173-ceb1-4f6a-8a76-2a8c869aa26b\") " pod="openstack/glance-default-external-api-0" Nov 28 13:37:15 crc kubenswrapper[4631]: I1128 13:37:15.379825 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-external-api-0\" (UID: \"e35e9173-ceb1-4f6a-8a76-2a8c869aa26b\") " pod="openstack/glance-default-external-api-0" Nov 28 13:37:15 crc kubenswrapper[4631]: I1128 13:37:15.379848 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/3050013d-2db4-4801-8626-1fb6398b714e-db-sync-config-data\") pod \"barbican-db-sync-z6tgb\" (UID: \"3050013d-2db4-4801-8626-1fb6398b714e\") " pod="openstack/barbican-db-sync-z6tgb" Nov 28 13:37:15 crc kubenswrapper[4631]: I1128 13:37:15.379876 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e35e9173-ceb1-4f6a-8a76-2a8c869aa26b-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"e35e9173-ceb1-4f6a-8a76-2a8c869aa26b\") " pod="openstack/glance-default-external-api-0" Nov 28 13:37:15 crc kubenswrapper[4631]: I1128 13:37:15.398075 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3050013d-2db4-4801-8626-1fb6398b714e-combined-ca-bundle\") pod \"barbican-db-sync-z6tgb\" (UID: \"3050013d-2db4-4801-8626-1fb6398b714e\") " pod="openstack/barbican-db-sync-z6tgb" Nov 28 13:37:15 crc kubenswrapper[4631]: I1128 13:37:15.407494 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qjbrq\" (UniqueName: \"kubernetes.io/projected/3050013d-2db4-4801-8626-1fb6398b714e-kube-api-access-qjbrq\") pod \"barbican-db-sync-z6tgb\" (UID: \"3050013d-2db4-4801-8626-1fb6398b714e\") " pod="openstack/barbican-db-sync-z6tgb" Nov 28 13:37:15 crc kubenswrapper[4631]: I1128 13:37:15.435112 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/3050013d-2db4-4801-8626-1fb6398b714e-db-sync-config-data\") pod \"barbican-db-sync-z6tgb\" (UID: \"3050013d-2db4-4801-8626-1fb6398b714e\") " pod="openstack/barbican-db-sync-z6tgb" Nov 28 13:37:15 crc kubenswrapper[4631]: I1128 13:37:15.452679 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 28 13:37:15 crc kubenswrapper[4631]: I1128 13:37:15.455839 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 28 13:37:15 crc kubenswrapper[4631]: I1128 13:37:15.466571 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-7ljwv"] Nov 28 13:37:15 crc kubenswrapper[4631]: I1128 13:37:15.483859 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-785d8bcb8c-7ljwv" Nov 28 13:37:15 crc kubenswrapper[4631]: I1128 13:37:15.497403 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-7ljwv"] Nov 28 13:37:15 crc kubenswrapper[4631]: I1128 13:37:15.507210 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qjbrq\" (UniqueName: \"kubernetes.io/projected/3050013d-2db4-4801-8626-1fb6398b714e-kube-api-access-qjbrq\") pod \"barbican-db-sync-z6tgb\" (UID: \"3050013d-2db4-4801-8626-1fb6398b714e\") " pod="openstack/barbican-db-sync-z6tgb" Nov 28 13:37:15 crc kubenswrapper[4631]: I1128 13:37:15.519886 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e35e9173-ceb1-4f6a-8a76-2a8c869aa26b-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"e35e9173-ceb1-4f6a-8a76-2a8c869aa26b\") " pod="openstack/glance-default-external-api-0" Nov 28 13:37:15 crc kubenswrapper[4631]: I1128 13:37:15.520450 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqbq9\" (UniqueName: \"kubernetes.io/projected/e35e9173-ceb1-4f6a-8a76-2a8c869aa26b-kube-api-access-cqbq9\") pod \"glance-default-external-api-0\" (UID: \"e35e9173-ceb1-4f6a-8a76-2a8c869aa26b\") " pod="openstack/glance-default-external-api-0" Nov 28 13:37:15 crc kubenswrapper[4631]: I1128 13:37:15.520480 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/cec87f6b-b6fd-467c-adb9-f6c6ff099a18-horizon-secret-key\") pod \"horizon-86566f7dbc-2zlk9\" (UID: \"cec87f6b-b6fd-467c-adb9-f6c6ff099a18\") " pod="openstack/horizon-86566f7dbc-2zlk9" Nov 28 13:37:15 crc kubenswrapper[4631]: I1128 13:37:15.520502 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e35e9173-ceb1-4f6a-8a76-2a8c869aa26b-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"e35e9173-ceb1-4f6a-8a76-2a8c869aa26b\") " pod="openstack/glance-default-external-api-0" Nov 28 13:37:15 crc kubenswrapper[4631]: I1128 13:37:15.520520 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e35e9173-ceb1-4f6a-8a76-2a8c869aa26b-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"e35e9173-ceb1-4f6a-8a76-2a8c869aa26b\") " pod="openstack/glance-default-external-api-0" Nov 28 13:37:15 crc kubenswrapper[4631]: I1128 13:37:15.520540 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j9kq8\" (UniqueName: \"kubernetes.io/projected/cec87f6b-b6fd-467c-adb9-f6c6ff099a18-kube-api-access-j9kq8\") pod \"horizon-86566f7dbc-2zlk9\" (UID: \"cec87f6b-b6fd-467c-adb9-f6c6ff099a18\") " pod="openstack/horizon-86566f7dbc-2zlk9" Nov 28 13:37:15 crc kubenswrapper[4631]: I1128 13:37:15.520573 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cec87f6b-b6fd-467c-adb9-f6c6ff099a18-logs\") pod \"horizon-86566f7dbc-2zlk9\" (UID: \"cec87f6b-b6fd-467c-adb9-f6c6ff099a18\") " pod="openstack/horizon-86566f7dbc-2zlk9" Nov 28 13:37:15 crc kubenswrapper[4631]: I1128 13:37:15.520595 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e35e9173-ceb1-4f6a-8a76-2a8c869aa26b-config-data\") pod \"glance-default-external-api-0\" (UID: \"e35e9173-ceb1-4f6a-8a76-2a8c869aa26b\") " pod="openstack/glance-default-external-api-0" Nov 28 13:37:15 crc kubenswrapper[4631]: I1128 13:37:15.520616 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e35e9173-ceb1-4f6a-8a76-2a8c869aa26b-scripts\") pod \"glance-default-external-api-0\" (UID: \"e35e9173-ceb1-4f6a-8a76-2a8c869aa26b\") " pod="openstack/glance-default-external-api-0" Nov 28 13:37:15 crc kubenswrapper[4631]: I1128 13:37:15.520635 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/cec87f6b-b6fd-467c-adb9-f6c6ff099a18-config-data\") pod \"horizon-86566f7dbc-2zlk9\" (UID: \"cec87f6b-b6fd-467c-adb9-f6c6ff099a18\") " pod="openstack/horizon-86566f7dbc-2zlk9" Nov 28 13:37:15 crc kubenswrapper[4631]: I1128 13:37:15.520658 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/cec87f6b-b6fd-467c-adb9-f6c6ff099a18-scripts\") pod \"horizon-86566f7dbc-2zlk9\" (UID: \"cec87f6b-b6fd-467c-adb9-f6c6ff099a18\") " pod="openstack/horizon-86566f7dbc-2zlk9" Nov 28 13:37:15 crc kubenswrapper[4631]: I1128 13:37:15.520690 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e35e9173-ceb1-4f6a-8a76-2a8c869aa26b-logs\") pod \"glance-default-external-api-0\" (UID: \"e35e9173-ceb1-4f6a-8a76-2a8c869aa26b\") " pod="openstack/glance-default-external-api-0" Nov 28 13:37:15 crc kubenswrapper[4631]: I1128 13:37:15.520708 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-external-api-0\" (UID: \"e35e9173-ceb1-4f6a-8a76-2a8c869aa26b\") " pod="openstack/glance-default-external-api-0" Nov 28 13:37:15 crc kubenswrapper[4631]: I1128 13:37:15.521120 4631 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-external-api-0\" (UID: \"e35e9173-ceb1-4f6a-8a76-2a8c869aa26b\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/glance-default-external-api-0" Nov 28 13:37:15 crc kubenswrapper[4631]: I1128 13:37:15.536408 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e35e9173-ceb1-4f6a-8a76-2a8c869aa26b-scripts\") pod \"glance-default-external-api-0\" (UID: \"e35e9173-ceb1-4f6a-8a76-2a8c869aa26b\") " pod="openstack/glance-default-external-api-0" Nov 28 13:37:15 crc kubenswrapper[4631]: I1128 13:37:15.539890 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cec87f6b-b6fd-467c-adb9-f6c6ff099a18-logs\") pod \"horizon-86566f7dbc-2zlk9\" (UID: \"cec87f6b-b6fd-467c-adb9-f6c6ff099a18\") " pod="openstack/horizon-86566f7dbc-2zlk9" Nov 28 13:37:15 crc kubenswrapper[4631]: I1128 13:37:15.542202 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/cec87f6b-b6fd-467c-adb9-f6c6ff099a18-scripts\") pod \"horizon-86566f7dbc-2zlk9\" (UID: \"cec87f6b-b6fd-467c-adb9-f6c6ff099a18\") " pod="openstack/horizon-86566f7dbc-2zlk9" Nov 28 13:37:15 crc kubenswrapper[4631]: I1128 13:37:15.542804 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e35e9173-ceb1-4f6a-8a76-2a8c869aa26b-logs\") pod \"glance-default-external-api-0\" (UID: \"e35e9173-ceb1-4f6a-8a76-2a8c869aa26b\") " pod="openstack/glance-default-external-api-0" Nov 28 13:37:15 crc kubenswrapper[4631]: I1128 13:37:15.543929 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-z6tgb" Nov 28 13:37:15 crc kubenswrapper[4631]: I1128 13:37:15.545092 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e35e9173-ceb1-4f6a-8a76-2a8c869aa26b-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"e35e9173-ceb1-4f6a-8a76-2a8c869aa26b\") " pod="openstack/glance-default-external-api-0" Nov 28 13:37:15 crc kubenswrapper[4631]: I1128 13:37:15.545731 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e35e9173-ceb1-4f6a-8a76-2a8c869aa26b-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"e35e9173-ceb1-4f6a-8a76-2a8c869aa26b\") " pod="openstack/glance-default-external-api-0" Nov 28 13:37:15 crc kubenswrapper[4631]: I1128 13:37:15.550040 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/cec87f6b-b6fd-467c-adb9-f6c6ff099a18-config-data\") pod \"horizon-86566f7dbc-2zlk9\" (UID: \"cec87f6b-b6fd-467c-adb9-f6c6ff099a18\") " pod="openstack/horizon-86566f7dbc-2zlk9" Nov 28 13:37:15 crc kubenswrapper[4631]: I1128 13:37:15.551746 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e35e9173-ceb1-4f6a-8a76-2a8c869aa26b-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"e35e9173-ceb1-4f6a-8a76-2a8c869aa26b\") " pod="openstack/glance-default-external-api-0" Nov 28 13:37:15 crc kubenswrapper[4631]: I1128 13:37:15.552922 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/cec87f6b-b6fd-467c-adb9-f6c6ff099a18-horizon-secret-key\") pod \"horizon-86566f7dbc-2zlk9\" (UID: \"cec87f6b-b6fd-467c-adb9-f6c6ff099a18\") " pod="openstack/horizon-86566f7dbc-2zlk9" Nov 28 13:37:15 crc kubenswrapper[4631]: I1128 13:37:15.554737 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e35e9173-ceb1-4f6a-8a76-2a8c869aa26b-config-data\") pod \"glance-default-external-api-0\" (UID: \"e35e9173-ceb1-4f6a-8a76-2a8c869aa26b\") " pod="openstack/glance-default-external-api-0" Nov 28 13:37:15 crc kubenswrapper[4631]: I1128 13:37:15.591029 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-xhft6"] Nov 28 13:37:15 crc kubenswrapper[4631]: I1128 13:37:15.608232 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqbq9\" (UniqueName: \"kubernetes.io/projected/e35e9173-ceb1-4f6a-8a76-2a8c869aa26b-kube-api-access-cqbq9\") pod \"glance-default-external-api-0\" (UID: \"e35e9173-ceb1-4f6a-8a76-2a8c869aa26b\") " pod="openstack/glance-default-external-api-0" Nov 28 13:37:15 crc kubenswrapper[4631]: I1128 13:37:15.609712 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-xhft6" Nov 28 13:37:15 crc kubenswrapper[4631]: I1128 13:37:15.618362 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-xhft6"] Nov 28 13:37:15 crc kubenswrapper[4631]: I1128 13:37:15.624519 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8d7961e8-77d2-4b92-8758-f06c6d216f73-config\") pod \"dnsmasq-dns-785d8bcb8c-7ljwv\" (UID: \"8d7961e8-77d2-4b92-8758-f06c6d216f73\") " pod="openstack/dnsmasq-dns-785d8bcb8c-7ljwv" Nov 28 13:37:15 crc kubenswrapper[4631]: I1128 13:37:15.634220 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jprwc\" (UniqueName: \"kubernetes.io/projected/8d7961e8-77d2-4b92-8758-f06c6d216f73-kube-api-access-jprwc\") pod \"dnsmasq-dns-785d8bcb8c-7ljwv\" (UID: \"8d7961e8-77d2-4b92-8758-f06c6d216f73\") " pod="openstack/dnsmasq-dns-785d8bcb8c-7ljwv" Nov 28 13:37:15 crc kubenswrapper[4631]: I1128 13:37:15.634409 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8d7961e8-77d2-4b92-8758-f06c6d216f73-dns-svc\") pod \"dnsmasq-dns-785d8bcb8c-7ljwv\" (UID: \"8d7961e8-77d2-4b92-8758-f06c6d216f73\") " pod="openstack/dnsmasq-dns-785d8bcb8c-7ljwv" Nov 28 13:37:15 crc kubenswrapper[4631]: I1128 13:37:15.634537 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8d7961e8-77d2-4b92-8758-f06c6d216f73-dns-swift-storage-0\") pod \"dnsmasq-dns-785d8bcb8c-7ljwv\" (UID: \"8d7961e8-77d2-4b92-8758-f06c6d216f73\") " pod="openstack/dnsmasq-dns-785d8bcb8c-7ljwv" Nov 28 13:37:15 crc kubenswrapper[4631]: I1128 13:37:15.634581 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8d7961e8-77d2-4b92-8758-f06c6d216f73-ovsdbserver-nb\") pod \"dnsmasq-dns-785d8bcb8c-7ljwv\" (UID: \"8d7961e8-77d2-4b92-8758-f06c6d216f73\") " pod="openstack/dnsmasq-dns-785d8bcb8c-7ljwv" Nov 28 13:37:15 crc kubenswrapper[4631]: I1128 13:37:15.634637 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8d7961e8-77d2-4b92-8758-f06c6d216f73-ovsdbserver-sb\") pod \"dnsmasq-dns-785d8bcb8c-7ljwv\" (UID: \"8d7961e8-77d2-4b92-8758-f06c6d216f73\") " pod="openstack/dnsmasq-dns-785d8bcb8c-7ljwv" Nov 28 13:37:15 crc kubenswrapper[4631]: I1128 13:37:15.638115 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Nov 28 13:37:15 crc kubenswrapper[4631]: I1128 13:37:15.653910 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Nov 28 13:37:15 crc kubenswrapper[4631]: I1128 13:37:15.654143 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-d778x" Nov 28 13:37:15 crc kubenswrapper[4631]: I1128 13:37:15.667157 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j9kq8\" (UniqueName: \"kubernetes.io/projected/cec87f6b-b6fd-467c-adb9-f6c6ff099a18-kube-api-access-j9kq8\") pod \"horizon-86566f7dbc-2zlk9\" (UID: \"cec87f6b-b6fd-467c-adb9-f6c6ff099a18\") " pod="openstack/horizon-86566f7dbc-2zlk9" Nov 28 13:37:15 crc kubenswrapper[4631]: I1128 13:37:15.679761 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 28 13:37:15 crc kubenswrapper[4631]: I1128 13:37:15.681510 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 28 13:37:15 crc kubenswrapper[4631]: I1128 13:37:15.692707 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Nov 28 13:37:15 crc kubenswrapper[4631]: I1128 13:37:15.696749 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Nov 28 13:37:15 crc kubenswrapper[4631]: I1128 13:37:15.716428 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-847c4cc679-mjczg" event={"ID":"ccdbbd8a-4aeb-4294-b047-754cac601f3d","Type":"ContainerStarted","Data":"960323b2d4d6df12a10929b4d8e9fb56a1796d1bc70c0c40474ae84fe26ce988"} Nov 28 13:37:15 crc kubenswrapper[4631]: I1128 13:37:15.739657 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/05e73b90-be19-4f14-9824-a5c27361adff-logs\") pod \"placement-db-sync-xhft6\" (UID: \"05e73b90-be19-4f14-9824-a5c27361adff\") " pod="openstack/placement-db-sync-xhft6" Nov 28 13:37:15 crc kubenswrapper[4631]: I1128 13:37:15.739788 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8d7961e8-77d2-4b92-8758-f06c6d216f73-config\") pod \"dnsmasq-dns-785d8bcb8c-7ljwv\" (UID: \"8d7961e8-77d2-4b92-8758-f06c6d216f73\") " pod="openstack/dnsmasq-dns-785d8bcb8c-7ljwv" Nov 28 13:37:15 crc kubenswrapper[4631]: I1128 13:37:15.739851 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jprwc\" (UniqueName: \"kubernetes.io/projected/8d7961e8-77d2-4b92-8758-f06c6d216f73-kube-api-access-jprwc\") pod \"dnsmasq-dns-785d8bcb8c-7ljwv\" (UID: \"8d7961e8-77d2-4b92-8758-f06c6d216f73\") " pod="openstack/dnsmasq-dns-785d8bcb8c-7ljwv" Nov 28 13:37:15 crc kubenswrapper[4631]: I1128 13:37:15.739903 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/05e73b90-be19-4f14-9824-a5c27361adff-config-data\") pod \"placement-db-sync-xhft6\" (UID: \"05e73b90-be19-4f14-9824-a5c27361adff\") " pod="openstack/placement-db-sync-xhft6" Nov 28 13:37:15 crc kubenswrapper[4631]: I1128 13:37:15.739969 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8d7961e8-77d2-4b92-8758-f06c6d216f73-dns-svc\") pod \"dnsmasq-dns-785d8bcb8c-7ljwv\" (UID: \"8d7961e8-77d2-4b92-8758-f06c6d216f73\") " pod="openstack/dnsmasq-dns-785d8bcb8c-7ljwv" Nov 28 13:37:15 crc kubenswrapper[4631]: I1128 13:37:15.739995 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t6gw4\" (UniqueName: \"kubernetes.io/projected/05e73b90-be19-4f14-9824-a5c27361adff-kube-api-access-t6gw4\") pod \"placement-db-sync-xhft6\" (UID: \"05e73b90-be19-4f14-9824-a5c27361adff\") " pod="openstack/placement-db-sync-xhft6" Nov 28 13:37:15 crc kubenswrapper[4631]: I1128 13:37:15.740127 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/05e73b90-be19-4f14-9824-a5c27361adff-combined-ca-bundle\") pod \"placement-db-sync-xhft6\" (UID: \"05e73b90-be19-4f14-9824-a5c27361adff\") " pod="openstack/placement-db-sync-xhft6" Nov 28 13:37:15 crc kubenswrapper[4631]: I1128 13:37:15.740221 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8d7961e8-77d2-4b92-8758-f06c6d216f73-dns-swift-storage-0\") pod \"dnsmasq-dns-785d8bcb8c-7ljwv\" (UID: \"8d7961e8-77d2-4b92-8758-f06c6d216f73\") " pod="openstack/dnsmasq-dns-785d8bcb8c-7ljwv" Nov 28 13:37:15 crc kubenswrapper[4631]: I1128 13:37:15.740395 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8d7961e8-77d2-4b92-8758-f06c6d216f73-ovsdbserver-nb\") pod \"dnsmasq-dns-785d8bcb8c-7ljwv\" (UID: \"8d7961e8-77d2-4b92-8758-f06c6d216f73\") " pod="openstack/dnsmasq-dns-785d8bcb8c-7ljwv" Nov 28 13:37:15 crc kubenswrapper[4631]: I1128 13:37:15.740474 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8d7961e8-77d2-4b92-8758-f06c6d216f73-ovsdbserver-sb\") pod \"dnsmasq-dns-785d8bcb8c-7ljwv\" (UID: \"8d7961e8-77d2-4b92-8758-f06c6d216f73\") " pod="openstack/dnsmasq-dns-785d8bcb8c-7ljwv" Nov 28 13:37:15 crc kubenswrapper[4631]: I1128 13:37:15.740497 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/05e73b90-be19-4f14-9824-a5c27361adff-scripts\") pod \"placement-db-sync-xhft6\" (UID: \"05e73b90-be19-4f14-9824-a5c27361adff\") " pod="openstack/placement-db-sync-xhft6" Nov 28 13:37:15 crc kubenswrapper[4631]: I1128 13:37:15.747050 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8d7961e8-77d2-4b92-8758-f06c6d216f73-dns-swift-storage-0\") pod \"dnsmasq-dns-785d8bcb8c-7ljwv\" (UID: \"8d7961e8-77d2-4b92-8758-f06c6d216f73\") " pod="openstack/dnsmasq-dns-785d8bcb8c-7ljwv" Nov 28 13:37:15 crc kubenswrapper[4631]: I1128 13:37:15.747513 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8d7961e8-77d2-4b92-8758-f06c6d216f73-dns-svc\") pod \"dnsmasq-dns-785d8bcb8c-7ljwv\" (UID: \"8d7961e8-77d2-4b92-8758-f06c6d216f73\") " pod="openstack/dnsmasq-dns-785d8bcb8c-7ljwv" Nov 28 13:37:15 crc kubenswrapper[4631]: I1128 13:37:15.748009 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8d7961e8-77d2-4b92-8758-f06c6d216f73-ovsdbserver-nb\") pod \"dnsmasq-dns-785d8bcb8c-7ljwv\" (UID: \"8d7961e8-77d2-4b92-8758-f06c6d216f73\") " pod="openstack/dnsmasq-dns-785d8bcb8c-7ljwv" Nov 28 13:37:15 crc kubenswrapper[4631]: I1128 13:37:15.750299 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8d7961e8-77d2-4b92-8758-f06c6d216f73-ovsdbserver-sb\") pod \"dnsmasq-dns-785d8bcb8c-7ljwv\" (UID: \"8d7961e8-77d2-4b92-8758-f06c6d216f73\") " pod="openstack/dnsmasq-dns-785d8bcb8c-7ljwv" Nov 28 13:37:15 crc kubenswrapper[4631]: I1128 13:37:15.755721 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8d7961e8-77d2-4b92-8758-f06c6d216f73-config\") pod \"dnsmasq-dns-785d8bcb8c-7ljwv\" (UID: \"8d7961e8-77d2-4b92-8758-f06c6d216f73\") " pod="openstack/dnsmasq-dns-785d8bcb8c-7ljwv" Nov 28 13:37:15 crc kubenswrapper[4631]: I1128 13:37:15.816536 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 28 13:37:15 crc kubenswrapper[4631]: I1128 13:37:15.842510 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-86566f7dbc-2zlk9" Nov 28 13:37:15 crc kubenswrapper[4631]: I1128 13:37:15.851204 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ed809148-a421-492c-a7c1-6a9e4eba0528-scripts\") pod \"glance-default-internal-api-0\" (UID: \"ed809148-a421-492c-a7c1-6a9e4eba0528\") " pod="openstack/glance-default-internal-api-0" Nov 28 13:37:15 crc kubenswrapper[4631]: I1128 13:37:15.851392 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/05e73b90-be19-4f14-9824-a5c27361adff-logs\") pod \"placement-db-sync-xhft6\" (UID: \"05e73b90-be19-4f14-9824-a5c27361adff\") " pod="openstack/placement-db-sync-xhft6" Nov 28 13:37:15 crc kubenswrapper[4631]: I1128 13:37:15.851484 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ed809148-a421-492c-a7c1-6a9e4eba0528-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"ed809148-a421-492c-a7c1-6a9e4eba0528\") " pod="openstack/glance-default-internal-api-0" Nov 28 13:37:15 crc kubenswrapper[4631]: I1128 13:37:15.852810 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/05e73b90-be19-4f14-9824-a5c27361adff-logs\") pod \"placement-db-sync-xhft6\" (UID: \"05e73b90-be19-4f14-9824-a5c27361adff\") " pod="openstack/placement-db-sync-xhft6" Nov 28 13:37:15 crc kubenswrapper[4631]: I1128 13:37:15.859456 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jprwc\" (UniqueName: \"kubernetes.io/projected/8d7961e8-77d2-4b92-8758-f06c6d216f73-kube-api-access-jprwc\") pod \"dnsmasq-dns-785d8bcb8c-7ljwv\" (UID: \"8d7961e8-77d2-4b92-8758-f06c6d216f73\") " pod="openstack/dnsmasq-dns-785d8bcb8c-7ljwv" Nov 28 13:37:15 crc kubenswrapper[4631]: I1128 13:37:15.862362 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-external-api-0\" (UID: \"e35e9173-ceb1-4f6a-8a76-2a8c869aa26b\") " pod="openstack/glance-default-external-api-0" Nov 28 13:37:15 crc kubenswrapper[4631]: I1128 13:37:15.865881 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ed809148-a421-492c-a7c1-6a9e4eba0528-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"ed809148-a421-492c-a7c1-6a9e4eba0528\") " pod="openstack/glance-default-internal-api-0" Nov 28 13:37:15 crc kubenswrapper[4631]: I1128 13:37:15.888859 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-internal-api-0\" (UID: \"ed809148-a421-492c-a7c1-6a9e4eba0528\") " pod="openstack/glance-default-internal-api-0" Nov 28 13:37:15 crc kubenswrapper[4631]: I1128 13:37:15.888960 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/05e73b90-be19-4f14-9824-a5c27361adff-config-data\") pod \"placement-db-sync-xhft6\" (UID: \"05e73b90-be19-4f14-9824-a5c27361adff\") " pod="openstack/placement-db-sync-xhft6" Nov 28 13:37:15 crc kubenswrapper[4631]: I1128 13:37:15.889059 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ed809148-a421-492c-a7c1-6a9e4eba0528-logs\") pod \"glance-default-internal-api-0\" (UID: \"ed809148-a421-492c-a7c1-6a9e4eba0528\") " pod="openstack/glance-default-internal-api-0" Nov 28 13:37:15 crc kubenswrapper[4631]: I1128 13:37:15.889092 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ed809148-a421-492c-a7c1-6a9e4eba0528-config-data\") pod \"glance-default-internal-api-0\" (UID: \"ed809148-a421-492c-a7c1-6a9e4eba0528\") " pod="openstack/glance-default-internal-api-0" Nov 28 13:37:15 crc kubenswrapper[4631]: I1128 13:37:15.889112 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ed809148-a421-492c-a7c1-6a9e4eba0528-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"ed809148-a421-492c-a7c1-6a9e4eba0528\") " pod="openstack/glance-default-internal-api-0" Nov 28 13:37:15 crc kubenswrapper[4631]: I1128 13:37:15.889140 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t6gw4\" (UniqueName: \"kubernetes.io/projected/05e73b90-be19-4f14-9824-a5c27361adff-kube-api-access-t6gw4\") pod \"placement-db-sync-xhft6\" (UID: \"05e73b90-be19-4f14-9824-a5c27361adff\") " pod="openstack/placement-db-sync-xhft6" Nov 28 13:37:15 crc kubenswrapper[4631]: I1128 13:37:15.889175 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/05e73b90-be19-4f14-9824-a5c27361adff-combined-ca-bundle\") pod \"placement-db-sync-xhft6\" (UID: \"05e73b90-be19-4f14-9824-a5c27361adff\") " pod="openstack/placement-db-sync-xhft6" Nov 28 13:37:15 crc kubenswrapper[4631]: I1128 13:37:15.889194 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6hlqb\" (UniqueName: \"kubernetes.io/projected/ed809148-a421-492c-a7c1-6a9e4eba0528-kube-api-access-6hlqb\") pod \"glance-default-internal-api-0\" (UID: \"ed809148-a421-492c-a7c1-6a9e4eba0528\") " pod="openstack/glance-default-internal-api-0" Nov 28 13:37:15 crc kubenswrapper[4631]: I1128 13:37:15.890968 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-external-api-0\" (UID: \"e35e9173-ceb1-4f6a-8a76-2a8c869aa26b\") " pod="openstack/glance-default-external-api-0" Nov 28 13:37:15 crc kubenswrapper[4631]: I1128 13:37:15.891039 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/05e73b90-be19-4f14-9824-a5c27361adff-scripts\") pod \"placement-db-sync-xhft6\" (UID: \"05e73b90-be19-4f14-9824-a5c27361adff\") " pod="openstack/placement-db-sync-xhft6" Nov 28 13:37:15 crc kubenswrapper[4631]: I1128 13:37:15.907479 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-external-api-0\" (UID: \"e35e9173-ceb1-4f6a-8a76-2a8c869aa26b\") " pod="openstack/glance-default-external-api-0" Nov 28 13:37:15 crc kubenswrapper[4631]: I1128 13:37:15.919836 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 28 13:37:15 crc kubenswrapper[4631]: I1128 13:37:15.924397 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/05e73b90-be19-4f14-9824-a5c27361adff-config-data\") pod \"placement-db-sync-xhft6\" (UID: \"05e73b90-be19-4f14-9824-a5c27361adff\") " pod="openstack/placement-db-sync-xhft6" Nov 28 13:37:15 crc kubenswrapper[4631]: I1128 13:37:15.938443 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/05e73b90-be19-4f14-9824-a5c27361adff-combined-ca-bundle\") pod \"placement-db-sync-xhft6\" (UID: \"05e73b90-be19-4f14-9824-a5c27361adff\") " pod="openstack/placement-db-sync-xhft6" Nov 28 13:37:15 crc kubenswrapper[4631]: I1128 13:37:15.966355 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-785d8bcb8c-7ljwv" Nov 28 13:37:15 crc kubenswrapper[4631]: I1128 13:37:15.988346 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t6gw4\" (UniqueName: \"kubernetes.io/projected/05e73b90-be19-4f14-9824-a5c27361adff-kube-api-access-t6gw4\") pod \"placement-db-sync-xhft6\" (UID: \"05e73b90-be19-4f14-9824-a5c27361adff\") " pod="openstack/placement-db-sync-xhft6" Nov 28 13:37:15 crc kubenswrapper[4631]: I1128 13:37:15.989754 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/05e73b90-be19-4f14-9824-a5c27361adff-scripts\") pod \"placement-db-sync-xhft6\" (UID: \"05e73b90-be19-4f14-9824-a5c27361adff\") " pod="openstack/placement-db-sync-xhft6" Nov 28 13:37:16 crc kubenswrapper[4631]: I1128 13:37:16.004779 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-xhft6" Nov 28 13:37:16 crc kubenswrapper[4631]: I1128 13:37:16.007675 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ed809148-a421-492c-a7c1-6a9e4eba0528-scripts\") pod \"glance-default-internal-api-0\" (UID: \"ed809148-a421-492c-a7c1-6a9e4eba0528\") " pod="openstack/glance-default-internal-api-0" Nov 28 13:37:16 crc kubenswrapper[4631]: I1128 13:37:16.007772 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ed809148-a421-492c-a7c1-6a9e4eba0528-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"ed809148-a421-492c-a7c1-6a9e4eba0528\") " pod="openstack/glance-default-internal-api-0" Nov 28 13:37:16 crc kubenswrapper[4631]: I1128 13:37:16.007869 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ed809148-a421-492c-a7c1-6a9e4eba0528-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"ed809148-a421-492c-a7c1-6a9e4eba0528\") " pod="openstack/glance-default-internal-api-0" Nov 28 13:37:16 crc kubenswrapper[4631]: I1128 13:37:16.007954 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-internal-api-0\" (UID: \"ed809148-a421-492c-a7c1-6a9e4eba0528\") " pod="openstack/glance-default-internal-api-0" Nov 28 13:37:16 crc kubenswrapper[4631]: I1128 13:37:16.008124 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ed809148-a421-492c-a7c1-6a9e4eba0528-logs\") pod \"glance-default-internal-api-0\" (UID: \"ed809148-a421-492c-a7c1-6a9e4eba0528\") " pod="openstack/glance-default-internal-api-0" Nov 28 13:37:16 crc kubenswrapper[4631]: I1128 13:37:16.008212 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ed809148-a421-492c-a7c1-6a9e4eba0528-config-data\") pod \"glance-default-internal-api-0\" (UID: \"ed809148-a421-492c-a7c1-6a9e4eba0528\") " pod="openstack/glance-default-internal-api-0" Nov 28 13:37:16 crc kubenswrapper[4631]: I1128 13:37:16.008308 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ed809148-a421-492c-a7c1-6a9e4eba0528-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"ed809148-a421-492c-a7c1-6a9e4eba0528\") " pod="openstack/glance-default-internal-api-0" Nov 28 13:37:16 crc kubenswrapper[4631]: I1128 13:37:16.008381 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6hlqb\" (UniqueName: \"kubernetes.io/projected/ed809148-a421-492c-a7c1-6a9e4eba0528-kube-api-access-6hlqb\") pod \"glance-default-internal-api-0\" (UID: \"ed809148-a421-492c-a7c1-6a9e4eba0528\") " pod="openstack/glance-default-internal-api-0" Nov 28 13:37:16 crc kubenswrapper[4631]: I1128 13:37:16.014634 4631 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-internal-api-0\" (UID: \"ed809148-a421-492c-a7c1-6a9e4eba0528\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/glance-default-internal-api-0" Nov 28 13:37:16 crc kubenswrapper[4631]: I1128 13:37:16.025633 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ed809148-a421-492c-a7c1-6a9e4eba0528-scripts\") pod \"glance-default-internal-api-0\" (UID: \"ed809148-a421-492c-a7c1-6a9e4eba0528\") " pod="openstack/glance-default-internal-api-0" Nov 28 13:37:16 crc kubenswrapper[4631]: I1128 13:37:16.025704 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-847c4cc679-mjczg"] Nov 28 13:37:16 crc kubenswrapper[4631]: I1128 13:37:16.026024 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ed809148-a421-492c-a7c1-6a9e4eba0528-logs\") pod \"glance-default-internal-api-0\" (UID: \"ed809148-a421-492c-a7c1-6a9e4eba0528\") " pod="openstack/glance-default-internal-api-0" Nov 28 13:37:16 crc kubenswrapper[4631]: I1128 13:37:16.034385 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ed809148-a421-492c-a7c1-6a9e4eba0528-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"ed809148-a421-492c-a7c1-6a9e4eba0528\") " pod="openstack/glance-default-internal-api-0" Nov 28 13:37:16 crc kubenswrapper[4631]: I1128 13:37:16.040133 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ed809148-a421-492c-a7c1-6a9e4eba0528-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"ed809148-a421-492c-a7c1-6a9e4eba0528\") " pod="openstack/glance-default-internal-api-0" Nov 28 13:37:16 crc kubenswrapper[4631]: I1128 13:37:16.049736 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ed809148-a421-492c-a7c1-6a9e4eba0528-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"ed809148-a421-492c-a7c1-6a9e4eba0528\") " pod="openstack/glance-default-internal-api-0" Nov 28 13:37:16 crc kubenswrapper[4631]: I1128 13:37:16.050252 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ed809148-a421-492c-a7c1-6a9e4eba0528-config-data\") pod \"glance-default-internal-api-0\" (UID: \"ed809148-a421-492c-a7c1-6a9e4eba0528\") " pod="openstack/glance-default-internal-api-0" Nov 28 13:37:16 crc kubenswrapper[4631]: I1128 13:37:16.103728 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6hlqb\" (UniqueName: \"kubernetes.io/projected/ed809148-a421-492c-a7c1-6a9e4eba0528-kube-api-access-6hlqb\") pod \"glance-default-internal-api-0\" (UID: \"ed809148-a421-492c-a7c1-6a9e4eba0528\") " pod="openstack/glance-default-internal-api-0" Nov 28 13:37:16 crc kubenswrapper[4631]: I1128 13:37:16.428997 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-65dxp"] Nov 28 13:37:16 crc kubenswrapper[4631]: I1128 13:37:16.433011 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-internal-api-0\" (UID: \"ed809148-a421-492c-a7c1-6a9e4eba0528\") " pod="openstack/glance-default-internal-api-0" Nov 28 13:37:16 crc kubenswrapper[4631]: I1128 13:37:16.512439 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-gprj7"] Nov 28 13:37:16 crc kubenswrapper[4631]: I1128 13:37:16.556935 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-6d58995c-nzh9k"] Nov 28 13:37:16 crc kubenswrapper[4631]: I1128 13:37:16.671931 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 28 13:37:16 crc kubenswrapper[4631]: I1128 13:37:16.712700 4631 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/openstack-cell1-galera-0" podUID="4bc20a10-95d5-4003-86d9-c6848bc1b188" containerName="galera" probeResult="failure" output="command timed out" Nov 28 13:37:16 crc kubenswrapper[4631]: I1128 13:37:16.769749 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-gprj7" event={"ID":"af357d67-237c-468a-81ac-990964e8e8b2","Type":"ContainerStarted","Data":"7a93010954a947c18f242d29b0ac230e7ee364211ff12c2bc539d2f8e230404d"} Nov 28 13:37:16 crc kubenswrapper[4631]: I1128 13:37:16.819911 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-65dxp" event={"ID":"edf3a98f-0cd6-4eb7-8bf9-0320ae2186c7","Type":"ContainerStarted","Data":"2c727f4e04248822c558942f4dd0a9912409ecbf0ba9d180e6db14b096b0c562"} Nov 28 13:37:16 crc kubenswrapper[4631]: I1128 13:37:16.828781 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6d58995c-nzh9k" event={"ID":"c3bcfdda-21af-4504-806c-50e6bc0f8770","Type":"ContainerStarted","Data":"04724a4cd88466cc6fc93aaeaed95d494fd770437900d2ca5b18383ea381e17a"} Nov 28 13:37:17 crc kubenswrapper[4631]: I1128 13:37:17.113914 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-z6tgb"] Nov 28 13:37:17 crc kubenswrapper[4631]: I1128 13:37:17.141598 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-6lwcj"] Nov 28 13:37:17 crc kubenswrapper[4631]: W1128 13:37:17.161121 4631 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3050013d_2db4_4801_8626_1fb6398b714e.slice/crio-2a49f5b0bb46e154c3c017389154cdcab6938ea4fe61c8d06e0c4be2c2069ad8 WatchSource:0}: Error finding container 2a49f5b0bb46e154c3c017389154cdcab6938ea4fe61c8d06e0c4be2c2069ad8: Status 404 returned error can't find the container with id 2a49f5b0bb46e154c3c017389154cdcab6938ea4fe61c8d06e0c4be2c2069ad8 Nov 28 13:37:17 crc kubenswrapper[4631]: W1128 13:37:17.164700 4631 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod268feb7e_4677_4c47_b837_e04bab4b8010.slice/crio-d05c339a07f8f431386a5d5539436c081d7d6b0bccd245b02b7ef81095085814 WatchSource:0}: Error finding container d05c339a07f8f431386a5d5539436c081d7d6b0bccd245b02b7ef81095085814: Status 404 returned error can't find the container with id d05c339a07f8f431386a5d5539436c081d7d6b0bccd245b02b7ef81095085814 Nov 28 13:37:17 crc kubenswrapper[4631]: I1128 13:37:17.202006 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-86566f7dbc-2zlk9"] Nov 28 13:37:17 crc kubenswrapper[4631]: I1128 13:37:17.263240 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 28 13:37:17 crc kubenswrapper[4631]: I1128 13:37:17.704376 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-7ljwv"] Nov 28 13:37:17 crc kubenswrapper[4631]: I1128 13:37:17.736662 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-xhft6"] Nov 28 13:37:17 crc kubenswrapper[4631]: I1128 13:37:17.893691 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-65dxp" event={"ID":"edf3a98f-0cd6-4eb7-8bf9-0320ae2186c7","Type":"ContainerStarted","Data":"2011893ca0d1cfaff73961c600f7e3a4976a22a496820c12fdceadc1ef74ba32"} Nov 28 13:37:17 crc kubenswrapper[4631]: I1128 13:37:17.893893 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 28 13:37:17 crc kubenswrapper[4631]: I1128 13:37:17.895805 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d95d3536-1392-4576-8ee5-390c09fe7a49","Type":"ContainerStarted","Data":"bfec386172ed3b954e0624e2432af7b8e50aea6ad51d561d1ba0a2aad259aecd"} Nov 28 13:37:17 crc kubenswrapper[4631]: I1128 13:37:17.898815 4631 generic.go:334] "Generic (PLEG): container finished" podID="ccdbbd8a-4aeb-4294-b047-754cac601f3d" containerID="755e45bec7af398835547344369325580a12e452bf9260e2b3b2667caaac2c9a" exitCode=0 Nov 28 13:37:17 crc kubenswrapper[4631]: I1128 13:37:17.899127 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-847c4cc679-mjczg" event={"ID":"ccdbbd8a-4aeb-4294-b047-754cac601f3d","Type":"ContainerDied","Data":"755e45bec7af398835547344369325580a12e452bf9260e2b3b2667caaac2c9a"} Nov 28 13:37:17 crc kubenswrapper[4631]: I1128 13:37:17.901689 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-86566f7dbc-2zlk9" event={"ID":"cec87f6b-b6fd-467c-adb9-f6c6ff099a18","Type":"ContainerStarted","Data":"54b60687a1b94197f0ec355b26757392cc2c90035dff06655d90153ac6ed7b7a"} Nov 28 13:37:17 crc kubenswrapper[4631]: I1128 13:37:17.904199 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-xhft6" event={"ID":"05e73b90-be19-4f14-9824-a5c27361adff","Type":"ContainerStarted","Data":"6711ceda49aba2a213632d2f369e6f3edeac4568c545e52d706d4124e2e414c0"} Nov 28 13:37:17 crc kubenswrapper[4631]: I1128 13:37:17.905883 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-6lwcj" event={"ID":"268feb7e-4677-4c47-b837-e04bab4b8010","Type":"ContainerStarted","Data":"22d98a24eacf769b7c3570fb24c8bf34f199e9c081e79a87c78dffae0d919cc9"} Nov 28 13:37:17 crc kubenswrapper[4631]: I1128 13:37:17.905903 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-6lwcj" event={"ID":"268feb7e-4677-4c47-b837-e04bab4b8010","Type":"ContainerStarted","Data":"d05c339a07f8f431386a5d5539436c081d7d6b0bccd245b02b7ef81095085814"} Nov 28 13:37:17 crc kubenswrapper[4631]: I1128 13:37:17.917975 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-z6tgb" event={"ID":"3050013d-2db4-4801-8626-1fb6398b714e","Type":"ContainerStarted","Data":"2a49f5b0bb46e154c3c017389154cdcab6938ea4fe61c8d06e0c4be2c2069ad8"} Nov 28 13:37:17 crc kubenswrapper[4631]: I1128 13:37:17.923768 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-65dxp" podStartSLOduration=4.923752962 podStartE2EDuration="4.923752962s" podCreationTimestamp="2025-11-28 13:37:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 13:37:17.920140153 +0000 UTC m=+994.727443507" watchObservedRunningTime="2025-11-28 13:37:17.923752962 +0000 UTC m=+994.731056306" Nov 28 13:37:17 crc kubenswrapper[4631]: I1128 13:37:17.949596 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-7ljwv" event={"ID":"8d7961e8-77d2-4b92-8758-f06c6d216f73","Type":"ContainerStarted","Data":"c06588ceb250932379b10bee3a4c8730bdff42b76db2c5d7d4d8af456e63e93f"} Nov 28 13:37:17 crc kubenswrapper[4631]: I1128 13:37:17.994105 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 28 13:37:18 crc kubenswrapper[4631]: I1128 13:37:18.010113 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-6lwcj" podStartSLOduration=4.010084683 podStartE2EDuration="4.010084683s" podCreationTimestamp="2025-11-28 13:37:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 13:37:17.998441747 +0000 UTC m=+994.805745101" watchObservedRunningTime="2025-11-28 13:37:18.010084683 +0000 UTC m=+994.817388027" Nov 28 13:37:18 crc kubenswrapper[4631]: I1128 13:37:18.409662 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-847c4cc679-mjczg" Nov 28 13:37:18 crc kubenswrapper[4631]: I1128 13:37:18.478259 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ccdbbd8a-4aeb-4294-b047-754cac601f3d-dns-swift-storage-0\") pod \"ccdbbd8a-4aeb-4294-b047-754cac601f3d\" (UID: \"ccdbbd8a-4aeb-4294-b047-754cac601f3d\") " Nov 28 13:37:18 crc kubenswrapper[4631]: I1128 13:37:18.478481 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qbmh5\" (UniqueName: \"kubernetes.io/projected/ccdbbd8a-4aeb-4294-b047-754cac601f3d-kube-api-access-qbmh5\") pod \"ccdbbd8a-4aeb-4294-b047-754cac601f3d\" (UID: \"ccdbbd8a-4aeb-4294-b047-754cac601f3d\") " Nov 28 13:37:18 crc kubenswrapper[4631]: I1128 13:37:18.478556 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ccdbbd8a-4aeb-4294-b047-754cac601f3d-ovsdbserver-nb\") pod \"ccdbbd8a-4aeb-4294-b047-754cac601f3d\" (UID: \"ccdbbd8a-4aeb-4294-b047-754cac601f3d\") " Nov 28 13:37:18 crc kubenswrapper[4631]: I1128 13:37:18.478627 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ccdbbd8a-4aeb-4294-b047-754cac601f3d-dns-svc\") pod \"ccdbbd8a-4aeb-4294-b047-754cac601f3d\" (UID: \"ccdbbd8a-4aeb-4294-b047-754cac601f3d\") " Nov 28 13:37:18 crc kubenswrapper[4631]: I1128 13:37:18.478797 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ccdbbd8a-4aeb-4294-b047-754cac601f3d-ovsdbserver-sb\") pod \"ccdbbd8a-4aeb-4294-b047-754cac601f3d\" (UID: \"ccdbbd8a-4aeb-4294-b047-754cac601f3d\") " Nov 28 13:37:18 crc kubenswrapper[4631]: I1128 13:37:18.478855 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ccdbbd8a-4aeb-4294-b047-754cac601f3d-config\") pod \"ccdbbd8a-4aeb-4294-b047-754cac601f3d\" (UID: \"ccdbbd8a-4aeb-4294-b047-754cac601f3d\") " Nov 28 13:37:18 crc kubenswrapper[4631]: I1128 13:37:18.552029 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ccdbbd8a-4aeb-4294-b047-754cac601f3d-kube-api-access-qbmh5" (OuterVolumeSpecName: "kube-api-access-qbmh5") pod "ccdbbd8a-4aeb-4294-b047-754cac601f3d" (UID: "ccdbbd8a-4aeb-4294-b047-754cac601f3d"). InnerVolumeSpecName "kube-api-access-qbmh5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:37:18 crc kubenswrapper[4631]: I1128 13:37:18.582470 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qbmh5\" (UniqueName: \"kubernetes.io/projected/ccdbbd8a-4aeb-4294-b047-754cac601f3d-kube-api-access-qbmh5\") on node \"crc\" DevicePath \"\"" Nov 28 13:37:18 crc kubenswrapper[4631]: I1128 13:37:18.584077 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ccdbbd8a-4aeb-4294-b047-754cac601f3d-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "ccdbbd8a-4aeb-4294-b047-754cac601f3d" (UID: "ccdbbd8a-4aeb-4294-b047-754cac601f3d"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 13:37:18 crc kubenswrapper[4631]: I1128 13:37:18.596689 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ccdbbd8a-4aeb-4294-b047-754cac601f3d-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "ccdbbd8a-4aeb-4294-b047-754cac601f3d" (UID: "ccdbbd8a-4aeb-4294-b047-754cac601f3d"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 13:37:18 crc kubenswrapper[4631]: I1128 13:37:18.661872 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ccdbbd8a-4aeb-4294-b047-754cac601f3d-config" (OuterVolumeSpecName: "config") pod "ccdbbd8a-4aeb-4294-b047-754cac601f3d" (UID: "ccdbbd8a-4aeb-4294-b047-754cac601f3d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 13:37:18 crc kubenswrapper[4631]: I1128 13:37:18.663607 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ccdbbd8a-4aeb-4294-b047-754cac601f3d-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "ccdbbd8a-4aeb-4294-b047-754cac601f3d" (UID: "ccdbbd8a-4aeb-4294-b047-754cac601f3d"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 13:37:18 crc kubenswrapper[4631]: I1128 13:37:18.696899 4631 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ccdbbd8a-4aeb-4294-b047-754cac601f3d-config\") on node \"crc\" DevicePath \"\"" Nov 28 13:37:18 crc kubenswrapper[4631]: I1128 13:37:18.696938 4631 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ccdbbd8a-4aeb-4294-b047-754cac601f3d-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 28 13:37:18 crc kubenswrapper[4631]: I1128 13:37:18.696951 4631 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ccdbbd8a-4aeb-4294-b047-754cac601f3d-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 28 13:37:18 crc kubenswrapper[4631]: I1128 13:37:18.696962 4631 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ccdbbd8a-4aeb-4294-b047-754cac601f3d-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 28 13:37:18 crc kubenswrapper[4631]: I1128 13:37:18.699816 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ccdbbd8a-4aeb-4294-b047-754cac601f3d-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "ccdbbd8a-4aeb-4294-b047-754cac601f3d" (UID: "ccdbbd8a-4aeb-4294-b047-754cac601f3d"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 13:37:18 crc kubenswrapper[4631]: I1128 13:37:18.746625 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 28 13:37:18 crc kubenswrapper[4631]: I1128 13:37:18.777690 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-6d58995c-nzh9k"] Nov 28 13:37:18 crc kubenswrapper[4631]: I1128 13:37:18.803444 4631 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ccdbbd8a-4aeb-4294-b047-754cac601f3d-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 28 13:37:18 crc kubenswrapper[4631]: I1128 13:37:18.848697 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-59495dc969-4lttm"] Nov 28 13:37:18 crc kubenswrapper[4631]: E1128 13:37:18.849165 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ccdbbd8a-4aeb-4294-b047-754cac601f3d" containerName="init" Nov 28 13:37:18 crc kubenswrapper[4631]: I1128 13:37:18.849179 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="ccdbbd8a-4aeb-4294-b047-754cac601f3d" containerName="init" Nov 28 13:37:18 crc kubenswrapper[4631]: I1128 13:37:18.849382 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="ccdbbd8a-4aeb-4294-b047-754cac601f3d" containerName="init" Nov 28 13:37:18 crc kubenswrapper[4631]: I1128 13:37:18.867829 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-59495dc969-4lttm" Nov 28 13:37:18 crc kubenswrapper[4631]: I1128 13:37:18.908431 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7f02c913-7539-481a-abfc-7e19a28529e1-scripts\") pod \"horizon-59495dc969-4lttm\" (UID: \"7f02c913-7539-481a-abfc-7e19a28529e1\") " pod="openstack/horizon-59495dc969-4lttm" Nov 28 13:37:18 crc kubenswrapper[4631]: I1128 13:37:18.908511 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/7f02c913-7539-481a-abfc-7e19a28529e1-horizon-secret-key\") pod \"horizon-59495dc969-4lttm\" (UID: \"7f02c913-7539-481a-abfc-7e19a28529e1\") " pod="openstack/horizon-59495dc969-4lttm" Nov 28 13:37:18 crc kubenswrapper[4631]: I1128 13:37:18.908552 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7f02c913-7539-481a-abfc-7e19a28529e1-logs\") pod \"horizon-59495dc969-4lttm\" (UID: \"7f02c913-7539-481a-abfc-7e19a28529e1\") " pod="openstack/horizon-59495dc969-4lttm" Nov 28 13:37:18 crc kubenswrapper[4631]: I1128 13:37:18.908605 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-44jtf\" (UniqueName: \"kubernetes.io/projected/7f02c913-7539-481a-abfc-7e19a28529e1-kube-api-access-44jtf\") pod \"horizon-59495dc969-4lttm\" (UID: \"7f02c913-7539-481a-abfc-7e19a28529e1\") " pod="openstack/horizon-59495dc969-4lttm" Nov 28 13:37:18 crc kubenswrapper[4631]: I1128 13:37:18.908680 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/7f02c913-7539-481a-abfc-7e19a28529e1-config-data\") pod \"horizon-59495dc969-4lttm\" (UID: \"7f02c913-7539-481a-abfc-7e19a28529e1\") " pod="openstack/horizon-59495dc969-4lttm" Nov 28 13:37:19 crc kubenswrapper[4631]: I1128 13:37:19.007689 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 28 13:37:19 crc kubenswrapper[4631]: I1128 13:37:19.030550 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7f02c913-7539-481a-abfc-7e19a28529e1-logs\") pod \"horizon-59495dc969-4lttm\" (UID: \"7f02c913-7539-481a-abfc-7e19a28529e1\") " pod="openstack/horizon-59495dc969-4lttm" Nov 28 13:37:19 crc kubenswrapper[4631]: I1128 13:37:19.030658 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-44jtf\" (UniqueName: \"kubernetes.io/projected/7f02c913-7539-481a-abfc-7e19a28529e1-kube-api-access-44jtf\") pod \"horizon-59495dc969-4lttm\" (UID: \"7f02c913-7539-481a-abfc-7e19a28529e1\") " pod="openstack/horizon-59495dc969-4lttm" Nov 28 13:37:19 crc kubenswrapper[4631]: I1128 13:37:19.030819 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/7f02c913-7539-481a-abfc-7e19a28529e1-config-data\") pod \"horizon-59495dc969-4lttm\" (UID: \"7f02c913-7539-481a-abfc-7e19a28529e1\") " pod="openstack/horizon-59495dc969-4lttm" Nov 28 13:37:19 crc kubenswrapper[4631]: I1128 13:37:19.030901 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7f02c913-7539-481a-abfc-7e19a28529e1-scripts\") pod \"horizon-59495dc969-4lttm\" (UID: \"7f02c913-7539-481a-abfc-7e19a28529e1\") " pod="openstack/horizon-59495dc969-4lttm" Nov 28 13:37:19 crc kubenswrapper[4631]: I1128 13:37:19.031594 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/7f02c913-7539-481a-abfc-7e19a28529e1-horizon-secret-key\") pod \"horizon-59495dc969-4lttm\" (UID: \"7f02c913-7539-481a-abfc-7e19a28529e1\") " pod="openstack/horizon-59495dc969-4lttm" Nov 28 13:37:19 crc kubenswrapper[4631]: I1128 13:37:19.036588 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7f02c913-7539-481a-abfc-7e19a28529e1-logs\") pod \"horizon-59495dc969-4lttm\" (UID: \"7f02c913-7539-481a-abfc-7e19a28529e1\") " pod="openstack/horizon-59495dc969-4lttm" Nov 28 13:37:19 crc kubenswrapper[4631]: I1128 13:37:19.038319 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/7f02c913-7539-481a-abfc-7e19a28529e1-config-data\") pod \"horizon-59495dc969-4lttm\" (UID: \"7f02c913-7539-481a-abfc-7e19a28529e1\") " pod="openstack/horizon-59495dc969-4lttm" Nov 28 13:37:19 crc kubenswrapper[4631]: I1128 13:37:19.038852 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7f02c913-7539-481a-abfc-7e19a28529e1-scripts\") pod \"horizon-59495dc969-4lttm\" (UID: \"7f02c913-7539-481a-abfc-7e19a28529e1\") " pod="openstack/horizon-59495dc969-4lttm" Nov 28 13:37:19 crc kubenswrapper[4631]: I1128 13:37:19.048567 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-59495dc969-4lttm"] Nov 28 13:37:19 crc kubenswrapper[4631]: I1128 13:37:19.064220 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/7f02c913-7539-481a-abfc-7e19a28529e1-horizon-secret-key\") pod \"horizon-59495dc969-4lttm\" (UID: \"7f02c913-7539-481a-abfc-7e19a28529e1\") " pod="openstack/horizon-59495dc969-4lttm" Nov 28 13:37:19 crc kubenswrapper[4631]: I1128 13:37:19.142252 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 28 13:37:19 crc kubenswrapper[4631]: I1128 13:37:19.144540 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-44jtf\" (UniqueName: \"kubernetes.io/projected/7f02c913-7539-481a-abfc-7e19a28529e1-kube-api-access-44jtf\") pod \"horizon-59495dc969-4lttm\" (UID: \"7f02c913-7539-481a-abfc-7e19a28529e1\") " pod="openstack/horizon-59495dc969-4lttm" Nov 28 13:37:19 crc kubenswrapper[4631]: I1128 13:37:19.150610 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"e35e9173-ceb1-4f6a-8a76-2a8c869aa26b","Type":"ContainerStarted","Data":"8bd92bf88a9c1ecb415236f8494315e9f848f24db7a89819912d8c8e52663575"} Nov 28 13:37:19 crc kubenswrapper[4631]: I1128 13:37:19.176680 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"ed809148-a421-492c-a7c1-6a9e4eba0528","Type":"ContainerStarted","Data":"320c0c7e4cd1d12b6176cf7431fda56ba41ef73386caf009b93629928efe7cf7"} Nov 28 13:37:19 crc kubenswrapper[4631]: I1128 13:37:19.241805 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-59495dc969-4lttm" Nov 28 13:37:19 crc kubenswrapper[4631]: I1128 13:37:19.248807 4631 generic.go:334] "Generic (PLEG): container finished" podID="8d7961e8-77d2-4b92-8758-f06c6d216f73" containerID="42dda41f0135561bf1f4863d37b0bf6d3670e53ab272554ff0dc38f4e2b83ee3" exitCode=0 Nov 28 13:37:19 crc kubenswrapper[4631]: I1128 13:37:19.248974 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-7ljwv" event={"ID":"8d7961e8-77d2-4b92-8758-f06c6d216f73","Type":"ContainerDied","Data":"42dda41f0135561bf1f4863d37b0bf6d3670e53ab272554ff0dc38f4e2b83ee3"} Nov 28 13:37:19 crc kubenswrapper[4631]: I1128 13:37:19.257555 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-847c4cc679-mjczg" event={"ID":"ccdbbd8a-4aeb-4294-b047-754cac601f3d","Type":"ContainerDied","Data":"960323b2d4d6df12a10929b4d8e9fb56a1796d1bc70c0c40474ae84fe26ce988"} Nov 28 13:37:19 crc kubenswrapper[4631]: I1128 13:37:19.257618 4631 scope.go:117] "RemoveContainer" containerID="755e45bec7af398835547344369325580a12e452bf9260e2b3b2667caaac2c9a" Nov 28 13:37:19 crc kubenswrapper[4631]: I1128 13:37:19.257777 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-847c4cc679-mjczg" Nov 28 13:37:19 crc kubenswrapper[4631]: I1128 13:37:19.450532 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-847c4cc679-mjczg"] Nov 28 13:37:19 crc kubenswrapper[4631]: I1128 13:37:19.492633 4631 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-847c4cc679-mjczg"] Nov 28 13:37:19 crc kubenswrapper[4631]: I1128 13:37:19.610373 4631 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ccdbbd8a-4aeb-4294-b047-754cac601f3d" path="/var/lib/kubelet/pods/ccdbbd8a-4aeb-4294-b047-754cac601f3d/volumes" Nov 28 13:37:19 crc kubenswrapper[4631]: I1128 13:37:19.634782 4631 patch_prober.go:28] interesting pod/machine-config-daemon-47sxw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 28 13:37:19 crc kubenswrapper[4631]: I1128 13:37:19.634844 4631 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" podUID="8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 28 13:37:21 crc kubenswrapper[4631]: I1128 13:37:21.060270 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-59495dc969-4lttm"] Nov 28 13:37:21 crc kubenswrapper[4631]: I1128 13:37:21.404802 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"e35e9173-ceb1-4f6a-8a76-2a8c869aa26b","Type":"ContainerStarted","Data":"d9680309b054cf63cc8abc1bb6e6a4bcd13b7ce334a2d1addf7497cb9c68398c"} Nov 28 13:37:21 crc kubenswrapper[4631]: I1128 13:37:21.411562 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"ed809148-a421-492c-a7c1-6a9e4eba0528","Type":"ContainerStarted","Data":"11e888d5e22b844bbb2536deb555c24c7a734fcdc9caae547b186690ef705ed9"} Nov 28 13:37:21 crc kubenswrapper[4631]: I1128 13:37:21.419425 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-59495dc969-4lttm" event={"ID":"7f02c913-7539-481a-abfc-7e19a28529e1","Type":"ContainerStarted","Data":"a466803046ef5f5b14028f4dd4a8e94216a4b7b4638b3d67a4c58eddb54eff68"} Nov 28 13:37:21 crc kubenswrapper[4631]: I1128 13:37:21.452228 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-7ljwv" event={"ID":"8d7961e8-77d2-4b92-8758-f06c6d216f73","Type":"ContainerStarted","Data":"5a249c121250725ea538cb440f5f420c7a93cf29d5adce655cb1291f01ea65f2"} Nov 28 13:37:21 crc kubenswrapper[4631]: I1128 13:37:21.453766 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-785d8bcb8c-7ljwv" Nov 28 13:37:21 crc kubenswrapper[4631]: I1128 13:37:21.487338 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-785d8bcb8c-7ljwv" podStartSLOduration=6.487320089 podStartE2EDuration="6.487320089s" podCreationTimestamp="2025-11-28 13:37:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 13:37:21.48044053 +0000 UTC m=+998.287743874" watchObservedRunningTime="2025-11-28 13:37:21.487320089 +0000 UTC m=+998.294623433" Nov 28 13:37:23 crc kubenswrapper[4631]: I1128 13:37:23.506179 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"ed809148-a421-492c-a7c1-6a9e4eba0528","Type":"ContainerStarted","Data":"9ae3553e198d5e5871f765c7eb232d5eb1cbd2d68f3a0039023bc82993806c9d"} Nov 28 13:37:23 crc kubenswrapper[4631]: I1128 13:37:23.506353 4631 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="ed809148-a421-492c-a7c1-6a9e4eba0528" containerName="glance-log" containerID="cri-o://11e888d5e22b844bbb2536deb555c24c7a734fcdc9caae547b186690ef705ed9" gracePeriod=30 Nov 28 13:37:23 crc kubenswrapper[4631]: I1128 13:37:23.506989 4631 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="ed809148-a421-492c-a7c1-6a9e4eba0528" containerName="glance-httpd" containerID="cri-o://9ae3553e198d5e5871f765c7eb232d5eb1cbd2d68f3a0039023bc82993806c9d" gracePeriod=30 Nov 28 13:37:23 crc kubenswrapper[4631]: I1128 13:37:23.521798 4631 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="e35e9173-ceb1-4f6a-8a76-2a8c869aa26b" containerName="glance-log" containerID="cri-o://d9680309b054cf63cc8abc1bb6e6a4bcd13b7ce334a2d1addf7497cb9c68398c" gracePeriod=30 Nov 28 13:37:23 crc kubenswrapper[4631]: I1128 13:37:23.522268 4631 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="e35e9173-ceb1-4f6a-8a76-2a8c869aa26b" containerName="glance-httpd" containerID="cri-o://9c22148dc08e7d0f7400463e44a7fe46239e46b30d3b118c3cac56d80d27c687" gracePeriod=30 Nov 28 13:37:23 crc kubenswrapper[4631]: I1128 13:37:23.528724 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"e35e9173-ceb1-4f6a-8a76-2a8c869aa26b","Type":"ContainerStarted","Data":"9c22148dc08e7d0f7400463e44a7fe46239e46b30d3b118c3cac56d80d27c687"} Nov 28 13:37:23 crc kubenswrapper[4631]: I1128 13:37:23.713197 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=8.713173925 podStartE2EDuration="8.713173925s" podCreationTimestamp="2025-11-28 13:37:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 13:37:23.589565406 +0000 UTC m=+1000.396868770" watchObservedRunningTime="2025-11-28 13:37:23.713173925 +0000 UTC m=+1000.520477269" Nov 28 13:37:23 crc kubenswrapper[4631]: I1128 13:37:23.908757 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=9.908731889 podStartE2EDuration="9.908731889s" podCreationTimestamp="2025-11-28 13:37:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 13:37:23.898972839 +0000 UTC m=+1000.706276183" watchObservedRunningTime="2025-11-28 13:37:23.908731889 +0000 UTC m=+1000.716035233" Nov 28 13:37:24 crc kubenswrapper[4631]: I1128 13:37:24.590948 4631 generic.go:334] "Generic (PLEG): container finished" podID="ed809148-a421-492c-a7c1-6a9e4eba0528" containerID="9ae3553e198d5e5871f765c7eb232d5eb1cbd2d68f3a0039023bc82993806c9d" exitCode=0 Nov 28 13:37:24 crc kubenswrapper[4631]: I1128 13:37:24.591522 4631 generic.go:334] "Generic (PLEG): container finished" podID="ed809148-a421-492c-a7c1-6a9e4eba0528" containerID="11e888d5e22b844bbb2536deb555c24c7a734fcdc9caae547b186690ef705ed9" exitCode=143 Nov 28 13:37:24 crc kubenswrapper[4631]: I1128 13:37:24.591241 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"ed809148-a421-492c-a7c1-6a9e4eba0528","Type":"ContainerDied","Data":"9ae3553e198d5e5871f765c7eb232d5eb1cbd2d68f3a0039023bc82993806c9d"} Nov 28 13:37:24 crc kubenswrapper[4631]: I1128 13:37:24.591722 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"ed809148-a421-492c-a7c1-6a9e4eba0528","Type":"ContainerDied","Data":"11e888d5e22b844bbb2536deb555c24c7a734fcdc9caae547b186690ef705ed9"} Nov 28 13:37:24 crc kubenswrapper[4631]: I1128 13:37:24.599657 4631 generic.go:334] "Generic (PLEG): container finished" podID="e35e9173-ceb1-4f6a-8a76-2a8c869aa26b" containerID="9c22148dc08e7d0f7400463e44a7fe46239e46b30d3b118c3cac56d80d27c687" exitCode=0 Nov 28 13:37:24 crc kubenswrapper[4631]: I1128 13:37:24.599696 4631 generic.go:334] "Generic (PLEG): container finished" podID="e35e9173-ceb1-4f6a-8a76-2a8c869aa26b" containerID="d9680309b054cf63cc8abc1bb6e6a4bcd13b7ce334a2d1addf7497cb9c68398c" exitCode=143 Nov 28 13:37:24 crc kubenswrapper[4631]: I1128 13:37:24.599792 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"e35e9173-ceb1-4f6a-8a76-2a8c869aa26b","Type":"ContainerDied","Data":"9c22148dc08e7d0f7400463e44a7fe46239e46b30d3b118c3cac56d80d27c687"} Nov 28 13:37:24 crc kubenswrapper[4631]: I1128 13:37:24.599831 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"e35e9173-ceb1-4f6a-8a76-2a8c869aa26b","Type":"ContainerDied","Data":"d9680309b054cf63cc8abc1bb6e6a4bcd13b7ce334a2d1addf7497cb9c68398c"} Nov 28 13:37:24 crc kubenswrapper[4631]: I1128 13:37:24.961315 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-86566f7dbc-2zlk9"] Nov 28 13:37:24 crc kubenswrapper[4631]: I1128 13:37:24.987902 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-64f8fdf97b-jn5mg"] Nov 28 13:37:24 crc kubenswrapper[4631]: I1128 13:37:24.989849 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-64f8fdf97b-jn5mg" Nov 28 13:37:24 crc kubenswrapper[4631]: I1128 13:37:24.993641 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-horizon-svc" Nov 28 13:37:25 crc kubenswrapper[4631]: I1128 13:37:25.028841 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-64f8fdf97b-jn5mg"] Nov 28 13:37:25 crc kubenswrapper[4631]: I1128 13:37:25.160084 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-59495dc969-4lttm"] Nov 28 13:37:25 crc kubenswrapper[4631]: I1128 13:37:25.189801 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/111c5349-528d-4856-a3e4-1277e3b10889-horizon-secret-key\") pod \"horizon-64f8fdf97b-jn5mg\" (UID: \"111c5349-528d-4856-a3e4-1277e3b10889\") " pod="openstack/horizon-64f8fdf97b-jn5mg" Nov 28 13:37:25 crc kubenswrapper[4631]: I1128 13:37:25.189864 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/111c5349-528d-4856-a3e4-1277e3b10889-scripts\") pod \"horizon-64f8fdf97b-jn5mg\" (UID: \"111c5349-528d-4856-a3e4-1277e3b10889\") " pod="openstack/horizon-64f8fdf97b-jn5mg" Nov 28 13:37:25 crc kubenswrapper[4631]: I1128 13:37:25.189893 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/111c5349-528d-4856-a3e4-1277e3b10889-config-data\") pod \"horizon-64f8fdf97b-jn5mg\" (UID: \"111c5349-528d-4856-a3e4-1277e3b10889\") " pod="openstack/horizon-64f8fdf97b-jn5mg" Nov 28 13:37:25 crc kubenswrapper[4631]: I1128 13:37:25.189919 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/111c5349-528d-4856-a3e4-1277e3b10889-logs\") pod \"horizon-64f8fdf97b-jn5mg\" (UID: \"111c5349-528d-4856-a3e4-1277e3b10889\") " pod="openstack/horizon-64f8fdf97b-jn5mg" Nov 28 13:37:25 crc kubenswrapper[4631]: I1128 13:37:25.189944 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k8tns\" (UniqueName: \"kubernetes.io/projected/111c5349-528d-4856-a3e4-1277e3b10889-kube-api-access-k8tns\") pod \"horizon-64f8fdf97b-jn5mg\" (UID: \"111c5349-528d-4856-a3e4-1277e3b10889\") " pod="openstack/horizon-64f8fdf97b-jn5mg" Nov 28 13:37:25 crc kubenswrapper[4631]: I1128 13:37:25.189979 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/111c5349-528d-4856-a3e4-1277e3b10889-combined-ca-bundle\") pod \"horizon-64f8fdf97b-jn5mg\" (UID: \"111c5349-528d-4856-a3e4-1277e3b10889\") " pod="openstack/horizon-64f8fdf97b-jn5mg" Nov 28 13:37:25 crc kubenswrapper[4631]: I1128 13:37:25.190071 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/111c5349-528d-4856-a3e4-1277e3b10889-horizon-tls-certs\") pod \"horizon-64f8fdf97b-jn5mg\" (UID: \"111c5349-528d-4856-a3e4-1277e3b10889\") " pod="openstack/horizon-64f8fdf97b-jn5mg" Nov 28 13:37:25 crc kubenswrapper[4631]: I1128 13:37:25.208725 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-544496c768-rzhsw"] Nov 28 13:37:25 crc kubenswrapper[4631]: I1128 13:37:25.210640 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-544496c768-rzhsw" Nov 28 13:37:25 crc kubenswrapper[4631]: I1128 13:37:25.220162 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-544496c768-rzhsw"] Nov 28 13:37:25 crc kubenswrapper[4631]: I1128 13:37:25.292055 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd607dae-568a-4b81-af81-3310c0e95854-combined-ca-bundle\") pod \"horizon-544496c768-rzhsw\" (UID: \"cd607dae-568a-4b81-af81-3310c0e95854\") " pod="openstack/horizon-544496c768-rzhsw" Nov 28 13:37:25 crc kubenswrapper[4631]: I1128 13:37:25.292115 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/cd607dae-568a-4b81-af81-3310c0e95854-scripts\") pod \"horizon-544496c768-rzhsw\" (UID: \"cd607dae-568a-4b81-af81-3310c0e95854\") " pod="openstack/horizon-544496c768-rzhsw" Nov 28 13:37:25 crc kubenswrapper[4631]: I1128 13:37:25.292148 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/111c5349-528d-4856-a3e4-1277e3b10889-horizon-secret-key\") pod \"horizon-64f8fdf97b-jn5mg\" (UID: \"111c5349-528d-4856-a3e4-1277e3b10889\") " pod="openstack/horizon-64f8fdf97b-jn5mg" Nov 28 13:37:25 crc kubenswrapper[4631]: I1128 13:37:25.292170 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/111c5349-528d-4856-a3e4-1277e3b10889-scripts\") pod \"horizon-64f8fdf97b-jn5mg\" (UID: \"111c5349-528d-4856-a3e4-1277e3b10889\") " pod="openstack/horizon-64f8fdf97b-jn5mg" Nov 28 13:37:25 crc kubenswrapper[4631]: I1128 13:37:25.292192 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/111c5349-528d-4856-a3e4-1277e3b10889-config-data\") pod \"horizon-64f8fdf97b-jn5mg\" (UID: \"111c5349-528d-4856-a3e4-1277e3b10889\") " pod="openstack/horizon-64f8fdf97b-jn5mg" Nov 28 13:37:25 crc kubenswrapper[4631]: I1128 13:37:25.292215 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/111c5349-528d-4856-a3e4-1277e3b10889-logs\") pod \"horizon-64f8fdf97b-jn5mg\" (UID: \"111c5349-528d-4856-a3e4-1277e3b10889\") " pod="openstack/horizon-64f8fdf97b-jn5mg" Nov 28 13:37:25 crc kubenswrapper[4631]: I1128 13:37:25.292236 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k8tns\" (UniqueName: \"kubernetes.io/projected/111c5349-528d-4856-a3e4-1277e3b10889-kube-api-access-k8tns\") pod \"horizon-64f8fdf97b-jn5mg\" (UID: \"111c5349-528d-4856-a3e4-1277e3b10889\") " pod="openstack/horizon-64f8fdf97b-jn5mg" Nov 28 13:37:25 crc kubenswrapper[4631]: I1128 13:37:25.292263 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/111c5349-528d-4856-a3e4-1277e3b10889-combined-ca-bundle\") pod \"horizon-64f8fdf97b-jn5mg\" (UID: \"111c5349-528d-4856-a3e4-1277e3b10889\") " pod="openstack/horizon-64f8fdf97b-jn5mg" Nov 28 13:37:25 crc kubenswrapper[4631]: I1128 13:37:25.292305 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/cd607dae-568a-4b81-af81-3310c0e95854-config-data\") pod \"horizon-544496c768-rzhsw\" (UID: \"cd607dae-568a-4b81-af81-3310c0e95854\") " pod="openstack/horizon-544496c768-rzhsw" Nov 28 13:37:25 crc kubenswrapper[4631]: I1128 13:37:25.292336 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/111c5349-528d-4856-a3e4-1277e3b10889-horizon-tls-certs\") pod \"horizon-64f8fdf97b-jn5mg\" (UID: \"111c5349-528d-4856-a3e4-1277e3b10889\") " pod="openstack/horizon-64f8fdf97b-jn5mg" Nov 28 13:37:25 crc kubenswrapper[4631]: I1128 13:37:25.292415 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/cd607dae-568a-4b81-af81-3310c0e95854-horizon-secret-key\") pod \"horizon-544496c768-rzhsw\" (UID: \"cd607dae-568a-4b81-af81-3310c0e95854\") " pod="openstack/horizon-544496c768-rzhsw" Nov 28 13:37:25 crc kubenswrapper[4631]: I1128 13:37:25.292457 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/cd607dae-568a-4b81-af81-3310c0e95854-horizon-tls-certs\") pod \"horizon-544496c768-rzhsw\" (UID: \"cd607dae-568a-4b81-af81-3310c0e95854\") " pod="openstack/horizon-544496c768-rzhsw" Nov 28 13:37:25 crc kubenswrapper[4631]: I1128 13:37:25.292484 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cd607dae-568a-4b81-af81-3310c0e95854-logs\") pod \"horizon-544496c768-rzhsw\" (UID: \"cd607dae-568a-4b81-af81-3310c0e95854\") " pod="openstack/horizon-544496c768-rzhsw" Nov 28 13:37:25 crc kubenswrapper[4631]: I1128 13:37:25.292510 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kjfzc\" (UniqueName: \"kubernetes.io/projected/cd607dae-568a-4b81-af81-3310c0e95854-kube-api-access-kjfzc\") pod \"horizon-544496c768-rzhsw\" (UID: \"cd607dae-568a-4b81-af81-3310c0e95854\") " pod="openstack/horizon-544496c768-rzhsw" Nov 28 13:37:25 crc kubenswrapper[4631]: I1128 13:37:25.292984 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/111c5349-528d-4856-a3e4-1277e3b10889-logs\") pod \"horizon-64f8fdf97b-jn5mg\" (UID: \"111c5349-528d-4856-a3e4-1277e3b10889\") " pod="openstack/horizon-64f8fdf97b-jn5mg" Nov 28 13:37:25 crc kubenswrapper[4631]: I1128 13:37:25.296063 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/111c5349-528d-4856-a3e4-1277e3b10889-config-data\") pod \"horizon-64f8fdf97b-jn5mg\" (UID: \"111c5349-528d-4856-a3e4-1277e3b10889\") " pod="openstack/horizon-64f8fdf97b-jn5mg" Nov 28 13:37:25 crc kubenswrapper[4631]: I1128 13:37:25.296220 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/111c5349-528d-4856-a3e4-1277e3b10889-scripts\") pod \"horizon-64f8fdf97b-jn5mg\" (UID: \"111c5349-528d-4856-a3e4-1277e3b10889\") " pod="openstack/horizon-64f8fdf97b-jn5mg" Nov 28 13:37:25 crc kubenswrapper[4631]: I1128 13:37:25.302506 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/111c5349-528d-4856-a3e4-1277e3b10889-horizon-secret-key\") pod \"horizon-64f8fdf97b-jn5mg\" (UID: \"111c5349-528d-4856-a3e4-1277e3b10889\") " pod="openstack/horizon-64f8fdf97b-jn5mg" Nov 28 13:37:25 crc kubenswrapper[4631]: I1128 13:37:25.305653 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/111c5349-528d-4856-a3e4-1277e3b10889-combined-ca-bundle\") pod \"horizon-64f8fdf97b-jn5mg\" (UID: \"111c5349-528d-4856-a3e4-1277e3b10889\") " pod="openstack/horizon-64f8fdf97b-jn5mg" Nov 28 13:37:25 crc kubenswrapper[4631]: I1128 13:37:25.305782 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/111c5349-528d-4856-a3e4-1277e3b10889-horizon-tls-certs\") pod \"horizon-64f8fdf97b-jn5mg\" (UID: \"111c5349-528d-4856-a3e4-1277e3b10889\") " pod="openstack/horizon-64f8fdf97b-jn5mg" Nov 28 13:37:25 crc kubenswrapper[4631]: I1128 13:37:25.326819 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k8tns\" (UniqueName: \"kubernetes.io/projected/111c5349-528d-4856-a3e4-1277e3b10889-kube-api-access-k8tns\") pod \"horizon-64f8fdf97b-jn5mg\" (UID: \"111c5349-528d-4856-a3e4-1277e3b10889\") " pod="openstack/horizon-64f8fdf97b-jn5mg" Nov 28 13:37:25 crc kubenswrapper[4631]: I1128 13:37:25.330986 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-64f8fdf97b-jn5mg" Nov 28 13:37:25 crc kubenswrapper[4631]: I1128 13:37:25.396797 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/cd607dae-568a-4b81-af81-3310c0e95854-config-data\") pod \"horizon-544496c768-rzhsw\" (UID: \"cd607dae-568a-4b81-af81-3310c0e95854\") " pod="openstack/horizon-544496c768-rzhsw" Nov 28 13:37:25 crc kubenswrapper[4631]: I1128 13:37:25.398778 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/cd607dae-568a-4b81-af81-3310c0e95854-horizon-secret-key\") pod \"horizon-544496c768-rzhsw\" (UID: \"cd607dae-568a-4b81-af81-3310c0e95854\") " pod="openstack/horizon-544496c768-rzhsw" Nov 28 13:37:25 crc kubenswrapper[4631]: I1128 13:37:25.398899 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/cd607dae-568a-4b81-af81-3310c0e95854-horizon-tls-certs\") pod \"horizon-544496c768-rzhsw\" (UID: \"cd607dae-568a-4b81-af81-3310c0e95854\") " pod="openstack/horizon-544496c768-rzhsw" Nov 28 13:37:25 crc kubenswrapper[4631]: I1128 13:37:25.398939 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cd607dae-568a-4b81-af81-3310c0e95854-logs\") pod \"horizon-544496c768-rzhsw\" (UID: \"cd607dae-568a-4b81-af81-3310c0e95854\") " pod="openstack/horizon-544496c768-rzhsw" Nov 28 13:37:25 crc kubenswrapper[4631]: I1128 13:37:25.398989 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kjfzc\" (UniqueName: \"kubernetes.io/projected/cd607dae-568a-4b81-af81-3310c0e95854-kube-api-access-kjfzc\") pod \"horizon-544496c768-rzhsw\" (UID: \"cd607dae-568a-4b81-af81-3310c0e95854\") " pod="openstack/horizon-544496c768-rzhsw" Nov 28 13:37:25 crc kubenswrapper[4631]: I1128 13:37:25.399060 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd607dae-568a-4b81-af81-3310c0e95854-combined-ca-bundle\") pod \"horizon-544496c768-rzhsw\" (UID: \"cd607dae-568a-4b81-af81-3310c0e95854\") " pod="openstack/horizon-544496c768-rzhsw" Nov 28 13:37:25 crc kubenswrapper[4631]: I1128 13:37:25.399149 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/cd607dae-568a-4b81-af81-3310c0e95854-scripts\") pod \"horizon-544496c768-rzhsw\" (UID: \"cd607dae-568a-4b81-af81-3310c0e95854\") " pod="openstack/horizon-544496c768-rzhsw" Nov 28 13:37:25 crc kubenswrapper[4631]: I1128 13:37:25.399717 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/cd607dae-568a-4b81-af81-3310c0e95854-config-data\") pod \"horizon-544496c768-rzhsw\" (UID: \"cd607dae-568a-4b81-af81-3310c0e95854\") " pod="openstack/horizon-544496c768-rzhsw" Nov 28 13:37:25 crc kubenswrapper[4631]: I1128 13:37:25.400555 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/cd607dae-568a-4b81-af81-3310c0e95854-scripts\") pod \"horizon-544496c768-rzhsw\" (UID: \"cd607dae-568a-4b81-af81-3310c0e95854\") " pod="openstack/horizon-544496c768-rzhsw" Nov 28 13:37:25 crc kubenswrapper[4631]: I1128 13:37:25.401691 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cd607dae-568a-4b81-af81-3310c0e95854-logs\") pod \"horizon-544496c768-rzhsw\" (UID: \"cd607dae-568a-4b81-af81-3310c0e95854\") " pod="openstack/horizon-544496c768-rzhsw" Nov 28 13:37:25 crc kubenswrapper[4631]: I1128 13:37:25.405986 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/cd607dae-568a-4b81-af81-3310c0e95854-horizon-secret-key\") pod \"horizon-544496c768-rzhsw\" (UID: \"cd607dae-568a-4b81-af81-3310c0e95854\") " pod="openstack/horizon-544496c768-rzhsw" Nov 28 13:37:25 crc kubenswrapper[4631]: I1128 13:37:25.406598 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd607dae-568a-4b81-af81-3310c0e95854-combined-ca-bundle\") pod \"horizon-544496c768-rzhsw\" (UID: \"cd607dae-568a-4b81-af81-3310c0e95854\") " pod="openstack/horizon-544496c768-rzhsw" Nov 28 13:37:25 crc kubenswrapper[4631]: I1128 13:37:25.422625 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/cd607dae-568a-4b81-af81-3310c0e95854-horizon-tls-certs\") pod \"horizon-544496c768-rzhsw\" (UID: \"cd607dae-568a-4b81-af81-3310c0e95854\") " pod="openstack/horizon-544496c768-rzhsw" Nov 28 13:37:25 crc kubenswrapper[4631]: I1128 13:37:25.425879 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kjfzc\" (UniqueName: \"kubernetes.io/projected/cd607dae-568a-4b81-af81-3310c0e95854-kube-api-access-kjfzc\") pod \"horizon-544496c768-rzhsw\" (UID: \"cd607dae-568a-4b81-af81-3310c0e95854\") " pod="openstack/horizon-544496c768-rzhsw" Nov 28 13:37:25 crc kubenswrapper[4631]: I1128 13:37:25.556057 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-544496c768-rzhsw" Nov 28 13:37:25 crc kubenswrapper[4631]: I1128 13:37:25.970675 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-785d8bcb8c-7ljwv" Nov 28 13:37:26 crc kubenswrapper[4631]: I1128 13:37:26.044000 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-74f6bcbc87-r8g9l"] Nov 28 13:37:26 crc kubenswrapper[4631]: I1128 13:37:26.044306 4631 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-74f6bcbc87-r8g9l" podUID="166f0f43-e5cb-42ad-a2f5-1b207accc673" containerName="dnsmasq-dns" containerID="cri-o://e9a7a7a5cf57eda39ae2aac9a52f8fe72136a44113973538ac6ec5d2590a54ef" gracePeriod=10 Nov 28 13:37:26 crc kubenswrapper[4631]: I1128 13:37:26.646069 4631 generic.go:334] "Generic (PLEG): container finished" podID="166f0f43-e5cb-42ad-a2f5-1b207accc673" containerID="e9a7a7a5cf57eda39ae2aac9a52f8fe72136a44113973538ac6ec5d2590a54ef" exitCode=0 Nov 28 13:37:26 crc kubenswrapper[4631]: I1128 13:37:26.646512 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74f6bcbc87-r8g9l" event={"ID":"166f0f43-e5cb-42ad-a2f5-1b207accc673","Type":"ContainerDied","Data":"e9a7a7a5cf57eda39ae2aac9a52f8fe72136a44113973538ac6ec5d2590a54ef"} Nov 28 13:37:27 crc kubenswrapper[4631]: I1128 13:37:27.668033 4631 generic.go:334] "Generic (PLEG): container finished" podID="edf3a98f-0cd6-4eb7-8bf9-0320ae2186c7" containerID="2011893ca0d1cfaff73961c600f7e3a4976a22a496820c12fdceadc1ef74ba32" exitCode=0 Nov 28 13:37:27 crc kubenswrapper[4631]: I1128 13:37:27.668130 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-65dxp" event={"ID":"edf3a98f-0cd6-4eb7-8bf9-0320ae2186c7","Type":"ContainerDied","Data":"2011893ca0d1cfaff73961c600f7e3a4976a22a496820c12fdceadc1ef74ba32"} Nov 28 13:37:29 crc kubenswrapper[4631]: I1128 13:37:29.268074 4631 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-74f6bcbc87-r8g9l" podUID="166f0f43-e5cb-42ad-a2f5-1b207accc673" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.132:5353: connect: connection refused" Nov 28 13:37:34 crc kubenswrapper[4631]: I1128 13:37:34.268250 4631 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-74f6bcbc87-r8g9l" podUID="166f0f43-e5cb-42ad-a2f5-1b207accc673" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.132:5353: connect: connection refused" Nov 28 13:37:44 crc kubenswrapper[4631]: I1128 13:37:44.268189 4631 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-74f6bcbc87-r8g9l" podUID="166f0f43-e5cb-42ad-a2f5-1b207accc673" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.132:5353: i/o timeout" Nov 28 13:37:44 crc kubenswrapper[4631]: I1128 13:37:44.269403 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-74f6bcbc87-r8g9l" Nov 28 13:37:45 crc kubenswrapper[4631]: I1128 13:37:45.864196 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-65dxp" event={"ID":"edf3a98f-0cd6-4eb7-8bf9-0320ae2186c7","Type":"ContainerDied","Data":"2c727f4e04248822c558942f4dd0a9912409ecbf0ba9d180e6db14b096b0c562"} Nov 28 13:37:45 crc kubenswrapper[4631]: I1128 13:37:45.864775 4631 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2c727f4e04248822c558942f4dd0a9912409ecbf0ba9d180e6db14b096b0c562" Nov 28 13:37:45 crc kubenswrapper[4631]: I1128 13:37:45.874721 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-65dxp" Nov 28 13:37:45 crc kubenswrapper[4631]: I1128 13:37:45.907672 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/edf3a98f-0cd6-4eb7-8bf9-0320ae2186c7-combined-ca-bundle\") pod \"edf3a98f-0cd6-4eb7-8bf9-0320ae2186c7\" (UID: \"edf3a98f-0cd6-4eb7-8bf9-0320ae2186c7\") " Nov 28 13:37:45 crc kubenswrapper[4631]: I1128 13:37:45.907757 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/edf3a98f-0cd6-4eb7-8bf9-0320ae2186c7-config-data\") pod \"edf3a98f-0cd6-4eb7-8bf9-0320ae2186c7\" (UID: \"edf3a98f-0cd6-4eb7-8bf9-0320ae2186c7\") " Nov 28 13:37:45 crc kubenswrapper[4631]: I1128 13:37:45.907851 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dfmp6\" (UniqueName: \"kubernetes.io/projected/edf3a98f-0cd6-4eb7-8bf9-0320ae2186c7-kube-api-access-dfmp6\") pod \"edf3a98f-0cd6-4eb7-8bf9-0320ae2186c7\" (UID: \"edf3a98f-0cd6-4eb7-8bf9-0320ae2186c7\") " Nov 28 13:37:45 crc kubenswrapper[4631]: I1128 13:37:45.907902 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/edf3a98f-0cd6-4eb7-8bf9-0320ae2186c7-scripts\") pod \"edf3a98f-0cd6-4eb7-8bf9-0320ae2186c7\" (UID: \"edf3a98f-0cd6-4eb7-8bf9-0320ae2186c7\") " Nov 28 13:37:45 crc kubenswrapper[4631]: I1128 13:37:45.907945 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/edf3a98f-0cd6-4eb7-8bf9-0320ae2186c7-credential-keys\") pod \"edf3a98f-0cd6-4eb7-8bf9-0320ae2186c7\" (UID: \"edf3a98f-0cd6-4eb7-8bf9-0320ae2186c7\") " Nov 28 13:37:45 crc kubenswrapper[4631]: I1128 13:37:45.907991 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/edf3a98f-0cd6-4eb7-8bf9-0320ae2186c7-fernet-keys\") pod \"edf3a98f-0cd6-4eb7-8bf9-0320ae2186c7\" (UID: \"edf3a98f-0cd6-4eb7-8bf9-0320ae2186c7\") " Nov 28 13:37:45 crc kubenswrapper[4631]: I1128 13:37:45.919754 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/edf3a98f-0cd6-4eb7-8bf9-0320ae2186c7-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "edf3a98f-0cd6-4eb7-8bf9-0320ae2186c7" (UID: "edf3a98f-0cd6-4eb7-8bf9-0320ae2186c7"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:37:45 crc kubenswrapper[4631]: I1128 13:37:45.920113 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/edf3a98f-0cd6-4eb7-8bf9-0320ae2186c7-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "edf3a98f-0cd6-4eb7-8bf9-0320ae2186c7" (UID: "edf3a98f-0cd6-4eb7-8bf9-0320ae2186c7"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:37:45 crc kubenswrapper[4631]: I1128 13:37:45.920548 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/edf3a98f-0cd6-4eb7-8bf9-0320ae2186c7-scripts" (OuterVolumeSpecName: "scripts") pod "edf3a98f-0cd6-4eb7-8bf9-0320ae2186c7" (UID: "edf3a98f-0cd6-4eb7-8bf9-0320ae2186c7"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:37:45 crc kubenswrapper[4631]: I1128 13:37:45.920579 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Nov 28 13:37:45 crc kubenswrapper[4631]: I1128 13:37:45.920711 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Nov 28 13:37:45 crc kubenswrapper[4631]: I1128 13:37:45.922970 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/edf3a98f-0cd6-4eb7-8bf9-0320ae2186c7-kube-api-access-dfmp6" (OuterVolumeSpecName: "kube-api-access-dfmp6") pod "edf3a98f-0cd6-4eb7-8bf9-0320ae2186c7" (UID: "edf3a98f-0cd6-4eb7-8bf9-0320ae2186c7"). InnerVolumeSpecName "kube-api-access-dfmp6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:37:45 crc kubenswrapper[4631]: I1128 13:37:45.951693 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/edf3a98f-0cd6-4eb7-8bf9-0320ae2186c7-config-data" (OuterVolumeSpecName: "config-data") pod "edf3a98f-0cd6-4eb7-8bf9-0320ae2186c7" (UID: "edf3a98f-0cd6-4eb7-8bf9-0320ae2186c7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:37:45 crc kubenswrapper[4631]: I1128 13:37:45.952004 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/edf3a98f-0cd6-4eb7-8bf9-0320ae2186c7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "edf3a98f-0cd6-4eb7-8bf9-0320ae2186c7" (UID: "edf3a98f-0cd6-4eb7-8bf9-0320ae2186c7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:37:46 crc kubenswrapper[4631]: I1128 13:37:46.011935 4631 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/edf3a98f-0cd6-4eb7-8bf9-0320ae2186c7-fernet-keys\") on node \"crc\" DevicePath \"\"" Nov 28 13:37:46 crc kubenswrapper[4631]: I1128 13:37:46.011979 4631 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/edf3a98f-0cd6-4eb7-8bf9-0320ae2186c7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 13:37:46 crc kubenswrapper[4631]: I1128 13:37:46.012000 4631 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/edf3a98f-0cd6-4eb7-8bf9-0320ae2186c7-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 13:37:46 crc kubenswrapper[4631]: I1128 13:37:46.012014 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dfmp6\" (UniqueName: \"kubernetes.io/projected/edf3a98f-0cd6-4eb7-8bf9-0320ae2186c7-kube-api-access-dfmp6\") on node \"crc\" DevicePath \"\"" Nov 28 13:37:46 crc kubenswrapper[4631]: I1128 13:37:46.012030 4631 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/edf3a98f-0cd6-4eb7-8bf9-0320ae2186c7-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 13:37:46 crc kubenswrapper[4631]: I1128 13:37:46.012047 4631 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/edf3a98f-0cd6-4eb7-8bf9-0320ae2186c7-credential-keys\") on node \"crc\" DevicePath \"\"" Nov 28 13:37:46 crc kubenswrapper[4631]: I1128 13:37:46.672999 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Nov 28 13:37:46 crc kubenswrapper[4631]: I1128 13:37:46.673069 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Nov 28 13:37:46 crc kubenswrapper[4631]: I1128 13:37:46.874503 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-65dxp" Nov 28 13:37:46 crc kubenswrapper[4631]: I1128 13:37:46.978550 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-65dxp"] Nov 28 13:37:46 crc kubenswrapper[4631]: I1128 13:37:46.987201 4631 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-65dxp"] Nov 28 13:37:47 crc kubenswrapper[4631]: I1128 13:37:47.061263 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-dcm9x"] Nov 28 13:37:47 crc kubenswrapper[4631]: E1128 13:37:47.061952 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="edf3a98f-0cd6-4eb7-8bf9-0320ae2186c7" containerName="keystone-bootstrap" Nov 28 13:37:47 crc kubenswrapper[4631]: I1128 13:37:47.061979 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="edf3a98f-0cd6-4eb7-8bf9-0320ae2186c7" containerName="keystone-bootstrap" Nov 28 13:37:47 crc kubenswrapper[4631]: I1128 13:37:47.062185 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="edf3a98f-0cd6-4eb7-8bf9-0320ae2186c7" containerName="keystone-bootstrap" Nov 28 13:37:47 crc kubenswrapper[4631]: I1128 13:37:47.063164 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-dcm9x" Nov 28 13:37:47 crc kubenswrapper[4631]: I1128 13:37:47.071138 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-szfjn" Nov 28 13:37:47 crc kubenswrapper[4631]: I1128 13:37:47.071566 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Nov 28 13:37:47 crc kubenswrapper[4631]: I1128 13:37:47.071593 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Nov 28 13:37:47 crc kubenswrapper[4631]: I1128 13:37:47.071746 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Nov 28 13:37:47 crc kubenswrapper[4631]: I1128 13:37:47.072939 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Nov 28 13:37:47 crc kubenswrapper[4631]: I1128 13:37:47.086916 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-dcm9x"] Nov 28 13:37:47 crc kubenswrapper[4631]: I1128 13:37:47.139672 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/33e9838e-f616-4504-90d5-799de265be04-scripts\") pod \"keystone-bootstrap-dcm9x\" (UID: \"33e9838e-f616-4504-90d5-799de265be04\") " pod="openstack/keystone-bootstrap-dcm9x" Nov 28 13:37:47 crc kubenswrapper[4631]: I1128 13:37:47.139737 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/33e9838e-f616-4504-90d5-799de265be04-credential-keys\") pod \"keystone-bootstrap-dcm9x\" (UID: \"33e9838e-f616-4504-90d5-799de265be04\") " pod="openstack/keystone-bootstrap-dcm9x" Nov 28 13:37:47 crc kubenswrapper[4631]: I1128 13:37:47.139774 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/33e9838e-f616-4504-90d5-799de265be04-fernet-keys\") pod \"keystone-bootstrap-dcm9x\" (UID: \"33e9838e-f616-4504-90d5-799de265be04\") " pod="openstack/keystone-bootstrap-dcm9x" Nov 28 13:37:47 crc kubenswrapper[4631]: I1128 13:37:47.139838 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vlgfj\" (UniqueName: \"kubernetes.io/projected/33e9838e-f616-4504-90d5-799de265be04-kube-api-access-vlgfj\") pod \"keystone-bootstrap-dcm9x\" (UID: \"33e9838e-f616-4504-90d5-799de265be04\") " pod="openstack/keystone-bootstrap-dcm9x" Nov 28 13:37:47 crc kubenswrapper[4631]: I1128 13:37:47.139888 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/33e9838e-f616-4504-90d5-799de265be04-config-data\") pod \"keystone-bootstrap-dcm9x\" (UID: \"33e9838e-f616-4504-90d5-799de265be04\") " pod="openstack/keystone-bootstrap-dcm9x" Nov 28 13:37:47 crc kubenswrapper[4631]: I1128 13:37:47.139921 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/33e9838e-f616-4504-90d5-799de265be04-combined-ca-bundle\") pod \"keystone-bootstrap-dcm9x\" (UID: \"33e9838e-f616-4504-90d5-799de265be04\") " pod="openstack/keystone-bootstrap-dcm9x" Nov 28 13:37:47 crc kubenswrapper[4631]: I1128 13:37:47.242054 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vlgfj\" (UniqueName: \"kubernetes.io/projected/33e9838e-f616-4504-90d5-799de265be04-kube-api-access-vlgfj\") pod \"keystone-bootstrap-dcm9x\" (UID: \"33e9838e-f616-4504-90d5-799de265be04\") " pod="openstack/keystone-bootstrap-dcm9x" Nov 28 13:37:47 crc kubenswrapper[4631]: I1128 13:37:47.242207 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/33e9838e-f616-4504-90d5-799de265be04-config-data\") pod \"keystone-bootstrap-dcm9x\" (UID: \"33e9838e-f616-4504-90d5-799de265be04\") " pod="openstack/keystone-bootstrap-dcm9x" Nov 28 13:37:47 crc kubenswrapper[4631]: I1128 13:37:47.242336 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/33e9838e-f616-4504-90d5-799de265be04-combined-ca-bundle\") pod \"keystone-bootstrap-dcm9x\" (UID: \"33e9838e-f616-4504-90d5-799de265be04\") " pod="openstack/keystone-bootstrap-dcm9x" Nov 28 13:37:47 crc kubenswrapper[4631]: I1128 13:37:47.244825 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/33e9838e-f616-4504-90d5-799de265be04-scripts\") pod \"keystone-bootstrap-dcm9x\" (UID: \"33e9838e-f616-4504-90d5-799de265be04\") " pod="openstack/keystone-bootstrap-dcm9x" Nov 28 13:37:47 crc kubenswrapper[4631]: I1128 13:37:47.245092 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/33e9838e-f616-4504-90d5-799de265be04-credential-keys\") pod \"keystone-bootstrap-dcm9x\" (UID: \"33e9838e-f616-4504-90d5-799de265be04\") " pod="openstack/keystone-bootstrap-dcm9x" Nov 28 13:37:47 crc kubenswrapper[4631]: I1128 13:37:47.246030 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/33e9838e-f616-4504-90d5-799de265be04-fernet-keys\") pod \"keystone-bootstrap-dcm9x\" (UID: \"33e9838e-f616-4504-90d5-799de265be04\") " pod="openstack/keystone-bootstrap-dcm9x" Nov 28 13:37:47 crc kubenswrapper[4631]: I1128 13:37:47.251349 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/33e9838e-f616-4504-90d5-799de265be04-scripts\") pod \"keystone-bootstrap-dcm9x\" (UID: \"33e9838e-f616-4504-90d5-799de265be04\") " pod="openstack/keystone-bootstrap-dcm9x" Nov 28 13:37:47 crc kubenswrapper[4631]: I1128 13:37:47.255724 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/33e9838e-f616-4504-90d5-799de265be04-combined-ca-bundle\") pod \"keystone-bootstrap-dcm9x\" (UID: \"33e9838e-f616-4504-90d5-799de265be04\") " pod="openstack/keystone-bootstrap-dcm9x" Nov 28 13:37:47 crc kubenswrapper[4631]: I1128 13:37:47.257598 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/33e9838e-f616-4504-90d5-799de265be04-credential-keys\") pod \"keystone-bootstrap-dcm9x\" (UID: \"33e9838e-f616-4504-90d5-799de265be04\") " pod="openstack/keystone-bootstrap-dcm9x" Nov 28 13:37:47 crc kubenswrapper[4631]: I1128 13:37:47.259854 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/33e9838e-f616-4504-90d5-799de265be04-config-data\") pod \"keystone-bootstrap-dcm9x\" (UID: \"33e9838e-f616-4504-90d5-799de265be04\") " pod="openstack/keystone-bootstrap-dcm9x" Nov 28 13:37:47 crc kubenswrapper[4631]: I1128 13:37:47.263629 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vlgfj\" (UniqueName: \"kubernetes.io/projected/33e9838e-f616-4504-90d5-799de265be04-kube-api-access-vlgfj\") pod \"keystone-bootstrap-dcm9x\" (UID: \"33e9838e-f616-4504-90d5-799de265be04\") " pod="openstack/keystone-bootstrap-dcm9x" Nov 28 13:37:47 crc kubenswrapper[4631]: I1128 13:37:47.264709 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/33e9838e-f616-4504-90d5-799de265be04-fernet-keys\") pod \"keystone-bootstrap-dcm9x\" (UID: \"33e9838e-f616-4504-90d5-799de265be04\") " pod="openstack/keystone-bootstrap-dcm9x" Nov 28 13:37:47 crc kubenswrapper[4631]: I1128 13:37:47.400202 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-dcm9x" Nov 28 13:37:47 crc kubenswrapper[4631]: I1128 13:37:47.528344 4631 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="edf3a98f-0cd6-4eb7-8bf9-0320ae2186c7" path="/var/lib/kubelet/pods/edf3a98f-0cd6-4eb7-8bf9-0320ae2186c7/volumes" Nov 28 13:37:47 crc kubenswrapper[4631]: E1128 13:37:47.624007 4631 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-horizon:current-podified" Nov 28 13:37:47 crc kubenswrapper[4631]: E1128 13:37:47.624258 4631 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizon-log,Image:quay.io/podified-antelope-centos9/openstack-horizon:current-podified,Command:[/bin/bash],Args:[-c tail -n+1 -F /var/log/horizon/horizon.log],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n686h684h64fhc5hfdh557h97h78h554h547h57bh644h9h588h68dh5f6h5d7h67dh546h5bbh664h5bfhfhbfhbbhf8h568h5ddh59dh68ch9ch64bq,ValueFrom:nil,},EnvVar{Name:ENABLE_DESIGNATE,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_HEAT,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_IRONIC,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_MANILA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_OCTAVIA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_WATCHER,Value:no,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:UNPACK_THEME,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/horizon,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-44jtf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*48,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*42400,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-59495dc969-4lttm_openstack(7f02c913-7539-481a-abfc-7e19a28529e1): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 28 13:37:47 crc kubenswrapper[4631]: E1128 13:37:47.627274 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-horizon:current-podified\\\"\"]" pod="openstack/horizon-59495dc969-4lttm" podUID="7f02c913-7539-481a-abfc-7e19a28529e1" Nov 28 13:37:47 crc kubenswrapper[4631]: E1128 13:37:47.640207 4631 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-placement-api:current-podified" Nov 28 13:37:47 crc kubenswrapper[4631]: E1128 13:37:47.640426 4631 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:placement-db-sync,Image:quay.io/podified-antelope-centos9/openstack-placement-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/placement,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:false,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:placement-dbsync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-t6gw4,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42482,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-db-sync-xhft6_openstack(05e73b90-be19-4f14-9824-a5c27361adff): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 28 13:37:47 crc kubenswrapper[4631]: E1128 13:37:47.641884 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"placement-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/placement-db-sync-xhft6" podUID="05e73b90-be19-4f14-9824-a5c27361adff" Nov 28 13:37:47 crc kubenswrapper[4631]: I1128 13:37:47.888117 4631 generic.go:334] "Generic (PLEG): container finished" podID="268feb7e-4677-4c47-b837-e04bab4b8010" containerID="22d98a24eacf769b7c3570fb24c8bf34f199e9c081e79a87c78dffae0d919cc9" exitCode=0 Nov 28 13:37:47 crc kubenswrapper[4631]: I1128 13:37:47.888219 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-6lwcj" event={"ID":"268feb7e-4677-4c47-b837-e04bab4b8010","Type":"ContainerDied","Data":"22d98a24eacf769b7c3570fb24c8bf34f199e9c081e79a87c78dffae0d919cc9"} Nov 28 13:37:47 crc kubenswrapper[4631]: E1128 13:37:47.891158 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"placement-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-placement-api:current-podified\\\"\"" pod="openstack/placement-db-sync-xhft6" podUID="05e73b90-be19-4f14-9824-a5c27361adff" Nov 28 13:37:49 crc kubenswrapper[4631]: I1128 13:37:49.270041 4631 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-74f6bcbc87-r8g9l" podUID="166f0f43-e5cb-42ad-a2f5-1b207accc673" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.132:5353: i/o timeout" Nov 28 13:37:49 crc kubenswrapper[4631]: I1128 13:37:49.635446 4631 patch_prober.go:28] interesting pod/machine-config-daemon-47sxw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 28 13:37:49 crc kubenswrapper[4631]: I1128 13:37:49.635960 4631 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" podUID="8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 28 13:37:49 crc kubenswrapper[4631]: I1128 13:37:49.636308 4631 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" Nov 28 13:37:49 crc kubenswrapper[4631]: I1128 13:37:49.637344 4631 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"a4b336d735e84003b86dbce21368f3f29fb0ded3e7a12a4562271fa2767a4ae8"} pod="openshift-machine-config-operator/machine-config-daemon-47sxw" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 28 13:37:49 crc kubenswrapper[4631]: I1128 13:37:49.637404 4631 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" podUID="8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1" containerName="machine-config-daemon" containerID="cri-o://a4b336d735e84003b86dbce21368f3f29fb0ded3e7a12a4562271fa2767a4ae8" gracePeriod=600 Nov 28 13:37:49 crc kubenswrapper[4631]: E1128 13:37:49.904211 4631 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-horizon:current-podified" Nov 28 13:37:49 crc kubenswrapper[4631]: E1128 13:37:49.904527 4631 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizon-log,Image:quay.io/podified-antelope-centos9/openstack-horizon:current-podified,Command:[/bin/bash],Args:[-c tail -n+1 -F /var/log/horizon/horizon.log],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n5b8hbh565h76h5ch66h657h5f6h55ch66bhfchd4h549h4h57h94h647h5fdh5dfh54fh556h577h65fh67h584h7h5cfhbbh59dhddh55bhf5q,ValueFrom:nil,},EnvVar{Name:ENABLE_DESIGNATE,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_HEAT,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_IRONIC,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_MANILA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_OCTAVIA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_WATCHER,Value:no,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:UNPACK_THEME,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/horizon,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-j9kq8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*48,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*42400,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-86566f7dbc-2zlk9_openstack(cec87f6b-b6fd-467c-adb9-f6c6ff099a18): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 28 13:37:49 crc kubenswrapper[4631]: E1128 13:37:49.911676 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-horizon:current-podified\\\"\"]" pod="openstack/horizon-86566f7dbc-2zlk9" podUID="cec87f6b-b6fd-467c-adb9-f6c6ff099a18" Nov 28 13:37:49 crc kubenswrapper[4631]: I1128 13:37:49.933911 4631 generic.go:334] "Generic (PLEG): container finished" podID="8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1" containerID="a4b336d735e84003b86dbce21368f3f29fb0ded3e7a12a4562271fa2767a4ae8" exitCode=0 Nov 28 13:37:49 crc kubenswrapper[4631]: I1128 13:37:49.934011 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" event={"ID":"8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1","Type":"ContainerDied","Data":"a4b336d735e84003b86dbce21368f3f29fb0ded3e7a12a4562271fa2767a4ae8"} Nov 28 13:37:49 crc kubenswrapper[4631]: I1128 13:37:49.934065 4631 scope.go:117] "RemoveContainer" containerID="914e9854c9ef96234991a24021b15fad0f64f1894afc1a19ce66cd4bc2c2af7d" Nov 28 13:37:50 crc kubenswrapper[4631]: I1128 13:37:50.042987 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74f6bcbc87-r8g9l" Nov 28 13:37:50 crc kubenswrapper[4631]: I1128 13:37:50.138372 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/166f0f43-e5cb-42ad-a2f5-1b207accc673-dns-svc\") pod \"166f0f43-e5cb-42ad-a2f5-1b207accc673\" (UID: \"166f0f43-e5cb-42ad-a2f5-1b207accc673\") " Nov 28 13:37:50 crc kubenswrapper[4631]: I1128 13:37:50.138426 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/166f0f43-e5cb-42ad-a2f5-1b207accc673-dns-swift-storage-0\") pod \"166f0f43-e5cb-42ad-a2f5-1b207accc673\" (UID: \"166f0f43-e5cb-42ad-a2f5-1b207accc673\") " Nov 28 13:37:50 crc kubenswrapper[4631]: I1128 13:37:50.138451 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/166f0f43-e5cb-42ad-a2f5-1b207accc673-ovsdbserver-nb\") pod \"166f0f43-e5cb-42ad-a2f5-1b207accc673\" (UID: \"166f0f43-e5cb-42ad-a2f5-1b207accc673\") " Nov 28 13:37:50 crc kubenswrapper[4631]: I1128 13:37:50.138514 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/166f0f43-e5cb-42ad-a2f5-1b207accc673-ovsdbserver-sb\") pod \"166f0f43-e5cb-42ad-a2f5-1b207accc673\" (UID: \"166f0f43-e5cb-42ad-a2f5-1b207accc673\") " Nov 28 13:37:50 crc kubenswrapper[4631]: I1128 13:37:50.138563 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/166f0f43-e5cb-42ad-a2f5-1b207accc673-config\") pod \"166f0f43-e5cb-42ad-a2f5-1b207accc673\" (UID: \"166f0f43-e5cb-42ad-a2f5-1b207accc673\") " Nov 28 13:37:50 crc kubenswrapper[4631]: I1128 13:37:50.138595 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-brqml\" (UniqueName: \"kubernetes.io/projected/166f0f43-e5cb-42ad-a2f5-1b207accc673-kube-api-access-brqml\") pod \"166f0f43-e5cb-42ad-a2f5-1b207accc673\" (UID: \"166f0f43-e5cb-42ad-a2f5-1b207accc673\") " Nov 28 13:37:50 crc kubenswrapper[4631]: I1128 13:37:50.166846 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/166f0f43-e5cb-42ad-a2f5-1b207accc673-kube-api-access-brqml" (OuterVolumeSpecName: "kube-api-access-brqml") pod "166f0f43-e5cb-42ad-a2f5-1b207accc673" (UID: "166f0f43-e5cb-42ad-a2f5-1b207accc673"). InnerVolumeSpecName "kube-api-access-brqml". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:37:50 crc kubenswrapper[4631]: I1128 13:37:50.226221 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/166f0f43-e5cb-42ad-a2f5-1b207accc673-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "166f0f43-e5cb-42ad-a2f5-1b207accc673" (UID: "166f0f43-e5cb-42ad-a2f5-1b207accc673"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 13:37:50 crc kubenswrapper[4631]: I1128 13:37:50.229700 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/166f0f43-e5cb-42ad-a2f5-1b207accc673-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "166f0f43-e5cb-42ad-a2f5-1b207accc673" (UID: "166f0f43-e5cb-42ad-a2f5-1b207accc673"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 13:37:50 crc kubenswrapper[4631]: I1128 13:37:50.230587 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/166f0f43-e5cb-42ad-a2f5-1b207accc673-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "166f0f43-e5cb-42ad-a2f5-1b207accc673" (UID: "166f0f43-e5cb-42ad-a2f5-1b207accc673"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 13:37:50 crc kubenswrapper[4631]: I1128 13:37:50.234376 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/166f0f43-e5cb-42ad-a2f5-1b207accc673-config" (OuterVolumeSpecName: "config") pod "166f0f43-e5cb-42ad-a2f5-1b207accc673" (UID: "166f0f43-e5cb-42ad-a2f5-1b207accc673"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 13:37:50 crc kubenswrapper[4631]: I1128 13:37:50.240743 4631 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/166f0f43-e5cb-42ad-a2f5-1b207accc673-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 28 13:37:50 crc kubenswrapper[4631]: I1128 13:37:50.241513 4631 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/166f0f43-e5cb-42ad-a2f5-1b207accc673-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 28 13:37:50 crc kubenswrapper[4631]: I1128 13:37:50.241550 4631 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/166f0f43-e5cb-42ad-a2f5-1b207accc673-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 28 13:37:50 crc kubenswrapper[4631]: I1128 13:37:50.241564 4631 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/166f0f43-e5cb-42ad-a2f5-1b207accc673-config\") on node \"crc\" DevicePath \"\"" Nov 28 13:37:50 crc kubenswrapper[4631]: I1128 13:37:50.241576 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-brqml\" (UniqueName: \"kubernetes.io/projected/166f0f43-e5cb-42ad-a2f5-1b207accc673-kube-api-access-brqml\") on node \"crc\" DevicePath \"\"" Nov 28 13:37:50 crc kubenswrapper[4631]: I1128 13:37:50.260599 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/166f0f43-e5cb-42ad-a2f5-1b207accc673-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "166f0f43-e5cb-42ad-a2f5-1b207accc673" (UID: "166f0f43-e5cb-42ad-a2f5-1b207accc673"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 13:37:50 crc kubenswrapper[4631]: I1128 13:37:50.368193 4631 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/166f0f43-e5cb-42ad-a2f5-1b207accc673-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 28 13:37:50 crc kubenswrapper[4631]: I1128 13:37:50.947889 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74f6bcbc87-r8g9l" event={"ID":"166f0f43-e5cb-42ad-a2f5-1b207accc673","Type":"ContainerDied","Data":"53ac8e91a456a4174a50d67629061bd3506db6839b116a69454647eb2c4a7266"} Nov 28 13:37:50 crc kubenswrapper[4631]: I1128 13:37:50.948109 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74f6bcbc87-r8g9l" Nov 28 13:37:51 crc kubenswrapper[4631]: I1128 13:37:51.014952 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-74f6bcbc87-r8g9l"] Nov 28 13:37:51 crc kubenswrapper[4631]: I1128 13:37:51.027724 4631 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-74f6bcbc87-r8g9l"] Nov 28 13:37:51 crc kubenswrapper[4631]: I1128 13:37:51.526781 4631 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="166f0f43-e5cb-42ad-a2f5-1b207accc673" path="/var/lib/kubelet/pods/166f0f43-e5cb-42ad-a2f5-1b207accc673/volumes" Nov 28 13:37:51 crc kubenswrapper[4631]: E1128 13:37:51.695493 4631 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified" Nov 28 13:37:51 crc kubenswrapper[4631]: E1128 13:37:51.695859 4631 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cinder-db-sync,Image:quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-machine-id,ReadOnly:true,MountPath:/etc/machine-id,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/merged,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/cinder/cinder.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-lrt64,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-db-sync-gprj7_openstack(af357d67-237c-468a-81ac-990964e8e8b2): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 28 13:37:51 crc kubenswrapper[4631]: E1128 13:37:51.704146 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/cinder-db-sync-gprj7" podUID="af357d67-237c-468a-81ac-990964e8e8b2" Nov 28 13:37:51 crc kubenswrapper[4631]: I1128 13:37:51.806710 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-59495dc969-4lttm" Nov 28 13:37:51 crc kubenswrapper[4631]: I1128 13:37:51.817574 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-6lwcj" Nov 28 13:37:51 crc kubenswrapper[4631]: I1128 13:37:51.917901 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/7f02c913-7539-481a-abfc-7e19a28529e1-config-data\") pod \"7f02c913-7539-481a-abfc-7e19a28529e1\" (UID: \"7f02c913-7539-481a-abfc-7e19a28529e1\") " Nov 28 13:37:51 crc kubenswrapper[4631]: I1128 13:37:51.917974 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-758jr\" (UniqueName: \"kubernetes.io/projected/268feb7e-4677-4c47-b837-e04bab4b8010-kube-api-access-758jr\") pod \"268feb7e-4677-4c47-b837-e04bab4b8010\" (UID: \"268feb7e-4677-4c47-b837-e04bab4b8010\") " Nov 28 13:37:51 crc kubenswrapper[4631]: I1128 13:37:51.918011 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/268feb7e-4677-4c47-b837-e04bab4b8010-combined-ca-bundle\") pod \"268feb7e-4677-4c47-b837-e04bab4b8010\" (UID: \"268feb7e-4677-4c47-b837-e04bab4b8010\") " Nov 28 13:37:51 crc kubenswrapper[4631]: I1128 13:37:51.918109 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/7f02c913-7539-481a-abfc-7e19a28529e1-horizon-secret-key\") pod \"7f02c913-7539-481a-abfc-7e19a28529e1\" (UID: \"7f02c913-7539-481a-abfc-7e19a28529e1\") " Nov 28 13:37:51 crc kubenswrapper[4631]: I1128 13:37:51.918138 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7f02c913-7539-481a-abfc-7e19a28529e1-scripts\") pod \"7f02c913-7539-481a-abfc-7e19a28529e1\" (UID: \"7f02c913-7539-481a-abfc-7e19a28529e1\") " Nov 28 13:37:51 crc kubenswrapper[4631]: I1128 13:37:51.918167 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/268feb7e-4677-4c47-b837-e04bab4b8010-config\") pod \"268feb7e-4677-4c47-b837-e04bab4b8010\" (UID: \"268feb7e-4677-4c47-b837-e04bab4b8010\") " Nov 28 13:37:51 crc kubenswrapper[4631]: I1128 13:37:51.918230 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-44jtf\" (UniqueName: \"kubernetes.io/projected/7f02c913-7539-481a-abfc-7e19a28529e1-kube-api-access-44jtf\") pod \"7f02c913-7539-481a-abfc-7e19a28529e1\" (UID: \"7f02c913-7539-481a-abfc-7e19a28529e1\") " Nov 28 13:37:51 crc kubenswrapper[4631]: I1128 13:37:51.918258 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7f02c913-7539-481a-abfc-7e19a28529e1-logs\") pod \"7f02c913-7539-481a-abfc-7e19a28529e1\" (UID: \"7f02c913-7539-481a-abfc-7e19a28529e1\") " Nov 28 13:37:51 crc kubenswrapper[4631]: I1128 13:37:51.919172 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7f02c913-7539-481a-abfc-7e19a28529e1-logs" (OuterVolumeSpecName: "logs") pod "7f02c913-7539-481a-abfc-7e19a28529e1" (UID: "7f02c913-7539-481a-abfc-7e19a28529e1"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 13:37:51 crc kubenswrapper[4631]: I1128 13:37:51.919535 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7f02c913-7539-481a-abfc-7e19a28529e1-scripts" (OuterVolumeSpecName: "scripts") pod "7f02c913-7539-481a-abfc-7e19a28529e1" (UID: "7f02c913-7539-481a-abfc-7e19a28529e1"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 13:37:51 crc kubenswrapper[4631]: I1128 13:37:51.920305 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7f02c913-7539-481a-abfc-7e19a28529e1-config-data" (OuterVolumeSpecName: "config-data") pod "7f02c913-7539-481a-abfc-7e19a28529e1" (UID: "7f02c913-7539-481a-abfc-7e19a28529e1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 13:37:51 crc kubenswrapper[4631]: I1128 13:37:51.926192 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7f02c913-7539-481a-abfc-7e19a28529e1-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "7f02c913-7539-481a-abfc-7e19a28529e1" (UID: "7f02c913-7539-481a-abfc-7e19a28529e1"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:37:51 crc kubenswrapper[4631]: I1128 13:37:51.926598 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7f02c913-7539-481a-abfc-7e19a28529e1-kube-api-access-44jtf" (OuterVolumeSpecName: "kube-api-access-44jtf") pod "7f02c913-7539-481a-abfc-7e19a28529e1" (UID: "7f02c913-7539-481a-abfc-7e19a28529e1"). InnerVolumeSpecName "kube-api-access-44jtf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:37:51 crc kubenswrapper[4631]: I1128 13:37:51.928243 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/268feb7e-4677-4c47-b837-e04bab4b8010-kube-api-access-758jr" (OuterVolumeSpecName: "kube-api-access-758jr") pod "268feb7e-4677-4c47-b837-e04bab4b8010" (UID: "268feb7e-4677-4c47-b837-e04bab4b8010"). InnerVolumeSpecName "kube-api-access-758jr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:37:51 crc kubenswrapper[4631]: I1128 13:37:51.955691 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/268feb7e-4677-4c47-b837-e04bab4b8010-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "268feb7e-4677-4c47-b837-e04bab4b8010" (UID: "268feb7e-4677-4c47-b837-e04bab4b8010"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:37:51 crc kubenswrapper[4631]: I1128 13:37:51.957509 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/268feb7e-4677-4c47-b837-e04bab4b8010-config" (OuterVolumeSpecName: "config") pod "268feb7e-4677-4c47-b837-e04bab4b8010" (UID: "268feb7e-4677-4c47-b837-e04bab4b8010"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:37:51 crc kubenswrapper[4631]: I1128 13:37:51.959531 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-59495dc969-4lttm" Nov 28 13:37:51 crc kubenswrapper[4631]: I1128 13:37:51.959521 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-59495dc969-4lttm" event={"ID":"7f02c913-7539-481a-abfc-7e19a28529e1","Type":"ContainerDied","Data":"a466803046ef5f5b14028f4dd4a8e94216a4b7b4638b3d67a4c58eddb54eff68"} Nov 28 13:37:51 crc kubenswrapper[4631]: I1128 13:37:51.961502 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-6lwcj" event={"ID":"268feb7e-4677-4c47-b837-e04bab4b8010","Type":"ContainerDied","Data":"d05c339a07f8f431386a5d5539436c081d7d6b0bccd245b02b7ef81095085814"} Nov 28 13:37:51 crc kubenswrapper[4631]: I1128 13:37:51.961543 4631 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d05c339a07f8f431386a5d5539436c081d7d6b0bccd245b02b7ef81095085814" Nov 28 13:37:51 crc kubenswrapper[4631]: I1128 13:37:51.961542 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-6lwcj" Nov 28 13:37:51 crc kubenswrapper[4631]: E1128 13:37:51.965502 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified\\\"\"" pod="openstack/cinder-db-sync-gprj7" podUID="af357d67-237c-468a-81ac-990964e8e8b2" Nov 28 13:37:52 crc kubenswrapper[4631]: I1128 13:37:52.021488 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-758jr\" (UniqueName: \"kubernetes.io/projected/268feb7e-4677-4c47-b837-e04bab4b8010-kube-api-access-758jr\") on node \"crc\" DevicePath \"\"" Nov 28 13:37:52 crc kubenswrapper[4631]: I1128 13:37:52.021519 4631 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/268feb7e-4677-4c47-b837-e04bab4b8010-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 13:37:52 crc kubenswrapper[4631]: I1128 13:37:52.021529 4631 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/7f02c913-7539-481a-abfc-7e19a28529e1-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Nov 28 13:37:52 crc kubenswrapper[4631]: I1128 13:37:52.021539 4631 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7f02c913-7539-481a-abfc-7e19a28529e1-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 13:37:52 crc kubenswrapper[4631]: I1128 13:37:52.021560 4631 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/268feb7e-4677-4c47-b837-e04bab4b8010-config\") on node \"crc\" DevicePath \"\"" Nov 28 13:37:52 crc kubenswrapper[4631]: I1128 13:37:52.021569 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-44jtf\" (UniqueName: \"kubernetes.io/projected/7f02c913-7539-481a-abfc-7e19a28529e1-kube-api-access-44jtf\") on node \"crc\" DevicePath \"\"" Nov 28 13:37:52 crc kubenswrapper[4631]: I1128 13:37:52.021578 4631 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7f02c913-7539-481a-abfc-7e19a28529e1-logs\") on node \"crc\" DevicePath \"\"" Nov 28 13:37:52 crc kubenswrapper[4631]: I1128 13:37:52.021591 4631 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/7f02c913-7539-481a-abfc-7e19a28529e1-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 13:37:52 crc kubenswrapper[4631]: I1128 13:37:52.081807 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-59495dc969-4lttm"] Nov 28 13:37:52 crc kubenswrapper[4631]: I1128 13:37:52.093729 4631 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-59495dc969-4lttm"] Nov 28 13:37:52 crc kubenswrapper[4631]: E1128 13:37:52.235871 4631 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7f02c913_7539_481a_abfc_7e19a28529e1.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7f02c913_7539_481a_abfc_7e19a28529e1.slice/crio-a466803046ef5f5b14028f4dd4a8e94216a4b7b4638b3d67a4c58eddb54eff68\": RecentStats: unable to find data in memory cache]" Nov 28 13:37:52 crc kubenswrapper[4631]: E1128 13:37:52.439431 4631 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified" Nov 28 13:37:52 crc kubenswrapper[4631]: E1128 13:37:52.440064 4631 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:barbican-db-sync,Image:quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified,Command:[/bin/bash],Args:[-c barbican-manage db upgrade],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/barbican/barbican.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-qjbrq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42403,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42403,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod barbican-db-sync-z6tgb_openstack(3050013d-2db4-4801-8626-1fb6398b714e): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 28 13:37:52 crc kubenswrapper[4631]: E1128 13:37:52.441678 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/barbican-db-sync-z6tgb" podUID="3050013d-2db4-4801-8626-1fb6398b714e" Nov 28 13:37:52 crc kubenswrapper[4631]: I1128 13:37:52.689364 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 28 13:37:52 crc kubenswrapper[4631]: I1128 13:37:52.721433 4631 scope.go:117] "RemoveContainer" containerID="e9a7a7a5cf57eda39ae2aac9a52f8fe72136a44113973538ac6ec5d2590a54ef" Nov 28 13:37:52 crc kubenswrapper[4631]: I1128 13:37:52.785192 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 28 13:37:52 crc kubenswrapper[4631]: I1128 13:37:52.800044 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-86566f7dbc-2zlk9" Nov 28 13:37:52 crc kubenswrapper[4631]: I1128 13:37:52.864406 4631 scope.go:117] "RemoveContainer" containerID="d35b76cd7932fe54552de2a63b3892d588152fa871b777d24b6c7c5406ac5703" Nov 28 13:37:52 crc kubenswrapper[4631]: I1128 13:37:52.864413 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"e35e9173-ceb1-4f6a-8a76-2a8c869aa26b\" (UID: \"e35e9173-ceb1-4f6a-8a76-2a8c869aa26b\") " Nov 28 13:37:52 crc kubenswrapper[4631]: I1128 13:37:52.864726 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e35e9173-ceb1-4f6a-8a76-2a8c869aa26b-public-tls-certs\") pod \"e35e9173-ceb1-4f6a-8a76-2a8c869aa26b\" (UID: \"e35e9173-ceb1-4f6a-8a76-2a8c869aa26b\") " Nov 28 13:37:52 crc kubenswrapper[4631]: I1128 13:37:52.864783 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6hlqb\" (UniqueName: \"kubernetes.io/projected/ed809148-a421-492c-a7c1-6a9e4eba0528-kube-api-access-6hlqb\") pod \"ed809148-a421-492c-a7c1-6a9e4eba0528\" (UID: \"ed809148-a421-492c-a7c1-6a9e4eba0528\") " Nov 28 13:37:52 crc kubenswrapper[4631]: I1128 13:37:52.864882 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cqbq9\" (UniqueName: \"kubernetes.io/projected/e35e9173-ceb1-4f6a-8a76-2a8c869aa26b-kube-api-access-cqbq9\") pod \"e35e9173-ceb1-4f6a-8a76-2a8c869aa26b\" (UID: \"e35e9173-ceb1-4f6a-8a76-2a8c869aa26b\") " Nov 28 13:37:52 crc kubenswrapper[4631]: I1128 13:37:52.864908 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e35e9173-ceb1-4f6a-8a76-2a8c869aa26b-combined-ca-bundle\") pod \"e35e9173-ceb1-4f6a-8a76-2a8c869aa26b\" (UID: \"e35e9173-ceb1-4f6a-8a76-2a8c869aa26b\") " Nov 28 13:37:52 crc kubenswrapper[4631]: I1128 13:37:52.865009 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e35e9173-ceb1-4f6a-8a76-2a8c869aa26b-config-data\") pod \"e35e9173-ceb1-4f6a-8a76-2a8c869aa26b\" (UID: \"e35e9173-ceb1-4f6a-8a76-2a8c869aa26b\") " Nov 28 13:37:52 crc kubenswrapper[4631]: I1128 13:37:52.865044 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ed809148-a421-492c-a7c1-6a9e4eba0528-httpd-run\") pod \"ed809148-a421-492c-a7c1-6a9e4eba0528\" (UID: \"ed809148-a421-492c-a7c1-6a9e4eba0528\") " Nov 28 13:37:52 crc kubenswrapper[4631]: I1128 13:37:52.865094 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e35e9173-ceb1-4f6a-8a76-2a8c869aa26b-logs\") pod \"e35e9173-ceb1-4f6a-8a76-2a8c869aa26b\" (UID: \"e35e9173-ceb1-4f6a-8a76-2a8c869aa26b\") " Nov 28 13:37:52 crc kubenswrapper[4631]: I1128 13:37:52.865126 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ed809148-a421-492c-a7c1-6a9e4eba0528-combined-ca-bundle\") pod \"ed809148-a421-492c-a7c1-6a9e4eba0528\" (UID: \"ed809148-a421-492c-a7c1-6a9e4eba0528\") " Nov 28 13:37:52 crc kubenswrapper[4631]: I1128 13:37:52.865201 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ed809148-a421-492c-a7c1-6a9e4eba0528-config-data\") pod \"ed809148-a421-492c-a7c1-6a9e4eba0528\" (UID: \"ed809148-a421-492c-a7c1-6a9e4eba0528\") " Nov 28 13:37:52 crc kubenswrapper[4631]: I1128 13:37:52.865230 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ed809148-a421-492c-a7c1-6a9e4eba0528-scripts\") pod \"ed809148-a421-492c-a7c1-6a9e4eba0528\" (UID: \"ed809148-a421-492c-a7c1-6a9e4eba0528\") " Nov 28 13:37:52 crc kubenswrapper[4631]: I1128 13:37:52.865258 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e35e9173-ceb1-4f6a-8a76-2a8c869aa26b-scripts\") pod \"e35e9173-ceb1-4f6a-8a76-2a8c869aa26b\" (UID: \"e35e9173-ceb1-4f6a-8a76-2a8c869aa26b\") " Nov 28 13:37:52 crc kubenswrapper[4631]: I1128 13:37:52.865325 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"ed809148-a421-492c-a7c1-6a9e4eba0528\" (UID: \"ed809148-a421-492c-a7c1-6a9e4eba0528\") " Nov 28 13:37:52 crc kubenswrapper[4631]: I1128 13:37:52.865354 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ed809148-a421-492c-a7c1-6a9e4eba0528-internal-tls-certs\") pod \"ed809148-a421-492c-a7c1-6a9e4eba0528\" (UID: \"ed809148-a421-492c-a7c1-6a9e4eba0528\") " Nov 28 13:37:52 crc kubenswrapper[4631]: I1128 13:37:52.865408 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e35e9173-ceb1-4f6a-8a76-2a8c869aa26b-httpd-run\") pod \"e35e9173-ceb1-4f6a-8a76-2a8c869aa26b\" (UID: \"e35e9173-ceb1-4f6a-8a76-2a8c869aa26b\") " Nov 28 13:37:52 crc kubenswrapper[4631]: I1128 13:37:52.865445 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ed809148-a421-492c-a7c1-6a9e4eba0528-logs\") pod \"ed809148-a421-492c-a7c1-6a9e4eba0528\" (UID: \"ed809148-a421-492c-a7c1-6a9e4eba0528\") " Nov 28 13:37:52 crc kubenswrapper[4631]: I1128 13:37:52.867797 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ed809148-a421-492c-a7c1-6a9e4eba0528-logs" (OuterVolumeSpecName: "logs") pod "ed809148-a421-492c-a7c1-6a9e4eba0528" (UID: "ed809148-a421-492c-a7c1-6a9e4eba0528"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 13:37:52 crc kubenswrapper[4631]: I1128 13:37:52.876662 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ed809148-a421-492c-a7c1-6a9e4eba0528-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "ed809148-a421-492c-a7c1-6a9e4eba0528" (UID: "ed809148-a421-492c-a7c1-6a9e4eba0528"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 13:37:52 crc kubenswrapper[4631]: I1128 13:37:52.877637 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e35e9173-ceb1-4f6a-8a76-2a8c869aa26b-logs" (OuterVolumeSpecName: "logs") pod "e35e9173-ceb1-4f6a-8a76-2a8c869aa26b" (UID: "e35e9173-ceb1-4f6a-8a76-2a8c869aa26b"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 13:37:52 crc kubenswrapper[4631]: I1128 13:37:52.877665 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage04-crc" (OuterVolumeSpecName: "glance") pod "e35e9173-ceb1-4f6a-8a76-2a8c869aa26b" (UID: "e35e9173-ceb1-4f6a-8a76-2a8c869aa26b"). InnerVolumeSpecName "local-storage04-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 28 13:37:52 crc kubenswrapper[4631]: I1128 13:37:52.881593 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e35e9173-ceb1-4f6a-8a76-2a8c869aa26b-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "e35e9173-ceb1-4f6a-8a76-2a8c869aa26b" (UID: "e35e9173-ceb1-4f6a-8a76-2a8c869aa26b"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 13:37:52 crc kubenswrapper[4631]: I1128 13:37:52.896711 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage08-crc" (OuterVolumeSpecName: "glance") pod "ed809148-a421-492c-a7c1-6a9e4eba0528" (UID: "ed809148-a421-492c-a7c1-6a9e4eba0528"). InnerVolumeSpecName "local-storage08-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 28 13:37:52 crc kubenswrapper[4631]: I1128 13:37:52.919648 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ed809148-a421-492c-a7c1-6a9e4eba0528-scripts" (OuterVolumeSpecName: "scripts") pod "ed809148-a421-492c-a7c1-6a9e4eba0528" (UID: "ed809148-a421-492c-a7c1-6a9e4eba0528"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:37:52 crc kubenswrapper[4631]: I1128 13:37:52.919710 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e35e9173-ceb1-4f6a-8a76-2a8c869aa26b-scripts" (OuterVolumeSpecName: "scripts") pod "e35e9173-ceb1-4f6a-8a76-2a8c869aa26b" (UID: "e35e9173-ceb1-4f6a-8a76-2a8c869aa26b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:37:52 crc kubenswrapper[4631]: I1128 13:37:52.919736 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ed809148-a421-492c-a7c1-6a9e4eba0528-kube-api-access-6hlqb" (OuterVolumeSpecName: "kube-api-access-6hlqb") pod "ed809148-a421-492c-a7c1-6a9e4eba0528" (UID: "ed809148-a421-492c-a7c1-6a9e4eba0528"). InnerVolumeSpecName "kube-api-access-6hlqb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:37:52 crc kubenswrapper[4631]: I1128 13:37:52.919858 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e35e9173-ceb1-4f6a-8a76-2a8c869aa26b-kube-api-access-cqbq9" (OuterVolumeSpecName: "kube-api-access-cqbq9") pod "e35e9173-ceb1-4f6a-8a76-2a8c869aa26b" (UID: "e35e9173-ceb1-4f6a-8a76-2a8c869aa26b"). InnerVolumeSpecName "kube-api-access-cqbq9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:37:52 crc kubenswrapper[4631]: I1128 13:37:52.979904 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/cec87f6b-b6fd-467c-adb9-f6c6ff099a18-horizon-secret-key\") pod \"cec87f6b-b6fd-467c-adb9-f6c6ff099a18\" (UID: \"cec87f6b-b6fd-467c-adb9-f6c6ff099a18\") " Nov 28 13:37:52 crc kubenswrapper[4631]: I1128 13:37:52.979952 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/cec87f6b-b6fd-467c-adb9-f6c6ff099a18-scripts\") pod \"cec87f6b-b6fd-467c-adb9-f6c6ff099a18\" (UID: \"cec87f6b-b6fd-467c-adb9-f6c6ff099a18\") " Nov 28 13:37:52 crc kubenswrapper[4631]: I1128 13:37:52.981452 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j9kq8\" (UniqueName: \"kubernetes.io/projected/cec87f6b-b6fd-467c-adb9-f6c6ff099a18-kube-api-access-j9kq8\") pod \"cec87f6b-b6fd-467c-adb9-f6c6ff099a18\" (UID: \"cec87f6b-b6fd-467c-adb9-f6c6ff099a18\") " Nov 28 13:37:52 crc kubenswrapper[4631]: I1128 13:37:52.981943 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/cec87f6b-b6fd-467c-adb9-f6c6ff099a18-config-data\") pod \"cec87f6b-b6fd-467c-adb9-f6c6ff099a18\" (UID: \"cec87f6b-b6fd-467c-adb9-f6c6ff099a18\") " Nov 28 13:37:52 crc kubenswrapper[4631]: I1128 13:37:52.982186 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cec87f6b-b6fd-467c-adb9-f6c6ff099a18-logs\") pod \"cec87f6b-b6fd-467c-adb9-f6c6ff099a18\" (UID: \"cec87f6b-b6fd-467c-adb9-f6c6ff099a18\") " Nov 28 13:37:52 crc kubenswrapper[4631]: I1128 13:37:52.987255 4631 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ed809148-a421-492c-a7c1-6a9e4eba0528-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 13:37:52 crc kubenswrapper[4631]: I1128 13:37:52.987346 4631 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e35e9173-ceb1-4f6a-8a76-2a8c869aa26b-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 13:37:52 crc kubenswrapper[4631]: I1128 13:37:52.987444 4631 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" " Nov 28 13:37:52 crc kubenswrapper[4631]: I1128 13:37:52.987544 4631 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e35e9173-ceb1-4f6a-8a76-2a8c869aa26b-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 28 13:37:52 crc kubenswrapper[4631]: I1128 13:37:52.987593 4631 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ed809148-a421-492c-a7c1-6a9e4eba0528-logs\") on node \"crc\" DevicePath \"\"" Nov 28 13:37:52 crc kubenswrapper[4631]: I1128 13:37:52.987619 4631 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" " Nov 28 13:37:52 crc kubenswrapper[4631]: I1128 13:37:52.987634 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6hlqb\" (UniqueName: \"kubernetes.io/projected/ed809148-a421-492c-a7c1-6a9e4eba0528-kube-api-access-6hlqb\") on node \"crc\" DevicePath \"\"" Nov 28 13:37:52 crc kubenswrapper[4631]: I1128 13:37:52.987685 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cqbq9\" (UniqueName: \"kubernetes.io/projected/e35e9173-ceb1-4f6a-8a76-2a8c869aa26b-kube-api-access-cqbq9\") on node \"crc\" DevicePath \"\"" Nov 28 13:37:52 crc kubenswrapper[4631]: I1128 13:37:52.987702 4631 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ed809148-a421-492c-a7c1-6a9e4eba0528-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 28 13:37:52 crc kubenswrapper[4631]: I1128 13:37:52.987713 4631 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e35e9173-ceb1-4f6a-8a76-2a8c869aa26b-logs\") on node \"crc\" DevicePath \"\"" Nov 28 13:37:52 crc kubenswrapper[4631]: I1128 13:37:52.999264 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cec87f6b-b6fd-467c-adb9-f6c6ff099a18-logs" (OuterVolumeSpecName: "logs") pod "cec87f6b-b6fd-467c-adb9-f6c6ff099a18" (UID: "cec87f6b-b6fd-467c-adb9-f6c6ff099a18"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 13:37:53 crc kubenswrapper[4631]: I1128 13:37:53.003875 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cec87f6b-b6fd-467c-adb9-f6c6ff099a18-scripts" (OuterVolumeSpecName: "scripts") pod "cec87f6b-b6fd-467c-adb9-f6c6ff099a18" (UID: "cec87f6b-b6fd-467c-adb9-f6c6ff099a18"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 13:37:53 crc kubenswrapper[4631]: I1128 13:37:53.024593 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cec87f6b-b6fd-467c-adb9-f6c6ff099a18-config-data" (OuterVolumeSpecName: "config-data") pod "cec87f6b-b6fd-467c-adb9-f6c6ff099a18" (UID: "cec87f6b-b6fd-467c-adb9-f6c6ff099a18"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 13:37:53 crc kubenswrapper[4631]: I1128 13:37:53.026452 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cec87f6b-b6fd-467c-adb9-f6c6ff099a18-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "cec87f6b-b6fd-467c-adb9-f6c6ff099a18" (UID: "cec87f6b-b6fd-467c-adb9-f6c6ff099a18"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:37:53 crc kubenswrapper[4631]: I1128 13:37:53.031164 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cec87f6b-b6fd-467c-adb9-f6c6ff099a18-kube-api-access-j9kq8" (OuterVolumeSpecName: "kube-api-access-j9kq8") pod "cec87f6b-b6fd-467c-adb9-f6c6ff099a18" (UID: "cec87f6b-b6fd-467c-adb9-f6c6ff099a18"). InnerVolumeSpecName "kube-api-access-j9kq8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:37:53 crc kubenswrapper[4631]: I1128 13:37:53.093920 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d95d3536-1392-4576-8ee5-390c09fe7a49","Type":"ContainerStarted","Data":"42664776e24655262092eae45a6b6b94e6ed84e5b6614b7c0abe60d7e38df756"} Nov 28 13:37:53 crc kubenswrapper[4631]: I1128 13:37:53.107123 4631 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/cec87f6b-b6fd-467c-adb9-f6c6ff099a18-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Nov 28 13:37:53 crc kubenswrapper[4631]: I1128 13:37:53.107167 4631 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/cec87f6b-b6fd-467c-adb9-f6c6ff099a18-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 13:37:53 crc kubenswrapper[4631]: I1128 13:37:53.107183 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j9kq8\" (UniqueName: \"kubernetes.io/projected/cec87f6b-b6fd-467c-adb9-f6c6ff099a18-kube-api-access-j9kq8\") on node \"crc\" DevicePath \"\"" Nov 28 13:37:53 crc kubenswrapper[4631]: I1128 13:37:53.107200 4631 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/cec87f6b-b6fd-467c-adb9-f6c6ff099a18-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 13:37:53 crc kubenswrapper[4631]: I1128 13:37:53.107214 4631 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cec87f6b-b6fd-467c-adb9-f6c6ff099a18-logs\") on node \"crc\" DevicePath \"\"" Nov 28 13:37:53 crc kubenswrapper[4631]: I1128 13:37:53.149036 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 28 13:37:53 crc kubenswrapper[4631]: I1128 13:37:53.159689 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-l4h89"] Nov 28 13:37:53 crc kubenswrapper[4631]: E1128 13:37:53.160124 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e35e9173-ceb1-4f6a-8a76-2a8c869aa26b" containerName="glance-httpd" Nov 28 13:37:53 crc kubenswrapper[4631]: I1128 13:37:53.160143 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="e35e9173-ceb1-4f6a-8a76-2a8c869aa26b" containerName="glance-httpd" Nov 28 13:37:53 crc kubenswrapper[4631]: E1128 13:37:53.160156 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="166f0f43-e5cb-42ad-a2f5-1b207accc673" containerName="init" Nov 28 13:37:53 crc kubenswrapper[4631]: I1128 13:37:53.160163 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="166f0f43-e5cb-42ad-a2f5-1b207accc673" containerName="init" Nov 28 13:37:53 crc kubenswrapper[4631]: E1128 13:37:53.160171 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ed809148-a421-492c-a7c1-6a9e4eba0528" containerName="glance-httpd" Nov 28 13:37:53 crc kubenswrapper[4631]: I1128 13:37:53.160177 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="ed809148-a421-492c-a7c1-6a9e4eba0528" containerName="glance-httpd" Nov 28 13:37:53 crc kubenswrapper[4631]: E1128 13:37:53.160197 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ed809148-a421-492c-a7c1-6a9e4eba0528" containerName="glance-log" Nov 28 13:37:53 crc kubenswrapper[4631]: I1128 13:37:53.160203 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="ed809148-a421-492c-a7c1-6a9e4eba0528" containerName="glance-log" Nov 28 13:37:53 crc kubenswrapper[4631]: E1128 13:37:53.160217 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e35e9173-ceb1-4f6a-8a76-2a8c869aa26b" containerName="glance-log" Nov 28 13:37:53 crc kubenswrapper[4631]: I1128 13:37:53.160224 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="e35e9173-ceb1-4f6a-8a76-2a8c869aa26b" containerName="glance-log" Nov 28 13:37:53 crc kubenswrapper[4631]: E1128 13:37:53.160235 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="268feb7e-4677-4c47-b837-e04bab4b8010" containerName="neutron-db-sync" Nov 28 13:37:53 crc kubenswrapper[4631]: I1128 13:37:53.160241 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="268feb7e-4677-4c47-b837-e04bab4b8010" containerName="neutron-db-sync" Nov 28 13:37:53 crc kubenswrapper[4631]: E1128 13:37:53.160257 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="166f0f43-e5cb-42ad-a2f5-1b207accc673" containerName="dnsmasq-dns" Nov 28 13:37:53 crc kubenswrapper[4631]: I1128 13:37:53.160262 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="166f0f43-e5cb-42ad-a2f5-1b207accc673" containerName="dnsmasq-dns" Nov 28 13:37:53 crc kubenswrapper[4631]: I1128 13:37:53.160475 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="268feb7e-4677-4c47-b837-e04bab4b8010" containerName="neutron-db-sync" Nov 28 13:37:53 crc kubenswrapper[4631]: I1128 13:37:53.160488 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="e35e9173-ceb1-4f6a-8a76-2a8c869aa26b" containerName="glance-log" Nov 28 13:37:53 crc kubenswrapper[4631]: I1128 13:37:53.160501 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="ed809148-a421-492c-a7c1-6a9e4eba0528" containerName="glance-httpd" Nov 28 13:37:53 crc kubenswrapper[4631]: I1128 13:37:53.160516 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="e35e9173-ceb1-4f6a-8a76-2a8c869aa26b" containerName="glance-httpd" Nov 28 13:37:53 crc kubenswrapper[4631]: I1128 13:37:53.160525 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="ed809148-a421-492c-a7c1-6a9e4eba0528" containerName="glance-log" Nov 28 13:37:53 crc kubenswrapper[4631]: I1128 13:37:53.160536 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="166f0f43-e5cb-42ad-a2f5-1b207accc673" containerName="dnsmasq-dns" Nov 28 13:37:53 crc kubenswrapper[4631]: I1128 13:37:53.161196 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e35e9173-ceb1-4f6a-8a76-2a8c869aa26b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e35e9173-ceb1-4f6a-8a76-2a8c869aa26b" (UID: "e35e9173-ceb1-4f6a-8a76-2a8c869aa26b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:37:53 crc kubenswrapper[4631]: I1128 13:37:53.175803 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"e35e9173-ceb1-4f6a-8a76-2a8c869aa26b","Type":"ContainerDied","Data":"8bd92bf88a9c1ecb415236f8494315e9f848f24db7a89819912d8c8e52663575"} Nov 28 13:37:53 crc kubenswrapper[4631]: I1128 13:37:53.176301 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55f844cf75-l4h89" Nov 28 13:37:53 crc kubenswrapper[4631]: I1128 13:37:53.203395 4631 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage04-crc" (UniqueName: "kubernetes.io/local-volume/local-storage04-crc") on node "crc" Nov 28 13:37:53 crc kubenswrapper[4631]: I1128 13:37:53.204686 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" event={"ID":"8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1","Type":"ContainerStarted","Data":"ae892f616f5c89a5af75dda3b0799a9e1ba33e9ae4579727d53b425382bdbdc2"} Nov 28 13:37:53 crc kubenswrapper[4631]: I1128 13:37:53.227887 4631 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e35e9173-ceb1-4f6a-8a76-2a8c869aa26b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 13:37:53 crc kubenswrapper[4631]: I1128 13:37:53.228259 4631 reconciler_common.go:293] "Volume detached for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" DevicePath \"\"" Nov 28 13:37:53 crc kubenswrapper[4631]: I1128 13:37:53.228739 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-86566f7dbc-2zlk9" event={"ID":"cec87f6b-b6fd-467c-adb9-f6c6ff099a18","Type":"ContainerDied","Data":"54b60687a1b94197f0ec355b26757392cc2c90035dff06655d90153ac6ed7b7a"} Nov 28 13:37:53 crc kubenswrapper[4631]: I1128 13:37:53.229052 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-86566f7dbc-2zlk9" Nov 28 13:37:53 crc kubenswrapper[4631]: I1128 13:37:53.234785 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"ed809148-a421-492c-a7c1-6a9e4eba0528","Type":"ContainerDied","Data":"320c0c7e4cd1d12b6176cf7431fda56ba41ef73386caf009b93629928efe7cf7"} Nov 28 13:37:53 crc kubenswrapper[4631]: I1128 13:37:53.235104 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 28 13:37:53 crc kubenswrapper[4631]: I1128 13:37:53.313138 4631 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage08-crc" (UniqueName: "kubernetes.io/local-volume/local-storage08-crc") on node "crc" Nov 28 13:37:53 crc kubenswrapper[4631]: I1128 13:37:53.322066 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6d58995c-nzh9k" event={"ID":"c3bcfdda-21af-4504-806c-50e6bc0f8770","Type":"ContainerStarted","Data":"f15628a29d87012b07a748a7da86e986a7407dc5babc30b157d6e5e5ac4c694c"} Nov 28 13:37:53 crc kubenswrapper[4631]: I1128 13:37:53.330807 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/83f364c4-bdf6-48ba-a5dc-31bb4fbb6f66-ovsdbserver-sb\") pod \"dnsmasq-dns-55f844cf75-l4h89\" (UID: \"83f364c4-bdf6-48ba-a5dc-31bb4fbb6f66\") " pod="openstack/dnsmasq-dns-55f844cf75-l4h89" Nov 28 13:37:53 crc kubenswrapper[4631]: I1128 13:37:53.331269 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/83f364c4-bdf6-48ba-a5dc-31bb4fbb6f66-ovsdbserver-nb\") pod \"dnsmasq-dns-55f844cf75-l4h89\" (UID: \"83f364c4-bdf6-48ba-a5dc-31bb4fbb6f66\") " pod="openstack/dnsmasq-dns-55f844cf75-l4h89" Nov 28 13:37:53 crc kubenswrapper[4631]: I1128 13:37:53.331387 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fgjjx\" (UniqueName: \"kubernetes.io/projected/83f364c4-bdf6-48ba-a5dc-31bb4fbb6f66-kube-api-access-fgjjx\") pod \"dnsmasq-dns-55f844cf75-l4h89\" (UID: \"83f364c4-bdf6-48ba-a5dc-31bb4fbb6f66\") " pod="openstack/dnsmasq-dns-55f844cf75-l4h89" Nov 28 13:37:53 crc kubenswrapper[4631]: I1128 13:37:53.331509 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/83f364c4-bdf6-48ba-a5dc-31bb4fbb6f66-config\") pod \"dnsmasq-dns-55f844cf75-l4h89\" (UID: \"83f364c4-bdf6-48ba-a5dc-31bb4fbb6f66\") " pod="openstack/dnsmasq-dns-55f844cf75-l4h89" Nov 28 13:37:53 crc kubenswrapper[4631]: I1128 13:37:53.331622 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/83f364c4-bdf6-48ba-a5dc-31bb4fbb6f66-dns-svc\") pod \"dnsmasq-dns-55f844cf75-l4h89\" (UID: \"83f364c4-bdf6-48ba-a5dc-31bb4fbb6f66\") " pod="openstack/dnsmasq-dns-55f844cf75-l4h89" Nov 28 13:37:53 crc kubenswrapper[4631]: I1128 13:37:53.331731 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/83f364c4-bdf6-48ba-a5dc-31bb4fbb6f66-dns-swift-storage-0\") pod \"dnsmasq-dns-55f844cf75-l4h89\" (UID: \"83f364c4-bdf6-48ba-a5dc-31bb4fbb6f66\") " pod="openstack/dnsmasq-dns-55f844cf75-l4h89" Nov 28 13:37:53 crc kubenswrapper[4631]: I1128 13:37:53.331839 4631 reconciler_common.go:293] "Volume detached for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" DevicePath \"\"" Nov 28 13:37:53 crc kubenswrapper[4631]: I1128 13:37:53.338504 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-l4h89"] Nov 28 13:37:53 crc kubenswrapper[4631]: I1128 13:37:53.344065 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e35e9173-ceb1-4f6a-8a76-2a8c869aa26b-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "e35e9173-ceb1-4f6a-8a76-2a8c869aa26b" (UID: "e35e9173-ceb1-4f6a-8a76-2a8c869aa26b"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:37:53 crc kubenswrapper[4631]: E1128 13:37:53.347153 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified\\\"\"" pod="openstack/barbican-db-sync-z6tgb" podUID="3050013d-2db4-4801-8626-1fb6398b714e" Nov 28 13:37:53 crc kubenswrapper[4631]: I1128 13:37:53.347934 4631 scope.go:117] "RemoveContainer" containerID="9c22148dc08e7d0f7400463e44a7fe46239e46b30d3b118c3cac56d80d27c687" Nov 28 13:37:53 crc kubenswrapper[4631]: E1128 13:37:53.362812 4631 kuberuntime_gc.go:389] "Failed to remove container log dead symlink" err="remove /var/log/containers/glance-default-external-api-0_openstack_glance-httpd-9c22148dc08e7d0f7400463e44a7fe46239e46b30d3b118c3cac56d80d27c687.log: no such file or directory" path="/var/log/containers/glance-default-external-api-0_openstack_glance-httpd-9c22148dc08e7d0f7400463e44a7fe46239e46b30d3b118c3cac56d80d27c687.log" Nov 28 13:37:53 crc kubenswrapper[4631]: I1128 13:37:53.382081 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-544496c768-rzhsw"] Nov 28 13:37:53 crc kubenswrapper[4631]: I1128 13:37:53.397589 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ed809148-a421-492c-a7c1-6a9e4eba0528-config-data" (OuterVolumeSpecName: "config-data") pod "ed809148-a421-492c-a7c1-6a9e4eba0528" (UID: "ed809148-a421-492c-a7c1-6a9e4eba0528"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:37:53 crc kubenswrapper[4631]: I1128 13:37:53.413496 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ed809148-a421-492c-a7c1-6a9e4eba0528-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "ed809148-a421-492c-a7c1-6a9e4eba0528" (UID: "ed809148-a421-492c-a7c1-6a9e4eba0528"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:37:53 crc kubenswrapper[4631]: I1128 13:37:53.437783 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ed809148-a421-492c-a7c1-6a9e4eba0528-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ed809148-a421-492c-a7c1-6a9e4eba0528" (UID: "ed809148-a421-492c-a7c1-6a9e4eba0528"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:37:53 crc kubenswrapper[4631]: I1128 13:37:53.438545 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ed809148-a421-492c-a7c1-6a9e4eba0528-combined-ca-bundle\") pod \"ed809148-a421-492c-a7c1-6a9e4eba0528\" (UID: \"ed809148-a421-492c-a7c1-6a9e4eba0528\") " Nov 28 13:37:53 crc kubenswrapper[4631]: I1128 13:37:53.439012 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/83f364c4-bdf6-48ba-a5dc-31bb4fbb6f66-config\") pod \"dnsmasq-dns-55f844cf75-l4h89\" (UID: \"83f364c4-bdf6-48ba-a5dc-31bb4fbb6f66\") " pod="openstack/dnsmasq-dns-55f844cf75-l4h89" Nov 28 13:37:53 crc kubenswrapper[4631]: I1128 13:37:53.439084 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/83f364c4-bdf6-48ba-a5dc-31bb4fbb6f66-dns-svc\") pod \"dnsmasq-dns-55f844cf75-l4h89\" (UID: \"83f364c4-bdf6-48ba-a5dc-31bb4fbb6f66\") " pod="openstack/dnsmasq-dns-55f844cf75-l4h89" Nov 28 13:37:53 crc kubenswrapper[4631]: I1128 13:37:53.439158 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/83f364c4-bdf6-48ba-a5dc-31bb4fbb6f66-dns-swift-storage-0\") pod \"dnsmasq-dns-55f844cf75-l4h89\" (UID: \"83f364c4-bdf6-48ba-a5dc-31bb4fbb6f66\") " pod="openstack/dnsmasq-dns-55f844cf75-l4h89" Nov 28 13:37:53 crc kubenswrapper[4631]: I1128 13:37:53.439253 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/83f364c4-bdf6-48ba-a5dc-31bb4fbb6f66-ovsdbserver-sb\") pod \"dnsmasq-dns-55f844cf75-l4h89\" (UID: \"83f364c4-bdf6-48ba-a5dc-31bb4fbb6f66\") " pod="openstack/dnsmasq-dns-55f844cf75-l4h89" Nov 28 13:37:53 crc kubenswrapper[4631]: I1128 13:37:53.439329 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/83f364c4-bdf6-48ba-a5dc-31bb4fbb6f66-ovsdbserver-nb\") pod \"dnsmasq-dns-55f844cf75-l4h89\" (UID: \"83f364c4-bdf6-48ba-a5dc-31bb4fbb6f66\") " pod="openstack/dnsmasq-dns-55f844cf75-l4h89" Nov 28 13:37:53 crc kubenswrapper[4631]: I1128 13:37:53.439361 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fgjjx\" (UniqueName: \"kubernetes.io/projected/83f364c4-bdf6-48ba-a5dc-31bb4fbb6f66-kube-api-access-fgjjx\") pod \"dnsmasq-dns-55f844cf75-l4h89\" (UID: \"83f364c4-bdf6-48ba-a5dc-31bb4fbb6f66\") " pod="openstack/dnsmasq-dns-55f844cf75-l4h89" Nov 28 13:37:53 crc kubenswrapper[4631]: I1128 13:37:53.439438 4631 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ed809148-a421-492c-a7c1-6a9e4eba0528-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 13:37:53 crc kubenswrapper[4631]: I1128 13:37:53.439454 4631 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ed809148-a421-492c-a7c1-6a9e4eba0528-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 28 13:37:53 crc kubenswrapper[4631]: I1128 13:37:53.439466 4631 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e35e9173-ceb1-4f6a-8a76-2a8c869aa26b-public-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 28 13:37:53 crc kubenswrapper[4631]: W1128 13:37:53.439826 4631 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/ed809148-a421-492c-a7c1-6a9e4eba0528/volumes/kubernetes.io~secret/combined-ca-bundle Nov 28 13:37:53 crc kubenswrapper[4631]: I1128 13:37:53.439846 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ed809148-a421-492c-a7c1-6a9e4eba0528-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ed809148-a421-492c-a7c1-6a9e4eba0528" (UID: "ed809148-a421-492c-a7c1-6a9e4eba0528"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:37:53 crc kubenswrapper[4631]: I1128 13:37:53.441469 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/83f364c4-bdf6-48ba-a5dc-31bb4fbb6f66-dns-svc\") pod \"dnsmasq-dns-55f844cf75-l4h89\" (UID: \"83f364c4-bdf6-48ba-a5dc-31bb4fbb6f66\") " pod="openstack/dnsmasq-dns-55f844cf75-l4h89" Nov 28 13:37:53 crc kubenswrapper[4631]: I1128 13:37:53.442687 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/83f364c4-bdf6-48ba-a5dc-31bb4fbb6f66-dns-swift-storage-0\") pod \"dnsmasq-dns-55f844cf75-l4h89\" (UID: \"83f364c4-bdf6-48ba-a5dc-31bb4fbb6f66\") " pod="openstack/dnsmasq-dns-55f844cf75-l4h89" Nov 28 13:37:53 crc kubenswrapper[4631]: I1128 13:37:53.443743 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/83f364c4-bdf6-48ba-a5dc-31bb4fbb6f66-ovsdbserver-sb\") pod \"dnsmasq-dns-55f844cf75-l4h89\" (UID: \"83f364c4-bdf6-48ba-a5dc-31bb4fbb6f66\") " pod="openstack/dnsmasq-dns-55f844cf75-l4h89" Nov 28 13:37:53 crc kubenswrapper[4631]: I1128 13:37:53.444011 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/83f364c4-bdf6-48ba-a5dc-31bb4fbb6f66-ovsdbserver-nb\") pod \"dnsmasq-dns-55f844cf75-l4h89\" (UID: \"83f364c4-bdf6-48ba-a5dc-31bb4fbb6f66\") " pod="openstack/dnsmasq-dns-55f844cf75-l4h89" Nov 28 13:37:53 crc kubenswrapper[4631]: I1128 13:37:53.444515 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/83f364c4-bdf6-48ba-a5dc-31bb4fbb6f66-config\") pod \"dnsmasq-dns-55f844cf75-l4h89\" (UID: \"83f364c4-bdf6-48ba-a5dc-31bb4fbb6f66\") " pod="openstack/dnsmasq-dns-55f844cf75-l4h89" Nov 28 13:37:53 crc kubenswrapper[4631]: I1128 13:37:53.468258 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e35e9173-ceb1-4f6a-8a76-2a8c869aa26b-config-data" (OuterVolumeSpecName: "config-data") pod "e35e9173-ceb1-4f6a-8a76-2a8c869aa26b" (UID: "e35e9173-ceb1-4f6a-8a76-2a8c869aa26b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:37:53 crc kubenswrapper[4631]: I1128 13:37:53.472557 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fgjjx\" (UniqueName: \"kubernetes.io/projected/83f364c4-bdf6-48ba-a5dc-31bb4fbb6f66-kube-api-access-fgjjx\") pod \"dnsmasq-dns-55f844cf75-l4h89\" (UID: \"83f364c4-bdf6-48ba-a5dc-31bb4fbb6f66\") " pod="openstack/dnsmasq-dns-55f844cf75-l4h89" Nov 28 13:37:53 crc kubenswrapper[4631]: I1128 13:37:53.495611 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-5fb67d6644-dbmzf"] Nov 28 13:37:53 crc kubenswrapper[4631]: I1128 13:37:53.500609 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5fb67d6644-dbmzf" Nov 28 13:37:53 crc kubenswrapper[4631]: I1128 13:37:53.506867 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Nov 28 13:37:53 crc kubenswrapper[4631]: I1128 13:37:53.511213 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Nov 28 13:37:53 crc kubenswrapper[4631]: I1128 13:37:53.512042 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-cfjnw" Nov 28 13:37:53 crc kubenswrapper[4631]: I1128 13:37:53.512315 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Nov 28 13:37:53 crc kubenswrapper[4631]: I1128 13:37:53.523616 4631 scope.go:117] "RemoveContainer" containerID="d9680309b054cf63cc8abc1bb6e6a4bcd13b7ce334a2d1addf7497cb9c68398c" Nov 28 13:37:53 crc kubenswrapper[4631]: I1128 13:37:53.552191 4631 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e35e9173-ceb1-4f6a-8a76-2a8c869aa26b-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 13:37:53 crc kubenswrapper[4631]: I1128 13:37:53.552224 4631 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ed809148-a421-492c-a7c1-6a9e4eba0528-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 13:37:53 crc kubenswrapper[4631]: I1128 13:37:53.571604 4631 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7f02c913-7539-481a-abfc-7e19a28529e1" path="/var/lib/kubelet/pods/7f02c913-7539-481a-abfc-7e19a28529e1/volumes" Nov 28 13:37:53 crc kubenswrapper[4631]: I1128 13:37:53.584157 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-5fb67d6644-dbmzf"] Nov 28 13:37:53 crc kubenswrapper[4631]: I1128 13:37:53.593880 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-86566f7dbc-2zlk9"] Nov 28 13:37:53 crc kubenswrapper[4631]: I1128 13:37:53.623064 4631 scope.go:117] "RemoveContainer" containerID="9ae3553e198d5e5871f765c7eb232d5eb1cbd2d68f3a0039023bc82993806c9d" Nov 28 13:37:53 crc kubenswrapper[4631]: I1128 13:37:53.656841 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55f844cf75-l4h89" Nov 28 13:37:53 crc kubenswrapper[4631]: I1128 13:37:53.660139 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/6a0a46ba-e48f-45a6-801e-7cb0ececabee-ovndb-tls-certs\") pod \"neutron-5fb67d6644-dbmzf\" (UID: \"6a0a46ba-e48f-45a6-801e-7cb0ececabee\") " pod="openstack/neutron-5fb67d6644-dbmzf" Nov 28 13:37:53 crc kubenswrapper[4631]: I1128 13:37:53.660517 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a0a46ba-e48f-45a6-801e-7cb0ececabee-combined-ca-bundle\") pod \"neutron-5fb67d6644-dbmzf\" (UID: \"6a0a46ba-e48f-45a6-801e-7cb0ececabee\") " pod="openstack/neutron-5fb67d6644-dbmzf" Nov 28 13:37:53 crc kubenswrapper[4631]: I1128 13:37:53.660795 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/6a0a46ba-e48f-45a6-801e-7cb0ececabee-httpd-config\") pod \"neutron-5fb67d6644-dbmzf\" (UID: \"6a0a46ba-e48f-45a6-801e-7cb0ececabee\") " pod="openstack/neutron-5fb67d6644-dbmzf" Nov 28 13:37:53 crc kubenswrapper[4631]: I1128 13:37:53.660863 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/6a0a46ba-e48f-45a6-801e-7cb0ececabee-config\") pod \"neutron-5fb67d6644-dbmzf\" (UID: \"6a0a46ba-e48f-45a6-801e-7cb0ececabee\") " pod="openstack/neutron-5fb67d6644-dbmzf" Nov 28 13:37:53 crc kubenswrapper[4631]: I1128 13:37:53.660904 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cblkg\" (UniqueName: \"kubernetes.io/projected/6a0a46ba-e48f-45a6-801e-7cb0ececabee-kube-api-access-cblkg\") pod \"neutron-5fb67d6644-dbmzf\" (UID: \"6a0a46ba-e48f-45a6-801e-7cb0ececabee\") " pod="openstack/neutron-5fb67d6644-dbmzf" Nov 28 13:37:53 crc kubenswrapper[4631]: I1128 13:37:53.700492 4631 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-86566f7dbc-2zlk9"] Nov 28 13:37:53 crc kubenswrapper[4631]: I1128 13:37:53.762899 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/6a0a46ba-e48f-45a6-801e-7cb0ececabee-httpd-config\") pod \"neutron-5fb67d6644-dbmzf\" (UID: \"6a0a46ba-e48f-45a6-801e-7cb0ececabee\") " pod="openstack/neutron-5fb67d6644-dbmzf" Nov 28 13:37:53 crc kubenswrapper[4631]: I1128 13:37:53.762951 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/6a0a46ba-e48f-45a6-801e-7cb0ececabee-config\") pod \"neutron-5fb67d6644-dbmzf\" (UID: \"6a0a46ba-e48f-45a6-801e-7cb0ececabee\") " pod="openstack/neutron-5fb67d6644-dbmzf" Nov 28 13:37:53 crc kubenswrapper[4631]: I1128 13:37:53.762974 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cblkg\" (UniqueName: \"kubernetes.io/projected/6a0a46ba-e48f-45a6-801e-7cb0ececabee-kube-api-access-cblkg\") pod \"neutron-5fb67d6644-dbmzf\" (UID: \"6a0a46ba-e48f-45a6-801e-7cb0ececabee\") " pod="openstack/neutron-5fb67d6644-dbmzf" Nov 28 13:37:53 crc kubenswrapper[4631]: I1128 13:37:53.763035 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/6a0a46ba-e48f-45a6-801e-7cb0ececabee-ovndb-tls-certs\") pod \"neutron-5fb67d6644-dbmzf\" (UID: \"6a0a46ba-e48f-45a6-801e-7cb0ececabee\") " pod="openstack/neutron-5fb67d6644-dbmzf" Nov 28 13:37:53 crc kubenswrapper[4631]: I1128 13:37:53.763086 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a0a46ba-e48f-45a6-801e-7cb0ececabee-combined-ca-bundle\") pod \"neutron-5fb67d6644-dbmzf\" (UID: \"6a0a46ba-e48f-45a6-801e-7cb0ececabee\") " pod="openstack/neutron-5fb67d6644-dbmzf" Nov 28 13:37:53 crc kubenswrapper[4631]: I1128 13:37:53.773191 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/6a0a46ba-e48f-45a6-801e-7cb0ececabee-config\") pod \"neutron-5fb67d6644-dbmzf\" (UID: \"6a0a46ba-e48f-45a6-801e-7cb0ececabee\") " pod="openstack/neutron-5fb67d6644-dbmzf" Nov 28 13:37:53 crc kubenswrapper[4631]: I1128 13:37:53.787676 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/6a0a46ba-e48f-45a6-801e-7cb0ececabee-ovndb-tls-certs\") pod \"neutron-5fb67d6644-dbmzf\" (UID: \"6a0a46ba-e48f-45a6-801e-7cb0ececabee\") " pod="openstack/neutron-5fb67d6644-dbmzf" Nov 28 13:37:53 crc kubenswrapper[4631]: I1128 13:37:53.788365 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a0a46ba-e48f-45a6-801e-7cb0ececabee-combined-ca-bundle\") pod \"neutron-5fb67d6644-dbmzf\" (UID: \"6a0a46ba-e48f-45a6-801e-7cb0ececabee\") " pod="openstack/neutron-5fb67d6644-dbmzf" Nov 28 13:37:53 crc kubenswrapper[4631]: I1128 13:37:53.789309 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/6a0a46ba-e48f-45a6-801e-7cb0ececabee-httpd-config\") pod \"neutron-5fb67d6644-dbmzf\" (UID: \"6a0a46ba-e48f-45a6-801e-7cb0ececabee\") " pod="openstack/neutron-5fb67d6644-dbmzf" Nov 28 13:37:53 crc kubenswrapper[4631]: I1128 13:37:53.825066 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cblkg\" (UniqueName: \"kubernetes.io/projected/6a0a46ba-e48f-45a6-801e-7cb0ececabee-kube-api-access-cblkg\") pod \"neutron-5fb67d6644-dbmzf\" (UID: \"6a0a46ba-e48f-45a6-801e-7cb0ececabee\") " pod="openstack/neutron-5fb67d6644-dbmzf" Nov 28 13:37:53 crc kubenswrapper[4631]: I1128 13:37:53.834470 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-dcm9x"] Nov 28 13:37:53 crc kubenswrapper[4631]: I1128 13:37:53.847279 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5fb67d6644-dbmzf" Nov 28 13:37:53 crc kubenswrapper[4631]: I1128 13:37:53.858924 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-64f8fdf97b-jn5mg"] Nov 28 13:37:53 crc kubenswrapper[4631]: I1128 13:37:53.878441 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 28 13:37:53 crc kubenswrapper[4631]: I1128 13:37:53.925982 4631 scope.go:117] "RemoveContainer" containerID="11e888d5e22b844bbb2536deb555c24c7a734fcdc9caae547b186690ef705ed9" Nov 28 13:37:53 crc kubenswrapper[4631]: I1128 13:37:53.967078 4631 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 28 13:37:54 crc kubenswrapper[4631]: I1128 13:37:54.013966 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 28 13:37:54 crc kubenswrapper[4631]: I1128 13:37:54.051778 4631 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 28 13:37:54 crc kubenswrapper[4631]: I1128 13:37:54.116627 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 28 13:37:54 crc kubenswrapper[4631]: I1128 13:37:54.118442 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 28 13:37:54 crc kubenswrapper[4631]: I1128 13:37:54.134499 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Nov 28 13:37:54 crc kubenswrapper[4631]: I1128 13:37:54.135338 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Nov 28 13:37:54 crc kubenswrapper[4631]: I1128 13:37:54.135483 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-f22sn" Nov 28 13:37:54 crc kubenswrapper[4631]: I1128 13:37:54.143784 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Nov 28 13:37:54 crc kubenswrapper[4631]: I1128 13:37:54.145140 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 28 13:37:54 crc kubenswrapper[4631]: I1128 13:37:54.184065 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Nov 28 13:37:54 crc kubenswrapper[4631]: I1128 13:37:54.185696 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 28 13:37:54 crc kubenswrapper[4631]: I1128 13:37:54.188862 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Nov 28 13:37:54 crc kubenswrapper[4631]: I1128 13:37:54.189037 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Nov 28 13:37:54 crc kubenswrapper[4631]: I1128 13:37:54.190426 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cb966e50-d246-48dc-b4a6-a9de4dbb8569-logs\") pod \"glance-default-internal-api-0\" (UID: \"cb966e50-d246-48dc-b4a6-a9de4dbb8569\") " pod="openstack/glance-default-internal-api-0" Nov 28 13:37:54 crc kubenswrapper[4631]: I1128 13:37:54.190486 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cb966e50-d246-48dc-b4a6-a9de4dbb8569-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"cb966e50-d246-48dc-b4a6-a9de4dbb8569\") " pod="openstack/glance-default-internal-api-0" Nov 28 13:37:54 crc kubenswrapper[4631]: I1128 13:37:54.190535 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/cb966e50-d246-48dc-b4a6-a9de4dbb8569-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"cb966e50-d246-48dc-b4a6-a9de4dbb8569\") " pod="openstack/glance-default-internal-api-0" Nov 28 13:37:54 crc kubenswrapper[4631]: I1128 13:37:54.190568 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jksgc\" (UniqueName: \"kubernetes.io/projected/cb966e50-d246-48dc-b4a6-a9de4dbb8569-kube-api-access-jksgc\") pod \"glance-default-internal-api-0\" (UID: \"cb966e50-d246-48dc-b4a6-a9de4dbb8569\") " pod="openstack/glance-default-internal-api-0" Nov 28 13:37:54 crc kubenswrapper[4631]: I1128 13:37:54.190589 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/cb966e50-d246-48dc-b4a6-a9de4dbb8569-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"cb966e50-d246-48dc-b4a6-a9de4dbb8569\") " pod="openstack/glance-default-internal-api-0" Nov 28 13:37:54 crc kubenswrapper[4631]: I1128 13:37:54.190631 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cb966e50-d246-48dc-b4a6-a9de4dbb8569-config-data\") pod \"glance-default-internal-api-0\" (UID: \"cb966e50-d246-48dc-b4a6-a9de4dbb8569\") " pod="openstack/glance-default-internal-api-0" Nov 28 13:37:54 crc kubenswrapper[4631]: I1128 13:37:54.190667 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cb966e50-d246-48dc-b4a6-a9de4dbb8569-scripts\") pod \"glance-default-internal-api-0\" (UID: \"cb966e50-d246-48dc-b4a6-a9de4dbb8569\") " pod="openstack/glance-default-internal-api-0" Nov 28 13:37:54 crc kubenswrapper[4631]: I1128 13:37:54.190702 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-internal-api-0\" (UID: \"cb966e50-d246-48dc-b4a6-a9de4dbb8569\") " pod="openstack/glance-default-internal-api-0" Nov 28 13:37:54 crc kubenswrapper[4631]: I1128 13:37:54.192921 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 28 13:37:54 crc kubenswrapper[4631]: I1128 13:37:54.272079 4631 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-74f6bcbc87-r8g9l" podUID="166f0f43-e5cb-42ad-a2f5-1b207accc673" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.132:5353: i/o timeout" Nov 28 13:37:54 crc kubenswrapper[4631]: I1128 13:37:54.292680 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e1d15959-31a9-464e-9e70-57547ddad347-scripts\") pod \"glance-default-external-api-0\" (UID: \"e1d15959-31a9-464e-9e70-57547ddad347\") " pod="openstack/glance-default-external-api-0" Nov 28 13:37:54 crc kubenswrapper[4631]: I1128 13:37:54.292760 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e1d15959-31a9-464e-9e70-57547ddad347-logs\") pod \"glance-default-external-api-0\" (UID: \"e1d15959-31a9-464e-9e70-57547ddad347\") " pod="openstack/glance-default-external-api-0" Nov 28 13:37:54 crc kubenswrapper[4631]: I1128 13:37:54.292793 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cb966e50-d246-48dc-b4a6-a9de4dbb8569-logs\") pod \"glance-default-internal-api-0\" (UID: \"cb966e50-d246-48dc-b4a6-a9de4dbb8569\") " pod="openstack/glance-default-internal-api-0" Nov 28 13:37:54 crc kubenswrapper[4631]: I1128 13:37:54.292836 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cb966e50-d246-48dc-b4a6-a9de4dbb8569-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"cb966e50-d246-48dc-b4a6-a9de4dbb8569\") " pod="openstack/glance-default-internal-api-0" Nov 28 13:37:54 crc kubenswrapper[4631]: I1128 13:37:54.292871 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-external-api-0\" (UID: \"e1d15959-31a9-464e-9e70-57547ddad347\") " pod="openstack/glance-default-external-api-0" Nov 28 13:37:54 crc kubenswrapper[4631]: I1128 13:37:54.292908 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e1d15959-31a9-464e-9e70-57547ddad347-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"e1d15959-31a9-464e-9e70-57547ddad347\") " pod="openstack/glance-default-external-api-0" Nov 28 13:37:54 crc kubenswrapper[4631]: I1128 13:37:54.292935 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kpkmg\" (UniqueName: \"kubernetes.io/projected/e1d15959-31a9-464e-9e70-57547ddad347-kube-api-access-kpkmg\") pod \"glance-default-external-api-0\" (UID: \"e1d15959-31a9-464e-9e70-57547ddad347\") " pod="openstack/glance-default-external-api-0" Nov 28 13:37:54 crc kubenswrapper[4631]: I1128 13:37:54.292979 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e1d15959-31a9-464e-9e70-57547ddad347-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"e1d15959-31a9-464e-9e70-57547ddad347\") " pod="openstack/glance-default-external-api-0" Nov 28 13:37:54 crc kubenswrapper[4631]: I1128 13:37:54.293016 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/cb966e50-d246-48dc-b4a6-a9de4dbb8569-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"cb966e50-d246-48dc-b4a6-a9de4dbb8569\") " pod="openstack/glance-default-internal-api-0" Nov 28 13:37:54 crc kubenswrapper[4631]: I1128 13:37:54.293055 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jksgc\" (UniqueName: \"kubernetes.io/projected/cb966e50-d246-48dc-b4a6-a9de4dbb8569-kube-api-access-jksgc\") pod \"glance-default-internal-api-0\" (UID: \"cb966e50-d246-48dc-b4a6-a9de4dbb8569\") " pod="openstack/glance-default-internal-api-0" Nov 28 13:37:54 crc kubenswrapper[4631]: I1128 13:37:54.293081 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/cb966e50-d246-48dc-b4a6-a9de4dbb8569-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"cb966e50-d246-48dc-b4a6-a9de4dbb8569\") " pod="openstack/glance-default-internal-api-0" Nov 28 13:37:54 crc kubenswrapper[4631]: I1128 13:37:54.293104 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e1d15959-31a9-464e-9e70-57547ddad347-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"e1d15959-31a9-464e-9e70-57547ddad347\") " pod="openstack/glance-default-external-api-0" Nov 28 13:37:54 crc kubenswrapper[4631]: I1128 13:37:54.293151 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e1d15959-31a9-464e-9e70-57547ddad347-config-data\") pod \"glance-default-external-api-0\" (UID: \"e1d15959-31a9-464e-9e70-57547ddad347\") " pod="openstack/glance-default-external-api-0" Nov 28 13:37:54 crc kubenswrapper[4631]: I1128 13:37:54.293237 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cb966e50-d246-48dc-b4a6-a9de4dbb8569-config-data\") pod \"glance-default-internal-api-0\" (UID: \"cb966e50-d246-48dc-b4a6-a9de4dbb8569\") " pod="openstack/glance-default-internal-api-0" Nov 28 13:37:54 crc kubenswrapper[4631]: I1128 13:37:54.293270 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cb966e50-d246-48dc-b4a6-a9de4dbb8569-scripts\") pod \"glance-default-internal-api-0\" (UID: \"cb966e50-d246-48dc-b4a6-a9de4dbb8569\") " pod="openstack/glance-default-internal-api-0" Nov 28 13:37:54 crc kubenswrapper[4631]: I1128 13:37:54.293371 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-internal-api-0\" (UID: \"cb966e50-d246-48dc-b4a6-a9de4dbb8569\") " pod="openstack/glance-default-internal-api-0" Nov 28 13:37:54 crc kubenswrapper[4631]: I1128 13:37:54.293966 4631 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-internal-api-0\" (UID: \"cb966e50-d246-48dc-b4a6-a9de4dbb8569\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/glance-default-internal-api-0" Nov 28 13:37:54 crc kubenswrapper[4631]: I1128 13:37:54.294735 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cb966e50-d246-48dc-b4a6-a9de4dbb8569-logs\") pod \"glance-default-internal-api-0\" (UID: \"cb966e50-d246-48dc-b4a6-a9de4dbb8569\") " pod="openstack/glance-default-internal-api-0" Nov 28 13:37:54 crc kubenswrapper[4631]: I1128 13:37:54.301591 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/cb966e50-d246-48dc-b4a6-a9de4dbb8569-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"cb966e50-d246-48dc-b4a6-a9de4dbb8569\") " pod="openstack/glance-default-internal-api-0" Nov 28 13:37:54 crc kubenswrapper[4631]: I1128 13:37:54.312890 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cb966e50-d246-48dc-b4a6-a9de4dbb8569-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"cb966e50-d246-48dc-b4a6-a9de4dbb8569\") " pod="openstack/glance-default-internal-api-0" Nov 28 13:37:54 crc kubenswrapper[4631]: I1128 13:37:54.313159 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cb966e50-d246-48dc-b4a6-a9de4dbb8569-config-data\") pod \"glance-default-internal-api-0\" (UID: \"cb966e50-d246-48dc-b4a6-a9de4dbb8569\") " pod="openstack/glance-default-internal-api-0" Nov 28 13:37:54 crc kubenswrapper[4631]: I1128 13:37:54.324992 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/cb966e50-d246-48dc-b4a6-a9de4dbb8569-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"cb966e50-d246-48dc-b4a6-a9de4dbb8569\") " pod="openstack/glance-default-internal-api-0" Nov 28 13:37:54 crc kubenswrapper[4631]: I1128 13:37:54.345726 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cb966e50-d246-48dc-b4a6-a9de4dbb8569-scripts\") pod \"glance-default-internal-api-0\" (UID: \"cb966e50-d246-48dc-b4a6-a9de4dbb8569\") " pod="openstack/glance-default-internal-api-0" Nov 28 13:37:54 crc kubenswrapper[4631]: I1128 13:37:54.360347 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-64f8fdf97b-jn5mg" event={"ID":"111c5349-528d-4856-a3e4-1277e3b10889","Type":"ContainerStarted","Data":"23b889f2631cf8acb4cb5fb4d20d550abb6dce9afa18c85b7c3660b3d4a38d3c"} Nov 28 13:37:54 crc kubenswrapper[4631]: I1128 13:37:54.377281 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jksgc\" (UniqueName: \"kubernetes.io/projected/cb966e50-d246-48dc-b4a6-a9de4dbb8569-kube-api-access-jksgc\") pod \"glance-default-internal-api-0\" (UID: \"cb966e50-d246-48dc-b4a6-a9de4dbb8569\") " pod="openstack/glance-default-internal-api-0" Nov 28 13:37:54 crc kubenswrapper[4631]: I1128 13:37:54.379242 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-dcm9x" event={"ID":"33e9838e-f616-4504-90d5-799de265be04","Type":"ContainerStarted","Data":"6476af066dead47f55a2df29345d1427b1180fd8cba959fa66e7109c5684f5c3"} Nov 28 13:37:54 crc kubenswrapper[4631]: I1128 13:37:54.403472 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e1d15959-31a9-464e-9e70-57547ddad347-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"e1d15959-31a9-464e-9e70-57547ddad347\") " pod="openstack/glance-default-external-api-0" Nov 28 13:37:54 crc kubenswrapper[4631]: I1128 13:37:54.403970 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e1d15959-31a9-464e-9e70-57547ddad347-config-data\") pod \"glance-default-external-api-0\" (UID: \"e1d15959-31a9-464e-9e70-57547ddad347\") " pod="openstack/glance-default-external-api-0" Nov 28 13:37:54 crc kubenswrapper[4631]: I1128 13:37:54.404102 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e1d15959-31a9-464e-9e70-57547ddad347-scripts\") pod \"glance-default-external-api-0\" (UID: \"e1d15959-31a9-464e-9e70-57547ddad347\") " pod="openstack/glance-default-external-api-0" Nov 28 13:37:54 crc kubenswrapper[4631]: I1128 13:37:54.404163 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e1d15959-31a9-464e-9e70-57547ddad347-logs\") pod \"glance-default-external-api-0\" (UID: \"e1d15959-31a9-464e-9e70-57547ddad347\") " pod="openstack/glance-default-external-api-0" Nov 28 13:37:54 crc kubenswrapper[4631]: I1128 13:37:54.404233 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-external-api-0\" (UID: \"e1d15959-31a9-464e-9e70-57547ddad347\") " pod="openstack/glance-default-external-api-0" Nov 28 13:37:54 crc kubenswrapper[4631]: I1128 13:37:54.404264 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kpkmg\" (UniqueName: \"kubernetes.io/projected/e1d15959-31a9-464e-9e70-57547ddad347-kube-api-access-kpkmg\") pod \"glance-default-external-api-0\" (UID: \"e1d15959-31a9-464e-9e70-57547ddad347\") " pod="openstack/glance-default-external-api-0" Nov 28 13:37:54 crc kubenswrapper[4631]: I1128 13:37:54.404332 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e1d15959-31a9-464e-9e70-57547ddad347-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"e1d15959-31a9-464e-9e70-57547ddad347\") " pod="openstack/glance-default-external-api-0" Nov 28 13:37:54 crc kubenswrapper[4631]: I1128 13:37:54.404359 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e1d15959-31a9-464e-9e70-57547ddad347-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"e1d15959-31a9-464e-9e70-57547ddad347\") " pod="openstack/glance-default-external-api-0" Nov 28 13:37:54 crc kubenswrapper[4631]: I1128 13:37:54.406051 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e1d15959-31a9-464e-9e70-57547ddad347-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"e1d15959-31a9-464e-9e70-57547ddad347\") " pod="openstack/glance-default-external-api-0" Nov 28 13:37:54 crc kubenswrapper[4631]: I1128 13:37:54.408792 4631 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-external-api-0\" (UID: \"e1d15959-31a9-464e-9e70-57547ddad347\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/glance-default-external-api-0" Nov 28 13:37:54 crc kubenswrapper[4631]: I1128 13:37:54.412235 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e1d15959-31a9-464e-9e70-57547ddad347-logs\") pod \"glance-default-external-api-0\" (UID: \"e1d15959-31a9-464e-9e70-57547ddad347\") " pod="openstack/glance-default-external-api-0" Nov 28 13:37:54 crc kubenswrapper[4631]: I1128 13:37:54.421126 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e1d15959-31a9-464e-9e70-57547ddad347-config-data\") pod \"glance-default-external-api-0\" (UID: \"e1d15959-31a9-464e-9e70-57547ddad347\") " pod="openstack/glance-default-external-api-0" Nov 28 13:37:54 crc kubenswrapper[4631]: I1128 13:37:54.435514 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e1d15959-31a9-464e-9e70-57547ddad347-scripts\") pod \"glance-default-external-api-0\" (UID: \"e1d15959-31a9-464e-9e70-57547ddad347\") " pod="openstack/glance-default-external-api-0" Nov 28 13:37:54 crc kubenswrapper[4631]: I1128 13:37:54.447376 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e1d15959-31a9-464e-9e70-57547ddad347-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"e1d15959-31a9-464e-9e70-57547ddad347\") " pod="openstack/glance-default-external-api-0" Nov 28 13:37:54 crc kubenswrapper[4631]: I1128 13:37:54.466232 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e1d15959-31a9-464e-9e70-57547ddad347-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"e1d15959-31a9-464e-9e70-57547ddad347\") " pod="openstack/glance-default-external-api-0" Nov 28 13:37:54 crc kubenswrapper[4631]: I1128 13:37:54.470418 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6d58995c-nzh9k" event={"ID":"c3bcfdda-21af-4504-806c-50e6bc0f8770","Type":"ContainerStarted","Data":"1ac446f19faea84032fc117fe1a86106333262fc6b9f72897a52070eea7a431c"} Nov 28 13:37:54 crc kubenswrapper[4631]: I1128 13:37:54.470630 4631 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-6d58995c-nzh9k" podUID="c3bcfdda-21af-4504-806c-50e6bc0f8770" containerName="horizon-log" containerID="cri-o://f15628a29d87012b07a748a7da86e986a7407dc5babc30b157d6e5e5ac4c694c" gracePeriod=30 Nov 28 13:37:54 crc kubenswrapper[4631]: I1128 13:37:54.470914 4631 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-6d58995c-nzh9k" podUID="c3bcfdda-21af-4504-806c-50e6bc0f8770" containerName="horizon" containerID="cri-o://1ac446f19faea84032fc117fe1a86106333262fc6b9f72897a52070eea7a431c" gracePeriod=30 Nov 28 13:37:54 crc kubenswrapper[4631]: I1128 13:37:54.555683 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-544496c768-rzhsw" event={"ID":"cd607dae-568a-4b81-af81-3310c0e95854","Type":"ContainerStarted","Data":"113aeaf934513c5f6d5f7a2c2cdcc5673675d86f6310ae3ebe9f349155fe85b9"} Nov 28 13:37:54 crc kubenswrapper[4631]: I1128 13:37:54.571071 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kpkmg\" (UniqueName: \"kubernetes.io/projected/e1d15959-31a9-464e-9e70-57547ddad347-kube-api-access-kpkmg\") pod \"glance-default-external-api-0\" (UID: \"e1d15959-31a9-464e-9e70-57547ddad347\") " pod="openstack/glance-default-external-api-0" Nov 28 13:37:54 crc kubenswrapper[4631]: I1128 13:37:54.588504 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-6d58995c-nzh9k" podStartSLOduration=4.791060205 podStartE2EDuration="40.588483636s" podCreationTimestamp="2025-11-28 13:37:14 +0000 UTC" firstStartedPulling="2025-11-28 13:37:16.615175037 +0000 UTC m=+993.422478381" lastFinishedPulling="2025-11-28 13:37:52.412598468 +0000 UTC m=+1029.219901812" observedRunningTime="2025-11-28 13:37:54.512922059 +0000 UTC m=+1031.320225403" watchObservedRunningTime="2025-11-28 13:37:54.588483636 +0000 UTC m=+1031.395786980" Nov 28 13:37:54 crc kubenswrapper[4631]: I1128 13:37:54.592035 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-external-api-0\" (UID: \"e1d15959-31a9-464e-9e70-57547ddad347\") " pod="openstack/glance-default-external-api-0" Nov 28 13:37:54 crc kubenswrapper[4631]: I1128 13:37:54.604780 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-internal-api-0\" (UID: \"cb966e50-d246-48dc-b4a6-a9de4dbb8569\") " pod="openstack/glance-default-internal-api-0" Nov 28 13:37:54 crc kubenswrapper[4631]: I1128 13:37:54.643505 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-l4h89"] Nov 28 13:37:54 crc kubenswrapper[4631]: W1128 13:37:54.672429 4631 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod83f364c4_bdf6_48ba_a5dc_31bb4fbb6f66.slice/crio-794b95d779176e79e82c45e7d2b377761a40a354cfd5048bb8e41b6c642b84d7 WatchSource:0}: Error finding container 794b95d779176e79e82c45e7d2b377761a40a354cfd5048bb8e41b6c642b84d7: Status 404 returned error can't find the container with id 794b95d779176e79e82c45e7d2b377761a40a354cfd5048bb8e41b6c642b84d7 Nov 28 13:37:54 crc kubenswrapper[4631]: I1128 13:37:54.792933 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 28 13:37:54 crc kubenswrapper[4631]: I1128 13:37:54.852906 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 28 13:37:54 crc kubenswrapper[4631]: I1128 13:37:54.867421 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-6d58995c-nzh9k" Nov 28 13:37:55 crc kubenswrapper[4631]: I1128 13:37:55.133543 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-5fb67d6644-dbmzf"] Nov 28 13:37:55 crc kubenswrapper[4631]: I1128 13:37:55.564540 4631 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cec87f6b-b6fd-467c-adb9-f6c6ff099a18" path="/var/lib/kubelet/pods/cec87f6b-b6fd-467c-adb9-f6c6ff099a18/volumes" Nov 28 13:37:55 crc kubenswrapper[4631]: I1128 13:37:55.568013 4631 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e35e9173-ceb1-4f6a-8a76-2a8c869aa26b" path="/var/lib/kubelet/pods/e35e9173-ceb1-4f6a-8a76-2a8c869aa26b/volumes" Nov 28 13:37:55 crc kubenswrapper[4631]: I1128 13:37:55.569233 4631 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ed809148-a421-492c-a7c1-6a9e4eba0528" path="/var/lib/kubelet/pods/ed809148-a421-492c-a7c1-6a9e4eba0528/volumes" Nov 28 13:37:55 crc kubenswrapper[4631]: I1128 13:37:55.618110 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55f844cf75-l4h89" event={"ID":"83f364c4-bdf6-48ba-a5dc-31bb4fbb6f66","Type":"ContainerStarted","Data":"794b95d779176e79e82c45e7d2b377761a40a354cfd5048bb8e41b6c642b84d7"} Nov 28 13:37:55 crc kubenswrapper[4631]: I1128 13:37:55.643715 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-544496c768-rzhsw" event={"ID":"cd607dae-568a-4b81-af81-3310c0e95854","Type":"ContainerStarted","Data":"eae1edde3af309c81483dc654c5de8326586db63504262cb5a4a2ea645c96988"} Nov 28 13:37:55 crc kubenswrapper[4631]: I1128 13:37:55.665132 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5fb67d6644-dbmzf" event={"ID":"6a0a46ba-e48f-45a6-801e-7cb0ececabee","Type":"ContainerStarted","Data":"630b1d6a6a9df29e837276658f20fa9c41ae2a7730d5d0daf69e9bd2a2b1eb75"} Nov 28 13:37:55 crc kubenswrapper[4631]: I1128 13:37:55.680227 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-dcm9x" event={"ID":"33e9838e-f616-4504-90d5-799de265be04","Type":"ContainerStarted","Data":"4cb68320336ce8ebb3c54401b0bf63ad92137da24f81bdcca4fff5bcff77299f"} Nov 28 13:37:55 crc kubenswrapper[4631]: I1128 13:37:55.683993 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-64f8fdf97b-jn5mg" event={"ID":"111c5349-528d-4856-a3e4-1277e3b10889","Type":"ContainerStarted","Data":"7bec69392b7f0e2c28b3311f3ec81e5bb4480735e00154f9e67c85a33543bc29"} Nov 28 13:37:55 crc kubenswrapper[4631]: I1128 13:37:55.718672 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-dcm9x" podStartSLOduration=8.718649427 podStartE2EDuration="8.718649427s" podCreationTimestamp="2025-11-28 13:37:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 13:37:55.71388193 +0000 UTC m=+1032.521185294" watchObservedRunningTime="2025-11-28 13:37:55.718649427 +0000 UTC m=+1032.525952771" Nov 28 13:37:55 crc kubenswrapper[4631]: I1128 13:37:55.935508 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 28 13:37:56 crc kubenswrapper[4631]: I1128 13:37:56.239519 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 28 13:37:56 crc kubenswrapper[4631]: I1128 13:37:56.757872 4631 generic.go:334] "Generic (PLEG): container finished" podID="83f364c4-bdf6-48ba-a5dc-31bb4fbb6f66" containerID="2f977a596e6188962f616acbe2f9777643926d465cba41a2749970823e0c0dd1" exitCode=0 Nov 28 13:37:56 crc kubenswrapper[4631]: I1128 13:37:56.757997 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55f844cf75-l4h89" event={"ID":"83f364c4-bdf6-48ba-a5dc-31bb4fbb6f66","Type":"ContainerDied","Data":"2f977a596e6188962f616acbe2f9777643926d465cba41a2749970823e0c0dd1"} Nov 28 13:37:56 crc kubenswrapper[4631]: I1128 13:37:56.776104 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-544496c768-rzhsw" event={"ID":"cd607dae-568a-4b81-af81-3310c0e95854","Type":"ContainerStarted","Data":"bd4617d23aba6bbc47167ded3acd30280726cfe208e8bfc920d75fb684ec48e8"} Nov 28 13:37:56 crc kubenswrapper[4631]: I1128 13:37:56.778363 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5fb67d6644-dbmzf" event={"ID":"6a0a46ba-e48f-45a6-801e-7cb0ececabee","Type":"ContainerStarted","Data":"821a3f2847ee539065f14edad868b3980fc7ccbc996dbb6dbf3a771daf646f34"} Nov 28 13:37:56 crc kubenswrapper[4631]: I1128 13:37:56.779054 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"cb966e50-d246-48dc-b4a6-a9de4dbb8569","Type":"ContainerStarted","Data":"598515cabf25ca4a2a23bcef4ffe5fa19863409e4a4da39027626c511adc8c82"} Nov 28 13:37:56 crc kubenswrapper[4631]: I1128 13:37:56.779773 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"e1d15959-31a9-464e-9e70-57547ddad347","Type":"ContainerStarted","Data":"889aaf4a8a635294c7495304ce6462870ca87110eefe5667edfc7352a064dc36"} Nov 28 13:37:56 crc kubenswrapper[4631]: I1128 13:37:56.784354 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-64f8fdf97b-jn5mg" event={"ID":"111c5349-528d-4856-a3e4-1277e3b10889","Type":"ContainerStarted","Data":"ffafadc828320f83436d72910a3912cda7f6bb048d8c4c01528f4529bf55b772"} Nov 28 13:37:56 crc kubenswrapper[4631]: I1128 13:37:56.849845 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-544496c768-rzhsw" podStartSLOduration=31.849824263 podStartE2EDuration="31.849824263s" podCreationTimestamp="2025-11-28 13:37:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 13:37:56.84196494 +0000 UTC m=+1033.649268284" watchObservedRunningTime="2025-11-28 13:37:56.849824263 +0000 UTC m=+1033.657127607" Nov 28 13:37:56 crc kubenswrapper[4631]: I1128 13:37:56.890910 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-64f8fdf97b-jn5mg" podStartSLOduration=32.890890213 podStartE2EDuration="32.890890213s" podCreationTimestamp="2025-11-28 13:37:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 13:37:56.890566564 +0000 UTC m=+1033.697869908" watchObservedRunningTime="2025-11-28 13:37:56.890890213 +0000 UTC m=+1033.698193557" Nov 28 13:37:57 crc kubenswrapper[4631]: I1128 13:37:57.186363 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-7f7bb9bf8f-42ssw"] Nov 28 13:37:57 crc kubenswrapper[4631]: I1128 13:37:57.188015 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-7f7bb9bf8f-42ssw" Nov 28 13:37:57 crc kubenswrapper[4631]: I1128 13:37:57.193661 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Nov 28 13:37:57 crc kubenswrapper[4631]: I1128 13:37:57.194097 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Nov 28 13:37:57 crc kubenswrapper[4631]: I1128 13:37:57.197658 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-7f7bb9bf8f-42ssw"] Nov 28 13:37:57 crc kubenswrapper[4631]: I1128 13:37:57.267132 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/d05d0efc-381c-4222-8c8c-384c2c22ac9c-ovndb-tls-certs\") pod \"neutron-7f7bb9bf8f-42ssw\" (UID: \"d05d0efc-381c-4222-8c8c-384c2c22ac9c\") " pod="openstack/neutron-7f7bb9bf8f-42ssw" Nov 28 13:37:57 crc kubenswrapper[4631]: I1128 13:37:57.267197 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/d05d0efc-381c-4222-8c8c-384c2c22ac9c-config\") pod \"neutron-7f7bb9bf8f-42ssw\" (UID: \"d05d0efc-381c-4222-8c8c-384c2c22ac9c\") " pod="openstack/neutron-7f7bb9bf8f-42ssw" Nov 28 13:37:57 crc kubenswrapper[4631]: I1128 13:37:57.267254 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d05d0efc-381c-4222-8c8c-384c2c22ac9c-combined-ca-bundle\") pod \"neutron-7f7bb9bf8f-42ssw\" (UID: \"d05d0efc-381c-4222-8c8c-384c2c22ac9c\") " pod="openstack/neutron-7f7bb9bf8f-42ssw" Nov 28 13:37:57 crc kubenswrapper[4631]: I1128 13:37:57.267278 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/d05d0efc-381c-4222-8c8c-384c2c22ac9c-httpd-config\") pod \"neutron-7f7bb9bf8f-42ssw\" (UID: \"d05d0efc-381c-4222-8c8c-384c2c22ac9c\") " pod="openstack/neutron-7f7bb9bf8f-42ssw" Nov 28 13:37:57 crc kubenswrapper[4631]: I1128 13:37:57.267334 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d05d0efc-381c-4222-8c8c-384c2c22ac9c-public-tls-certs\") pod \"neutron-7f7bb9bf8f-42ssw\" (UID: \"d05d0efc-381c-4222-8c8c-384c2c22ac9c\") " pod="openstack/neutron-7f7bb9bf8f-42ssw" Nov 28 13:37:57 crc kubenswrapper[4631]: I1128 13:37:57.268612 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d05d0efc-381c-4222-8c8c-384c2c22ac9c-internal-tls-certs\") pod \"neutron-7f7bb9bf8f-42ssw\" (UID: \"d05d0efc-381c-4222-8c8c-384c2c22ac9c\") " pod="openstack/neutron-7f7bb9bf8f-42ssw" Nov 28 13:37:57 crc kubenswrapper[4631]: I1128 13:37:57.268692 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jd6b4\" (UniqueName: \"kubernetes.io/projected/d05d0efc-381c-4222-8c8c-384c2c22ac9c-kube-api-access-jd6b4\") pod \"neutron-7f7bb9bf8f-42ssw\" (UID: \"d05d0efc-381c-4222-8c8c-384c2c22ac9c\") " pod="openstack/neutron-7f7bb9bf8f-42ssw" Nov 28 13:37:57 crc kubenswrapper[4631]: I1128 13:37:57.370454 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jd6b4\" (UniqueName: \"kubernetes.io/projected/d05d0efc-381c-4222-8c8c-384c2c22ac9c-kube-api-access-jd6b4\") pod \"neutron-7f7bb9bf8f-42ssw\" (UID: \"d05d0efc-381c-4222-8c8c-384c2c22ac9c\") " pod="openstack/neutron-7f7bb9bf8f-42ssw" Nov 28 13:37:57 crc kubenswrapper[4631]: I1128 13:37:57.370551 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/d05d0efc-381c-4222-8c8c-384c2c22ac9c-ovndb-tls-certs\") pod \"neutron-7f7bb9bf8f-42ssw\" (UID: \"d05d0efc-381c-4222-8c8c-384c2c22ac9c\") " pod="openstack/neutron-7f7bb9bf8f-42ssw" Nov 28 13:37:57 crc kubenswrapper[4631]: I1128 13:37:57.370587 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/d05d0efc-381c-4222-8c8c-384c2c22ac9c-config\") pod \"neutron-7f7bb9bf8f-42ssw\" (UID: \"d05d0efc-381c-4222-8c8c-384c2c22ac9c\") " pod="openstack/neutron-7f7bb9bf8f-42ssw" Nov 28 13:37:57 crc kubenswrapper[4631]: I1128 13:37:57.370644 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d05d0efc-381c-4222-8c8c-384c2c22ac9c-combined-ca-bundle\") pod \"neutron-7f7bb9bf8f-42ssw\" (UID: \"d05d0efc-381c-4222-8c8c-384c2c22ac9c\") " pod="openstack/neutron-7f7bb9bf8f-42ssw" Nov 28 13:37:57 crc kubenswrapper[4631]: I1128 13:37:57.370669 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/d05d0efc-381c-4222-8c8c-384c2c22ac9c-httpd-config\") pod \"neutron-7f7bb9bf8f-42ssw\" (UID: \"d05d0efc-381c-4222-8c8c-384c2c22ac9c\") " pod="openstack/neutron-7f7bb9bf8f-42ssw" Nov 28 13:37:57 crc kubenswrapper[4631]: I1128 13:37:57.370697 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d05d0efc-381c-4222-8c8c-384c2c22ac9c-public-tls-certs\") pod \"neutron-7f7bb9bf8f-42ssw\" (UID: \"d05d0efc-381c-4222-8c8c-384c2c22ac9c\") " pod="openstack/neutron-7f7bb9bf8f-42ssw" Nov 28 13:37:57 crc kubenswrapper[4631]: I1128 13:37:57.370755 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d05d0efc-381c-4222-8c8c-384c2c22ac9c-internal-tls-certs\") pod \"neutron-7f7bb9bf8f-42ssw\" (UID: \"d05d0efc-381c-4222-8c8c-384c2c22ac9c\") " pod="openstack/neutron-7f7bb9bf8f-42ssw" Nov 28 13:37:57 crc kubenswrapper[4631]: I1128 13:37:57.382604 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/d05d0efc-381c-4222-8c8c-384c2c22ac9c-httpd-config\") pod \"neutron-7f7bb9bf8f-42ssw\" (UID: \"d05d0efc-381c-4222-8c8c-384c2c22ac9c\") " pod="openstack/neutron-7f7bb9bf8f-42ssw" Nov 28 13:37:57 crc kubenswrapper[4631]: I1128 13:37:57.390601 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d05d0efc-381c-4222-8c8c-384c2c22ac9c-public-tls-certs\") pod \"neutron-7f7bb9bf8f-42ssw\" (UID: \"d05d0efc-381c-4222-8c8c-384c2c22ac9c\") " pod="openstack/neutron-7f7bb9bf8f-42ssw" Nov 28 13:37:57 crc kubenswrapper[4631]: I1128 13:37:57.395465 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/d05d0efc-381c-4222-8c8c-384c2c22ac9c-config\") pod \"neutron-7f7bb9bf8f-42ssw\" (UID: \"d05d0efc-381c-4222-8c8c-384c2c22ac9c\") " pod="openstack/neutron-7f7bb9bf8f-42ssw" Nov 28 13:37:57 crc kubenswrapper[4631]: I1128 13:37:57.400121 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jd6b4\" (UniqueName: \"kubernetes.io/projected/d05d0efc-381c-4222-8c8c-384c2c22ac9c-kube-api-access-jd6b4\") pod \"neutron-7f7bb9bf8f-42ssw\" (UID: \"d05d0efc-381c-4222-8c8c-384c2c22ac9c\") " pod="openstack/neutron-7f7bb9bf8f-42ssw" Nov 28 13:37:57 crc kubenswrapper[4631]: I1128 13:37:57.400371 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d05d0efc-381c-4222-8c8c-384c2c22ac9c-combined-ca-bundle\") pod \"neutron-7f7bb9bf8f-42ssw\" (UID: \"d05d0efc-381c-4222-8c8c-384c2c22ac9c\") " pod="openstack/neutron-7f7bb9bf8f-42ssw" Nov 28 13:37:57 crc kubenswrapper[4631]: I1128 13:37:57.401934 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d05d0efc-381c-4222-8c8c-384c2c22ac9c-internal-tls-certs\") pod \"neutron-7f7bb9bf8f-42ssw\" (UID: \"d05d0efc-381c-4222-8c8c-384c2c22ac9c\") " pod="openstack/neutron-7f7bb9bf8f-42ssw" Nov 28 13:37:57 crc kubenswrapper[4631]: I1128 13:37:57.415126 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/d05d0efc-381c-4222-8c8c-384c2c22ac9c-ovndb-tls-certs\") pod \"neutron-7f7bb9bf8f-42ssw\" (UID: \"d05d0efc-381c-4222-8c8c-384c2c22ac9c\") " pod="openstack/neutron-7f7bb9bf8f-42ssw" Nov 28 13:37:57 crc kubenswrapper[4631]: I1128 13:37:57.533730 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-7f7bb9bf8f-42ssw" Nov 28 13:37:57 crc kubenswrapper[4631]: I1128 13:37:57.794342 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"cb966e50-d246-48dc-b4a6-a9de4dbb8569","Type":"ContainerStarted","Data":"6a153b478ff34b35b8edd10fed13a31e0f77eee95649346d1c104a57495f9a84"} Nov 28 13:37:57 crc kubenswrapper[4631]: I1128 13:37:57.799086 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"e1d15959-31a9-464e-9e70-57547ddad347","Type":"ContainerStarted","Data":"91af2216f4119efeb231926969248c0ed81dd41d8674ea566c2a0befef60820f"} Nov 28 13:37:59 crc kubenswrapper[4631]: I1128 13:37:59.336918 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-7f7bb9bf8f-42ssw"] Nov 28 13:37:59 crc kubenswrapper[4631]: W1128 13:37:59.350603 4631 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd05d0efc_381c_4222_8c8c_384c2c22ac9c.slice/crio-0f9c74a5880827406e0bdba0ad89fe1ef7c1b74ef2d1da5bf2ba69f84a011bd7 WatchSource:0}: Error finding container 0f9c74a5880827406e0bdba0ad89fe1ef7c1b74ef2d1da5bf2ba69f84a011bd7: Status 404 returned error can't find the container with id 0f9c74a5880827406e0bdba0ad89fe1ef7c1b74ef2d1da5bf2ba69f84a011bd7 Nov 28 13:37:59 crc kubenswrapper[4631]: I1128 13:37:59.912024 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7f7bb9bf8f-42ssw" event={"ID":"d05d0efc-381c-4222-8c8c-384c2c22ac9c","Type":"ContainerStarted","Data":"0f9c74a5880827406e0bdba0ad89fe1ef7c1b74ef2d1da5bf2ba69f84a011bd7"} Nov 28 13:38:00 crc kubenswrapper[4631]: I1128 13:38:00.956315 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55f844cf75-l4h89" event={"ID":"83f364c4-bdf6-48ba-a5dc-31bb4fbb6f66","Type":"ContainerStarted","Data":"ea19af2e351b50e1211a9c464d27d1c84979bdf113d275a5aed20975b3bff54b"} Nov 28 13:38:00 crc kubenswrapper[4631]: I1128 13:38:00.957061 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-55f844cf75-l4h89" Nov 28 13:38:00 crc kubenswrapper[4631]: I1128 13:38:00.967851 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7f7bb9bf8f-42ssw" event={"ID":"d05d0efc-381c-4222-8c8c-384c2c22ac9c","Type":"ContainerStarted","Data":"7b66bbbcde63e2f3eb793ba678554a3ff25a1f2b65813d0cdcd4a867b3776e9c"} Nov 28 13:38:00 crc kubenswrapper[4631]: I1128 13:38:00.967909 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7f7bb9bf8f-42ssw" event={"ID":"d05d0efc-381c-4222-8c8c-384c2c22ac9c","Type":"ContainerStarted","Data":"7eacc4150af8f65de48171cc8a9ad39a71bf5954f1a9e5ad96a691609ff418d8"} Nov 28 13:38:00 crc kubenswrapper[4631]: I1128 13:38:00.968859 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-7f7bb9bf8f-42ssw" Nov 28 13:38:00 crc kubenswrapper[4631]: I1128 13:38:00.974476 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d95d3536-1392-4576-8ee5-390c09fe7a49","Type":"ContainerStarted","Data":"2141eebe891186cff2651a86531e3f94d1b78cbeeead6dde52101d7a46457ff5"} Nov 28 13:38:00 crc kubenswrapper[4631]: I1128 13:38:00.980539 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5fb67d6644-dbmzf" event={"ID":"6a0a46ba-e48f-45a6-801e-7cb0ececabee","Type":"ContainerStarted","Data":"42641fee468bc31b2584b9a1325c1fdc14069913633b3c778d6d1079d6e2d587"} Nov 28 13:38:00 crc kubenswrapper[4631]: I1128 13:38:00.980742 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-5fb67d6644-dbmzf" Nov 28 13:38:00 crc kubenswrapper[4631]: I1128 13:38:00.989134 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"cb966e50-d246-48dc-b4a6-a9de4dbb8569","Type":"ContainerStarted","Data":"a869370244f2d82eb690074a62782ef88fa1793d2d6fd2b724555af086cd397f"} Nov 28 13:38:00 crc kubenswrapper[4631]: I1128 13:38:00.995396 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-55f844cf75-l4h89" podStartSLOduration=7.995368351 podStartE2EDuration="7.995368351s" podCreationTimestamp="2025-11-28 13:37:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 13:38:00.986812131 +0000 UTC m=+1037.794115485" watchObservedRunningTime="2025-11-28 13:38:00.995368351 +0000 UTC m=+1037.802671695" Nov 28 13:38:00 crc kubenswrapper[4631]: I1128 13:38:00.997516 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"e1d15959-31a9-464e-9e70-57547ddad347","Type":"ContainerStarted","Data":"998cd494d09a844d61a201f6da645a13de219f74073a03e8f64f574ed113822b"} Nov 28 13:38:01 crc kubenswrapper[4631]: I1128 13:38:01.024391 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-5fb67d6644-dbmzf" podStartSLOduration=8.024362954 podStartE2EDuration="8.024362954s" podCreationTimestamp="2025-11-28 13:37:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 13:38:01.012345238 +0000 UTC m=+1037.819648592" watchObservedRunningTime="2025-11-28 13:38:01.024362954 +0000 UTC m=+1037.831666298" Nov 28 13:38:01 crc kubenswrapper[4631]: I1128 13:38:01.036612 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-7f7bb9bf8f-42ssw" podStartSLOduration=4.036579604 podStartE2EDuration="4.036579604s" podCreationTimestamp="2025-11-28 13:37:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 13:38:01.034056522 +0000 UTC m=+1037.841359886" watchObservedRunningTime="2025-11-28 13:38:01.036579604 +0000 UTC m=+1037.843882958" Nov 28 13:38:01 crc kubenswrapper[4631]: I1128 13:38:01.079272 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=8.079246092 podStartE2EDuration="8.079246092s" podCreationTimestamp="2025-11-28 13:37:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 13:38:01.063795242 +0000 UTC m=+1037.871098596" watchObservedRunningTime="2025-11-28 13:38:01.079246092 +0000 UTC m=+1037.886549436" Nov 28 13:38:01 crc kubenswrapper[4631]: I1128 13:38:01.110842 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=8.110803808 podStartE2EDuration="8.110803808s" podCreationTimestamp="2025-11-28 13:37:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 13:38:01.098590648 +0000 UTC m=+1037.905894002" watchObservedRunningTime="2025-11-28 13:38:01.110803808 +0000 UTC m=+1037.918107152" Nov 28 13:38:04 crc kubenswrapper[4631]: I1128 13:38:04.046150 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-xhft6" event={"ID":"05e73b90-be19-4f14-9824-a5c27361adff","Type":"ContainerStarted","Data":"0aaa7f880cc9f4f5eb075f9907b370aaa7455dc78316ee92682613ee25465639"} Nov 28 13:38:04 crc kubenswrapper[4631]: I1128 13:38:04.072530 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-xhft6" podStartSLOduration=3.759141751 podStartE2EDuration="49.072505655s" podCreationTimestamp="2025-11-28 13:37:15 +0000 UTC" firstStartedPulling="2025-11-28 13:37:17.807155898 +0000 UTC m=+994.614459242" lastFinishedPulling="2025-11-28 13:38:03.120519802 +0000 UTC m=+1039.927823146" observedRunningTime="2025-11-28 13:38:04.064427126 +0000 UTC m=+1040.871730470" watchObservedRunningTime="2025-11-28 13:38:04.072505655 +0000 UTC m=+1040.879809009" Nov 28 13:38:04 crc kubenswrapper[4631]: I1128 13:38:04.793830 4631 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Nov 28 13:38:04 crc kubenswrapper[4631]: I1128 13:38:04.796647 4631 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Nov 28 13:38:04 crc kubenswrapper[4631]: I1128 13:38:04.857073 4631 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Nov 28 13:38:04 crc kubenswrapper[4631]: I1128 13:38:04.857153 4631 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Nov 28 13:38:04 crc kubenswrapper[4631]: I1128 13:38:04.904122 4631 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Nov 28 13:38:04 crc kubenswrapper[4631]: I1128 13:38:04.914731 4631 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Nov 28 13:38:04 crc kubenswrapper[4631]: I1128 13:38:04.965216 4631 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Nov 28 13:38:04 crc kubenswrapper[4631]: I1128 13:38:04.984026 4631 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Nov 28 13:38:05 crc kubenswrapper[4631]: I1128 13:38:05.097253 4631 generic.go:334] "Generic (PLEG): container finished" podID="33e9838e-f616-4504-90d5-799de265be04" containerID="4cb68320336ce8ebb3c54401b0bf63ad92137da24f81bdcca4fff5bcff77299f" exitCode=0 Nov 28 13:38:05 crc kubenswrapper[4631]: I1128 13:38:05.097477 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-dcm9x" event={"ID":"33e9838e-f616-4504-90d5-799de265be04","Type":"ContainerDied","Data":"4cb68320336ce8ebb3c54401b0bf63ad92137da24f81bdcca4fff5bcff77299f"} Nov 28 13:38:05 crc kubenswrapper[4631]: I1128 13:38:05.098147 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Nov 28 13:38:05 crc kubenswrapper[4631]: I1128 13:38:05.098195 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Nov 28 13:38:05 crc kubenswrapper[4631]: I1128 13:38:05.098624 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Nov 28 13:38:05 crc kubenswrapper[4631]: I1128 13:38:05.098659 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Nov 28 13:38:05 crc kubenswrapper[4631]: I1128 13:38:05.332317 4631 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-64f8fdf97b-jn5mg" Nov 28 13:38:05 crc kubenswrapper[4631]: I1128 13:38:05.332412 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-64f8fdf97b-jn5mg" Nov 28 13:38:05 crc kubenswrapper[4631]: I1128 13:38:05.336230 4631 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-64f8fdf97b-jn5mg" podUID="111c5349-528d-4856-a3e4-1277e3b10889" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.146:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.146:8443: connect: connection refused" Nov 28 13:38:05 crc kubenswrapper[4631]: I1128 13:38:05.558673 4631 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-544496c768-rzhsw" podUID="cd607dae-568a-4b81-af81-3310c0e95854" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.147:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.147:8443: connect: connection refused" Nov 28 13:38:05 crc kubenswrapper[4631]: I1128 13:38:05.563408 4631 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-544496c768-rzhsw" Nov 28 13:38:05 crc kubenswrapper[4631]: I1128 13:38:05.563462 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-544496c768-rzhsw" Nov 28 13:38:08 crc kubenswrapper[4631]: I1128 13:38:08.667546 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-55f844cf75-l4h89" Nov 28 13:38:08 crc kubenswrapper[4631]: I1128 13:38:08.796102 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-7ljwv"] Nov 28 13:38:08 crc kubenswrapper[4631]: I1128 13:38:08.802073 4631 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-785d8bcb8c-7ljwv" podUID="8d7961e8-77d2-4b92-8758-f06c6d216f73" containerName="dnsmasq-dns" containerID="cri-o://5a249c121250725ea538cb440f5f420c7a93cf29d5adce655cb1291f01ea65f2" gracePeriod=10 Nov 28 13:38:09 crc kubenswrapper[4631]: I1128 13:38:09.159966 4631 generic.go:334] "Generic (PLEG): container finished" podID="8d7961e8-77d2-4b92-8758-f06c6d216f73" containerID="5a249c121250725ea538cb440f5f420c7a93cf29d5adce655cb1291f01ea65f2" exitCode=0 Nov 28 13:38:09 crc kubenswrapper[4631]: I1128 13:38:09.160092 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-7ljwv" event={"ID":"8d7961e8-77d2-4b92-8758-f06c6d216f73","Type":"ContainerDied","Data":"5a249c121250725ea538cb440f5f420c7a93cf29d5adce655cb1291f01ea65f2"} Nov 28 13:38:09 crc kubenswrapper[4631]: I1128 13:38:09.166317 4631 generic.go:334] "Generic (PLEG): container finished" podID="05e73b90-be19-4f14-9824-a5c27361adff" containerID="0aaa7f880cc9f4f5eb075f9907b370aaa7455dc78316ee92682613ee25465639" exitCode=0 Nov 28 13:38:09 crc kubenswrapper[4631]: I1128 13:38:09.166379 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-xhft6" event={"ID":"05e73b90-be19-4f14-9824-a5c27361adff","Type":"ContainerDied","Data":"0aaa7f880cc9f4f5eb075f9907b370aaa7455dc78316ee92682613ee25465639"} Nov 28 13:38:10 crc kubenswrapper[4631]: I1128 13:38:10.965238 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Nov 28 13:38:10 crc kubenswrapper[4631]: I1128 13:38:10.965745 4631 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 28 13:38:10 crc kubenswrapper[4631]: I1128 13:38:10.969064 4631 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-785d8bcb8c-7ljwv" podUID="8d7961e8-77d2-4b92-8758-f06c6d216f73" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.142:5353: connect: connection refused" Nov 28 13:38:10 crc kubenswrapper[4631]: I1128 13:38:10.978997 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Nov 28 13:38:11 crc kubenswrapper[4631]: I1128 13:38:11.802452 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Nov 28 13:38:11 crc kubenswrapper[4631]: I1128 13:38:11.802682 4631 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 28 13:38:12 crc kubenswrapper[4631]: I1128 13:38:12.072840 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Nov 28 13:38:13 crc kubenswrapper[4631]: I1128 13:38:13.802501 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-xhft6" Nov 28 13:38:13 crc kubenswrapper[4631]: I1128 13:38:13.826836 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-dcm9x" Nov 28 13:38:13 crc kubenswrapper[4631]: I1128 13:38:13.885022 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/33e9838e-f616-4504-90d5-799de265be04-scripts\") pod \"33e9838e-f616-4504-90d5-799de265be04\" (UID: \"33e9838e-f616-4504-90d5-799de265be04\") " Nov 28 13:38:13 crc kubenswrapper[4631]: I1128 13:38:13.885082 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t6gw4\" (UniqueName: \"kubernetes.io/projected/05e73b90-be19-4f14-9824-a5c27361adff-kube-api-access-t6gw4\") pod \"05e73b90-be19-4f14-9824-a5c27361adff\" (UID: \"05e73b90-be19-4f14-9824-a5c27361adff\") " Nov 28 13:38:13 crc kubenswrapper[4631]: I1128 13:38:13.885106 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/05e73b90-be19-4f14-9824-a5c27361adff-config-data\") pod \"05e73b90-be19-4f14-9824-a5c27361adff\" (UID: \"05e73b90-be19-4f14-9824-a5c27361adff\") " Nov 28 13:38:13 crc kubenswrapper[4631]: I1128 13:38:13.885166 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/05e73b90-be19-4f14-9824-a5c27361adff-logs\") pod \"05e73b90-be19-4f14-9824-a5c27361adff\" (UID: \"05e73b90-be19-4f14-9824-a5c27361adff\") " Nov 28 13:38:13 crc kubenswrapper[4631]: I1128 13:38:13.885198 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/33e9838e-f616-4504-90d5-799de265be04-combined-ca-bundle\") pod \"33e9838e-f616-4504-90d5-799de265be04\" (UID: \"33e9838e-f616-4504-90d5-799de265be04\") " Nov 28 13:38:13 crc kubenswrapper[4631]: I1128 13:38:13.885271 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/33e9838e-f616-4504-90d5-799de265be04-credential-keys\") pod \"33e9838e-f616-4504-90d5-799de265be04\" (UID: \"33e9838e-f616-4504-90d5-799de265be04\") " Nov 28 13:38:13 crc kubenswrapper[4631]: I1128 13:38:13.885310 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vlgfj\" (UniqueName: \"kubernetes.io/projected/33e9838e-f616-4504-90d5-799de265be04-kube-api-access-vlgfj\") pod \"33e9838e-f616-4504-90d5-799de265be04\" (UID: \"33e9838e-f616-4504-90d5-799de265be04\") " Nov 28 13:38:13 crc kubenswrapper[4631]: I1128 13:38:13.885333 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/05e73b90-be19-4f14-9824-a5c27361adff-scripts\") pod \"05e73b90-be19-4f14-9824-a5c27361adff\" (UID: \"05e73b90-be19-4f14-9824-a5c27361adff\") " Nov 28 13:38:13 crc kubenswrapper[4631]: I1128 13:38:13.885359 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/33e9838e-f616-4504-90d5-799de265be04-config-data\") pod \"33e9838e-f616-4504-90d5-799de265be04\" (UID: \"33e9838e-f616-4504-90d5-799de265be04\") " Nov 28 13:38:13 crc kubenswrapper[4631]: I1128 13:38:13.885430 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/05e73b90-be19-4f14-9824-a5c27361adff-combined-ca-bundle\") pod \"05e73b90-be19-4f14-9824-a5c27361adff\" (UID: \"05e73b90-be19-4f14-9824-a5c27361adff\") " Nov 28 13:38:13 crc kubenswrapper[4631]: I1128 13:38:13.885455 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/33e9838e-f616-4504-90d5-799de265be04-fernet-keys\") pod \"33e9838e-f616-4504-90d5-799de265be04\" (UID: \"33e9838e-f616-4504-90d5-799de265be04\") " Nov 28 13:38:13 crc kubenswrapper[4631]: I1128 13:38:13.918000 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/05e73b90-be19-4f14-9824-a5c27361adff-kube-api-access-t6gw4" (OuterVolumeSpecName: "kube-api-access-t6gw4") pod "05e73b90-be19-4f14-9824-a5c27361adff" (UID: "05e73b90-be19-4f14-9824-a5c27361adff"). InnerVolumeSpecName "kube-api-access-t6gw4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:38:13 crc kubenswrapper[4631]: I1128 13:38:13.923547 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/05e73b90-be19-4f14-9824-a5c27361adff-logs" (OuterVolumeSpecName: "logs") pod "05e73b90-be19-4f14-9824-a5c27361adff" (UID: "05e73b90-be19-4f14-9824-a5c27361adff"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 13:38:13 crc kubenswrapper[4631]: I1128 13:38:13.935130 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/33e9838e-f616-4504-90d5-799de265be04-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "33e9838e-f616-4504-90d5-799de265be04" (UID: "33e9838e-f616-4504-90d5-799de265be04"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:38:13 crc kubenswrapper[4631]: I1128 13:38:13.935216 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/05e73b90-be19-4f14-9824-a5c27361adff-scripts" (OuterVolumeSpecName: "scripts") pod "05e73b90-be19-4f14-9824-a5c27361adff" (UID: "05e73b90-be19-4f14-9824-a5c27361adff"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:38:13 crc kubenswrapper[4631]: I1128 13:38:13.935262 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/33e9838e-f616-4504-90d5-799de265be04-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "33e9838e-f616-4504-90d5-799de265be04" (UID: "33e9838e-f616-4504-90d5-799de265be04"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:38:13 crc kubenswrapper[4631]: I1128 13:38:13.935394 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/33e9838e-f616-4504-90d5-799de265be04-kube-api-access-vlgfj" (OuterVolumeSpecName: "kube-api-access-vlgfj") pod "33e9838e-f616-4504-90d5-799de265be04" (UID: "33e9838e-f616-4504-90d5-799de265be04"). InnerVolumeSpecName "kube-api-access-vlgfj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:38:13 crc kubenswrapper[4631]: I1128 13:38:13.944096 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/33e9838e-f616-4504-90d5-799de265be04-scripts" (OuterVolumeSpecName: "scripts") pod "33e9838e-f616-4504-90d5-799de265be04" (UID: "33e9838e-f616-4504-90d5-799de265be04"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:38:14 crc kubenswrapper[4631]: I1128 13:38:13.989597 4631 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/33e9838e-f616-4504-90d5-799de265be04-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 13:38:14 crc kubenswrapper[4631]: I1128 13:38:13.989641 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t6gw4\" (UniqueName: \"kubernetes.io/projected/05e73b90-be19-4f14-9824-a5c27361adff-kube-api-access-t6gw4\") on node \"crc\" DevicePath \"\"" Nov 28 13:38:14 crc kubenswrapper[4631]: I1128 13:38:13.989654 4631 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/05e73b90-be19-4f14-9824-a5c27361adff-logs\") on node \"crc\" DevicePath \"\"" Nov 28 13:38:14 crc kubenswrapper[4631]: I1128 13:38:13.989665 4631 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/33e9838e-f616-4504-90d5-799de265be04-credential-keys\") on node \"crc\" DevicePath \"\"" Nov 28 13:38:14 crc kubenswrapper[4631]: I1128 13:38:13.989674 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vlgfj\" (UniqueName: \"kubernetes.io/projected/33e9838e-f616-4504-90d5-799de265be04-kube-api-access-vlgfj\") on node \"crc\" DevicePath \"\"" Nov 28 13:38:14 crc kubenswrapper[4631]: I1128 13:38:13.989683 4631 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/05e73b90-be19-4f14-9824-a5c27361adff-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 13:38:14 crc kubenswrapper[4631]: I1128 13:38:13.989691 4631 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/33e9838e-f616-4504-90d5-799de265be04-fernet-keys\") on node \"crc\" DevicePath \"\"" Nov 28 13:38:14 crc kubenswrapper[4631]: I1128 13:38:13.995785 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/33e9838e-f616-4504-90d5-799de265be04-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "33e9838e-f616-4504-90d5-799de265be04" (UID: "33e9838e-f616-4504-90d5-799de265be04"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:38:14 crc kubenswrapper[4631]: I1128 13:38:14.108802 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/33e9838e-f616-4504-90d5-799de265be04-config-data" (OuterVolumeSpecName: "config-data") pod "33e9838e-f616-4504-90d5-799de265be04" (UID: "33e9838e-f616-4504-90d5-799de265be04"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:38:14 crc kubenswrapper[4631]: I1128 13:38:14.112493 4631 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/33e9838e-f616-4504-90d5-799de265be04-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 13:38:14 crc kubenswrapper[4631]: I1128 13:38:14.112542 4631 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/33e9838e-f616-4504-90d5-799de265be04-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 13:38:14 crc kubenswrapper[4631]: I1128 13:38:14.202975 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/05e73b90-be19-4f14-9824-a5c27361adff-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "05e73b90-be19-4f14-9824-a5c27361adff" (UID: "05e73b90-be19-4f14-9824-a5c27361adff"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:38:14 crc kubenswrapper[4631]: I1128 13:38:14.204211 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-785d8bcb8c-7ljwv" Nov 28 13:38:14 crc kubenswrapper[4631]: I1128 13:38:14.214243 4631 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/05e73b90-be19-4f14-9824-a5c27361adff-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 13:38:14 crc kubenswrapper[4631]: I1128 13:38:14.218439 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/05e73b90-be19-4f14-9824-a5c27361adff-config-data" (OuterVolumeSpecName: "config-data") pod "05e73b90-be19-4f14-9824-a5c27361adff" (UID: "05e73b90-be19-4f14-9824-a5c27361adff"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:38:14 crc kubenswrapper[4631]: I1128 13:38:14.293922 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-xhft6" event={"ID":"05e73b90-be19-4f14-9824-a5c27361adff","Type":"ContainerDied","Data":"6711ceda49aba2a213632d2f369e6f3edeac4568c545e52d706d4124e2e414c0"} Nov 28 13:38:14 crc kubenswrapper[4631]: I1128 13:38:14.293990 4631 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6711ceda49aba2a213632d2f369e6f3edeac4568c545e52d706d4124e2e414c0" Nov 28 13:38:14 crc kubenswrapper[4631]: I1128 13:38:14.294114 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-xhft6" Nov 28 13:38:14 crc kubenswrapper[4631]: I1128 13:38:14.304897 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-dcm9x" event={"ID":"33e9838e-f616-4504-90d5-799de265be04","Type":"ContainerDied","Data":"6476af066dead47f55a2df29345d1427b1180fd8cba959fa66e7109c5684f5c3"} Nov 28 13:38:14 crc kubenswrapper[4631]: I1128 13:38:14.304965 4631 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6476af066dead47f55a2df29345d1427b1180fd8cba959fa66e7109c5684f5c3" Nov 28 13:38:14 crc kubenswrapper[4631]: I1128 13:38:14.305069 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-dcm9x" Nov 28 13:38:14 crc kubenswrapper[4631]: I1128 13:38:14.316159 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8d7961e8-77d2-4b92-8758-f06c6d216f73-config\") pod \"8d7961e8-77d2-4b92-8758-f06c6d216f73\" (UID: \"8d7961e8-77d2-4b92-8758-f06c6d216f73\") " Nov 28 13:38:14 crc kubenswrapper[4631]: I1128 13:38:14.316886 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8d7961e8-77d2-4b92-8758-f06c6d216f73-dns-swift-storage-0\") pod \"8d7961e8-77d2-4b92-8758-f06c6d216f73\" (UID: \"8d7961e8-77d2-4b92-8758-f06c6d216f73\") " Nov 28 13:38:14 crc kubenswrapper[4631]: I1128 13:38:14.317022 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8d7961e8-77d2-4b92-8758-f06c6d216f73-dns-svc\") pod \"8d7961e8-77d2-4b92-8758-f06c6d216f73\" (UID: \"8d7961e8-77d2-4b92-8758-f06c6d216f73\") " Nov 28 13:38:14 crc kubenswrapper[4631]: I1128 13:38:14.317284 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8d7961e8-77d2-4b92-8758-f06c6d216f73-ovsdbserver-nb\") pod \"8d7961e8-77d2-4b92-8758-f06c6d216f73\" (UID: \"8d7961e8-77d2-4b92-8758-f06c6d216f73\") " Nov 28 13:38:14 crc kubenswrapper[4631]: I1128 13:38:14.317415 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jprwc\" (UniqueName: \"kubernetes.io/projected/8d7961e8-77d2-4b92-8758-f06c6d216f73-kube-api-access-jprwc\") pod \"8d7961e8-77d2-4b92-8758-f06c6d216f73\" (UID: \"8d7961e8-77d2-4b92-8758-f06c6d216f73\") " Nov 28 13:38:14 crc kubenswrapper[4631]: I1128 13:38:14.317613 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8d7961e8-77d2-4b92-8758-f06c6d216f73-ovsdbserver-sb\") pod \"8d7961e8-77d2-4b92-8758-f06c6d216f73\" (UID: \"8d7961e8-77d2-4b92-8758-f06c6d216f73\") " Nov 28 13:38:14 crc kubenswrapper[4631]: I1128 13:38:14.318112 4631 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/05e73b90-be19-4f14-9824-a5c27361adff-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 13:38:14 crc kubenswrapper[4631]: I1128 13:38:14.323762 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-7ljwv" event={"ID":"8d7961e8-77d2-4b92-8758-f06c6d216f73","Type":"ContainerDied","Data":"c06588ceb250932379b10bee3a4c8730bdff42b76db2c5d7d4d8af456e63e93f"} Nov 28 13:38:14 crc kubenswrapper[4631]: I1128 13:38:14.323843 4631 scope.go:117] "RemoveContainer" containerID="5a249c121250725ea538cb440f5f420c7a93cf29d5adce655cb1291f01ea65f2" Nov 28 13:38:14 crc kubenswrapper[4631]: I1128 13:38:14.324020 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-785d8bcb8c-7ljwv" Nov 28 13:38:14 crc kubenswrapper[4631]: I1128 13:38:14.338881 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8d7961e8-77d2-4b92-8758-f06c6d216f73-kube-api-access-jprwc" (OuterVolumeSpecName: "kube-api-access-jprwc") pod "8d7961e8-77d2-4b92-8758-f06c6d216f73" (UID: "8d7961e8-77d2-4b92-8758-f06c6d216f73"). InnerVolumeSpecName "kube-api-access-jprwc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:38:14 crc kubenswrapper[4631]: I1128 13:38:14.427538 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jprwc\" (UniqueName: \"kubernetes.io/projected/8d7961e8-77d2-4b92-8758-f06c6d216f73-kube-api-access-jprwc\") on node \"crc\" DevicePath \"\"" Nov 28 13:38:14 crc kubenswrapper[4631]: I1128 13:38:14.459912 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8d7961e8-77d2-4b92-8758-f06c6d216f73-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "8d7961e8-77d2-4b92-8758-f06c6d216f73" (UID: "8d7961e8-77d2-4b92-8758-f06c6d216f73"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 13:38:14 crc kubenswrapper[4631]: I1128 13:38:14.471330 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8d7961e8-77d2-4b92-8758-f06c6d216f73-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "8d7961e8-77d2-4b92-8758-f06c6d216f73" (UID: "8d7961e8-77d2-4b92-8758-f06c6d216f73"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 13:38:14 crc kubenswrapper[4631]: I1128 13:38:14.478143 4631 scope.go:117] "RemoveContainer" containerID="42dda41f0135561bf1f4863d37b0bf6d3670e53ab272554ff0dc38f4e2b83ee3" Nov 28 13:38:14 crc kubenswrapper[4631]: I1128 13:38:14.479522 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8d7961e8-77d2-4b92-8758-f06c6d216f73-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "8d7961e8-77d2-4b92-8758-f06c6d216f73" (UID: "8d7961e8-77d2-4b92-8758-f06c6d216f73"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 13:38:14 crc kubenswrapper[4631]: I1128 13:38:14.506432 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8d7961e8-77d2-4b92-8758-f06c6d216f73-config" (OuterVolumeSpecName: "config") pod "8d7961e8-77d2-4b92-8758-f06c6d216f73" (UID: "8d7961e8-77d2-4b92-8758-f06c6d216f73"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 13:38:14 crc kubenswrapper[4631]: I1128 13:38:14.550762 4631 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8d7961e8-77d2-4b92-8758-f06c6d216f73-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 28 13:38:14 crc kubenswrapper[4631]: I1128 13:38:14.550837 4631 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8d7961e8-77d2-4b92-8758-f06c6d216f73-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 28 13:38:14 crc kubenswrapper[4631]: I1128 13:38:14.550854 4631 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8d7961e8-77d2-4b92-8758-f06c6d216f73-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 28 13:38:14 crc kubenswrapper[4631]: I1128 13:38:14.550863 4631 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8d7961e8-77d2-4b92-8758-f06c6d216f73-config\") on node \"crc\" DevicePath \"\"" Nov 28 13:38:14 crc kubenswrapper[4631]: I1128 13:38:14.580160 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8d7961e8-77d2-4b92-8758-f06c6d216f73-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "8d7961e8-77d2-4b92-8758-f06c6d216f73" (UID: "8d7961e8-77d2-4b92-8758-f06c6d216f73"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 13:38:14 crc kubenswrapper[4631]: I1128 13:38:14.658781 4631 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8d7961e8-77d2-4b92-8758-f06c6d216f73-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 28 13:38:14 crc kubenswrapper[4631]: I1128 13:38:14.771782 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-7ljwv"] Nov 28 13:38:14 crc kubenswrapper[4631]: I1128 13:38:14.781032 4631 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-7ljwv"] Nov 28 13:38:14 crc kubenswrapper[4631]: I1128 13:38:14.976005 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-769746444-8zhp8"] Nov 28 13:38:14 crc kubenswrapper[4631]: E1128 13:38:14.976562 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="05e73b90-be19-4f14-9824-a5c27361adff" containerName="placement-db-sync" Nov 28 13:38:14 crc kubenswrapper[4631]: I1128 13:38:14.976584 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="05e73b90-be19-4f14-9824-a5c27361adff" containerName="placement-db-sync" Nov 28 13:38:14 crc kubenswrapper[4631]: E1128 13:38:14.976608 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="33e9838e-f616-4504-90d5-799de265be04" containerName="keystone-bootstrap" Nov 28 13:38:14 crc kubenswrapper[4631]: I1128 13:38:14.976617 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="33e9838e-f616-4504-90d5-799de265be04" containerName="keystone-bootstrap" Nov 28 13:38:14 crc kubenswrapper[4631]: E1128 13:38:14.976634 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d7961e8-77d2-4b92-8758-f06c6d216f73" containerName="init" Nov 28 13:38:14 crc kubenswrapper[4631]: I1128 13:38:14.976642 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d7961e8-77d2-4b92-8758-f06c6d216f73" containerName="init" Nov 28 13:38:14 crc kubenswrapper[4631]: E1128 13:38:14.976659 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d7961e8-77d2-4b92-8758-f06c6d216f73" containerName="dnsmasq-dns" Nov 28 13:38:14 crc kubenswrapper[4631]: I1128 13:38:14.976666 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d7961e8-77d2-4b92-8758-f06c6d216f73" containerName="dnsmasq-dns" Nov 28 13:38:14 crc kubenswrapper[4631]: I1128 13:38:14.976910 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="33e9838e-f616-4504-90d5-799de265be04" containerName="keystone-bootstrap" Nov 28 13:38:14 crc kubenswrapper[4631]: I1128 13:38:14.981891 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="8d7961e8-77d2-4b92-8758-f06c6d216f73" containerName="dnsmasq-dns" Nov 28 13:38:14 crc kubenswrapper[4631]: I1128 13:38:14.981974 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="05e73b90-be19-4f14-9824-a5c27361adff" containerName="placement-db-sync" Nov 28 13:38:14 crc kubenswrapper[4631]: I1128 13:38:14.983408 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-769746444-8zhp8" Nov 28 13:38:14 crc kubenswrapper[4631]: I1128 13:38:14.996110 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Nov 28 13:38:14 crc kubenswrapper[4631]: I1128 13:38:14.996429 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Nov 28 13:38:15 crc kubenswrapper[4631]: I1128 13:38:15.000509 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Nov 28 13:38:15 crc kubenswrapper[4631]: I1128 13:38:15.004466 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-769746444-8zhp8"] Nov 28 13:38:15 crc kubenswrapper[4631]: I1128 13:38:15.000586 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Nov 28 13:38:15 crc kubenswrapper[4631]: I1128 13:38:15.001006 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-d778x" Nov 28 13:38:15 crc kubenswrapper[4631]: I1128 13:38:15.216532 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/176c2dec-4940-449b-ab1a-95cb11cdfeff-combined-ca-bundle\") pod \"placement-769746444-8zhp8\" (UID: \"176c2dec-4940-449b-ab1a-95cb11cdfeff\") " pod="openstack/placement-769746444-8zhp8" Nov 28 13:38:15 crc kubenswrapper[4631]: I1128 13:38:15.216642 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/176c2dec-4940-449b-ab1a-95cb11cdfeff-scripts\") pod \"placement-769746444-8zhp8\" (UID: \"176c2dec-4940-449b-ab1a-95cb11cdfeff\") " pod="openstack/placement-769746444-8zhp8" Nov 28 13:38:15 crc kubenswrapper[4631]: I1128 13:38:15.216683 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/176c2dec-4940-449b-ab1a-95cb11cdfeff-public-tls-certs\") pod \"placement-769746444-8zhp8\" (UID: \"176c2dec-4940-449b-ab1a-95cb11cdfeff\") " pod="openstack/placement-769746444-8zhp8" Nov 28 13:38:15 crc kubenswrapper[4631]: I1128 13:38:15.216726 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/176c2dec-4940-449b-ab1a-95cb11cdfeff-internal-tls-certs\") pod \"placement-769746444-8zhp8\" (UID: \"176c2dec-4940-449b-ab1a-95cb11cdfeff\") " pod="openstack/placement-769746444-8zhp8" Nov 28 13:38:15 crc kubenswrapper[4631]: I1128 13:38:15.216754 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/176c2dec-4940-449b-ab1a-95cb11cdfeff-logs\") pod \"placement-769746444-8zhp8\" (UID: \"176c2dec-4940-449b-ab1a-95cb11cdfeff\") " pod="openstack/placement-769746444-8zhp8" Nov 28 13:38:15 crc kubenswrapper[4631]: I1128 13:38:15.216818 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/176c2dec-4940-449b-ab1a-95cb11cdfeff-config-data\") pod \"placement-769746444-8zhp8\" (UID: \"176c2dec-4940-449b-ab1a-95cb11cdfeff\") " pod="openstack/placement-769746444-8zhp8" Nov 28 13:38:15 crc kubenswrapper[4631]: I1128 13:38:15.216843 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jspdq\" (UniqueName: \"kubernetes.io/projected/176c2dec-4940-449b-ab1a-95cb11cdfeff-kube-api-access-jspdq\") pod \"placement-769746444-8zhp8\" (UID: \"176c2dec-4940-449b-ab1a-95cb11cdfeff\") " pod="openstack/placement-769746444-8zhp8" Nov 28 13:38:15 crc kubenswrapper[4631]: I1128 13:38:15.264386 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-6bd6cfc878-65bg5"] Nov 28 13:38:15 crc kubenswrapper[4631]: I1128 13:38:15.269500 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-6bd6cfc878-65bg5" Nov 28 13:38:15 crc kubenswrapper[4631]: I1128 13:38:15.277346 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Nov 28 13:38:15 crc kubenswrapper[4631]: I1128 13:38:15.277747 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-szfjn" Nov 28 13:38:15 crc kubenswrapper[4631]: I1128 13:38:15.278201 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Nov 28 13:38:15 crc kubenswrapper[4631]: I1128 13:38:15.278567 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Nov 28 13:38:15 crc kubenswrapper[4631]: I1128 13:38:15.280735 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Nov 28 13:38:15 crc kubenswrapper[4631]: I1128 13:38:15.282158 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-6bd6cfc878-65bg5"] Nov 28 13:38:15 crc kubenswrapper[4631]: I1128 13:38:15.286027 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Nov 28 13:38:15 crc kubenswrapper[4631]: I1128 13:38:15.320306 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2adae216-4db3-49d7-8d11-24bbd7d47f8f-config-data\") pod \"keystone-6bd6cfc878-65bg5\" (UID: \"2adae216-4db3-49d7-8d11-24bbd7d47f8f\") " pod="openstack/keystone-6bd6cfc878-65bg5" Nov 28 13:38:15 crc kubenswrapper[4631]: I1128 13:38:15.320404 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/176c2dec-4940-449b-ab1a-95cb11cdfeff-scripts\") pod \"placement-769746444-8zhp8\" (UID: \"176c2dec-4940-449b-ab1a-95cb11cdfeff\") " pod="openstack/placement-769746444-8zhp8" Nov 28 13:38:15 crc kubenswrapper[4631]: I1128 13:38:15.320472 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/176c2dec-4940-449b-ab1a-95cb11cdfeff-public-tls-certs\") pod \"placement-769746444-8zhp8\" (UID: \"176c2dec-4940-449b-ab1a-95cb11cdfeff\") " pod="openstack/placement-769746444-8zhp8" Nov 28 13:38:15 crc kubenswrapper[4631]: I1128 13:38:15.320502 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-btg7k\" (UniqueName: \"kubernetes.io/projected/2adae216-4db3-49d7-8d11-24bbd7d47f8f-kube-api-access-btg7k\") pod \"keystone-6bd6cfc878-65bg5\" (UID: \"2adae216-4db3-49d7-8d11-24bbd7d47f8f\") " pod="openstack/keystone-6bd6cfc878-65bg5" Nov 28 13:38:15 crc kubenswrapper[4631]: I1128 13:38:15.320523 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/2adae216-4db3-49d7-8d11-24bbd7d47f8f-fernet-keys\") pod \"keystone-6bd6cfc878-65bg5\" (UID: \"2adae216-4db3-49d7-8d11-24bbd7d47f8f\") " pod="openstack/keystone-6bd6cfc878-65bg5" Nov 28 13:38:15 crc kubenswrapper[4631]: I1128 13:38:15.320542 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2adae216-4db3-49d7-8d11-24bbd7d47f8f-public-tls-certs\") pod \"keystone-6bd6cfc878-65bg5\" (UID: \"2adae216-4db3-49d7-8d11-24bbd7d47f8f\") " pod="openstack/keystone-6bd6cfc878-65bg5" Nov 28 13:38:15 crc kubenswrapper[4631]: I1128 13:38:15.320576 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/176c2dec-4940-449b-ab1a-95cb11cdfeff-internal-tls-certs\") pod \"placement-769746444-8zhp8\" (UID: \"176c2dec-4940-449b-ab1a-95cb11cdfeff\") " pod="openstack/placement-769746444-8zhp8" Nov 28 13:38:15 crc kubenswrapper[4631]: I1128 13:38:15.320604 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/176c2dec-4940-449b-ab1a-95cb11cdfeff-logs\") pod \"placement-769746444-8zhp8\" (UID: \"176c2dec-4940-449b-ab1a-95cb11cdfeff\") " pod="openstack/placement-769746444-8zhp8" Nov 28 13:38:15 crc kubenswrapper[4631]: I1128 13:38:15.320635 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2adae216-4db3-49d7-8d11-24bbd7d47f8f-combined-ca-bundle\") pod \"keystone-6bd6cfc878-65bg5\" (UID: \"2adae216-4db3-49d7-8d11-24bbd7d47f8f\") " pod="openstack/keystone-6bd6cfc878-65bg5" Nov 28 13:38:15 crc kubenswrapper[4631]: I1128 13:38:15.321072 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/176c2dec-4940-449b-ab1a-95cb11cdfeff-config-data\") pod \"placement-769746444-8zhp8\" (UID: \"176c2dec-4940-449b-ab1a-95cb11cdfeff\") " pod="openstack/placement-769746444-8zhp8" Nov 28 13:38:15 crc kubenswrapper[4631]: I1128 13:38:15.321103 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2adae216-4db3-49d7-8d11-24bbd7d47f8f-scripts\") pod \"keystone-6bd6cfc878-65bg5\" (UID: \"2adae216-4db3-49d7-8d11-24bbd7d47f8f\") " pod="openstack/keystone-6bd6cfc878-65bg5" Nov 28 13:38:15 crc kubenswrapper[4631]: I1128 13:38:15.321153 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jspdq\" (UniqueName: \"kubernetes.io/projected/176c2dec-4940-449b-ab1a-95cb11cdfeff-kube-api-access-jspdq\") pod \"placement-769746444-8zhp8\" (UID: \"176c2dec-4940-449b-ab1a-95cb11cdfeff\") " pod="openstack/placement-769746444-8zhp8" Nov 28 13:38:15 crc kubenswrapper[4631]: I1128 13:38:15.321207 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/2adae216-4db3-49d7-8d11-24bbd7d47f8f-credential-keys\") pod \"keystone-6bd6cfc878-65bg5\" (UID: \"2adae216-4db3-49d7-8d11-24bbd7d47f8f\") " pod="openstack/keystone-6bd6cfc878-65bg5" Nov 28 13:38:15 crc kubenswrapper[4631]: I1128 13:38:15.321260 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2adae216-4db3-49d7-8d11-24bbd7d47f8f-internal-tls-certs\") pod \"keystone-6bd6cfc878-65bg5\" (UID: \"2adae216-4db3-49d7-8d11-24bbd7d47f8f\") " pod="openstack/keystone-6bd6cfc878-65bg5" Nov 28 13:38:15 crc kubenswrapper[4631]: I1128 13:38:15.321317 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/176c2dec-4940-449b-ab1a-95cb11cdfeff-combined-ca-bundle\") pod \"placement-769746444-8zhp8\" (UID: \"176c2dec-4940-449b-ab1a-95cb11cdfeff\") " pod="openstack/placement-769746444-8zhp8" Nov 28 13:38:15 crc kubenswrapper[4631]: I1128 13:38:15.322649 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/176c2dec-4940-449b-ab1a-95cb11cdfeff-logs\") pod \"placement-769746444-8zhp8\" (UID: \"176c2dec-4940-449b-ab1a-95cb11cdfeff\") " pod="openstack/placement-769746444-8zhp8" Nov 28 13:38:15 crc kubenswrapper[4631]: I1128 13:38:15.335560 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/176c2dec-4940-449b-ab1a-95cb11cdfeff-combined-ca-bundle\") pod \"placement-769746444-8zhp8\" (UID: \"176c2dec-4940-449b-ab1a-95cb11cdfeff\") " pod="openstack/placement-769746444-8zhp8" Nov 28 13:38:15 crc kubenswrapper[4631]: I1128 13:38:15.336427 4631 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-64f8fdf97b-jn5mg" podUID="111c5349-528d-4856-a3e4-1277e3b10889" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.146:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.146:8443: connect: connection refused" Nov 28 13:38:15 crc kubenswrapper[4631]: I1128 13:38:15.381884 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/176c2dec-4940-449b-ab1a-95cb11cdfeff-config-data\") pod \"placement-769746444-8zhp8\" (UID: \"176c2dec-4940-449b-ab1a-95cb11cdfeff\") " pod="openstack/placement-769746444-8zhp8" Nov 28 13:38:15 crc kubenswrapper[4631]: I1128 13:38:15.386379 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-z6tgb" event={"ID":"3050013d-2db4-4801-8626-1fb6398b714e","Type":"ContainerStarted","Data":"8659b5cf744dfcf5eb7c3313fb4013f0f03ffb7680a0047f33139bfff995c446"} Nov 28 13:38:15 crc kubenswrapper[4631]: I1128 13:38:15.386568 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/176c2dec-4940-449b-ab1a-95cb11cdfeff-scripts\") pod \"placement-769746444-8zhp8\" (UID: \"176c2dec-4940-449b-ab1a-95cb11cdfeff\") " pod="openstack/placement-769746444-8zhp8" Nov 28 13:38:15 crc kubenswrapper[4631]: I1128 13:38:15.386988 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/176c2dec-4940-449b-ab1a-95cb11cdfeff-internal-tls-certs\") pod \"placement-769746444-8zhp8\" (UID: \"176c2dec-4940-449b-ab1a-95cb11cdfeff\") " pod="openstack/placement-769746444-8zhp8" Nov 28 13:38:15 crc kubenswrapper[4631]: I1128 13:38:15.394223 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d95d3536-1392-4576-8ee5-390c09fe7a49","Type":"ContainerStarted","Data":"aed54e68e791750f2a7b381406c45fd57ec0e772fa0e75496b255dcade224d18"} Nov 28 13:38:15 crc kubenswrapper[4631]: I1128 13:38:15.423196 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/176c2dec-4940-449b-ab1a-95cb11cdfeff-public-tls-certs\") pod \"placement-769746444-8zhp8\" (UID: \"176c2dec-4940-449b-ab1a-95cb11cdfeff\") " pod="openstack/placement-769746444-8zhp8" Nov 28 13:38:15 crc kubenswrapper[4631]: I1128 13:38:15.425021 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jspdq\" (UniqueName: \"kubernetes.io/projected/176c2dec-4940-449b-ab1a-95cb11cdfeff-kube-api-access-jspdq\") pod \"placement-769746444-8zhp8\" (UID: \"176c2dec-4940-449b-ab1a-95cb11cdfeff\") " pod="openstack/placement-769746444-8zhp8" Nov 28 13:38:15 crc kubenswrapper[4631]: I1128 13:38:15.436934 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2adae216-4db3-49d7-8d11-24bbd7d47f8f-scripts\") pod \"keystone-6bd6cfc878-65bg5\" (UID: \"2adae216-4db3-49d7-8d11-24bbd7d47f8f\") " pod="openstack/keystone-6bd6cfc878-65bg5" Nov 28 13:38:15 crc kubenswrapper[4631]: I1128 13:38:15.437004 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/2adae216-4db3-49d7-8d11-24bbd7d47f8f-credential-keys\") pod \"keystone-6bd6cfc878-65bg5\" (UID: \"2adae216-4db3-49d7-8d11-24bbd7d47f8f\") " pod="openstack/keystone-6bd6cfc878-65bg5" Nov 28 13:38:15 crc kubenswrapper[4631]: I1128 13:38:15.437044 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2adae216-4db3-49d7-8d11-24bbd7d47f8f-internal-tls-certs\") pod \"keystone-6bd6cfc878-65bg5\" (UID: \"2adae216-4db3-49d7-8d11-24bbd7d47f8f\") " pod="openstack/keystone-6bd6cfc878-65bg5" Nov 28 13:38:15 crc kubenswrapper[4631]: I1128 13:38:15.437084 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2adae216-4db3-49d7-8d11-24bbd7d47f8f-config-data\") pod \"keystone-6bd6cfc878-65bg5\" (UID: \"2adae216-4db3-49d7-8d11-24bbd7d47f8f\") " pod="openstack/keystone-6bd6cfc878-65bg5" Nov 28 13:38:15 crc kubenswrapper[4631]: I1128 13:38:15.437139 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-btg7k\" (UniqueName: \"kubernetes.io/projected/2adae216-4db3-49d7-8d11-24bbd7d47f8f-kube-api-access-btg7k\") pod \"keystone-6bd6cfc878-65bg5\" (UID: \"2adae216-4db3-49d7-8d11-24bbd7d47f8f\") " pod="openstack/keystone-6bd6cfc878-65bg5" Nov 28 13:38:15 crc kubenswrapper[4631]: I1128 13:38:15.437155 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/2adae216-4db3-49d7-8d11-24bbd7d47f8f-fernet-keys\") pod \"keystone-6bd6cfc878-65bg5\" (UID: \"2adae216-4db3-49d7-8d11-24bbd7d47f8f\") " pod="openstack/keystone-6bd6cfc878-65bg5" Nov 28 13:38:15 crc kubenswrapper[4631]: I1128 13:38:15.437173 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2adae216-4db3-49d7-8d11-24bbd7d47f8f-public-tls-certs\") pod \"keystone-6bd6cfc878-65bg5\" (UID: \"2adae216-4db3-49d7-8d11-24bbd7d47f8f\") " pod="openstack/keystone-6bd6cfc878-65bg5" Nov 28 13:38:15 crc kubenswrapper[4631]: I1128 13:38:15.437211 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2adae216-4db3-49d7-8d11-24bbd7d47f8f-combined-ca-bundle\") pod \"keystone-6bd6cfc878-65bg5\" (UID: \"2adae216-4db3-49d7-8d11-24bbd7d47f8f\") " pod="openstack/keystone-6bd6cfc878-65bg5" Nov 28 13:38:15 crc kubenswrapper[4631]: I1128 13:38:15.470065 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/2adae216-4db3-49d7-8d11-24bbd7d47f8f-credential-keys\") pod \"keystone-6bd6cfc878-65bg5\" (UID: \"2adae216-4db3-49d7-8d11-24bbd7d47f8f\") " pod="openstack/keystone-6bd6cfc878-65bg5" Nov 28 13:38:15 crc kubenswrapper[4631]: I1128 13:38:15.470777 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2adae216-4db3-49d7-8d11-24bbd7d47f8f-internal-tls-certs\") pod \"keystone-6bd6cfc878-65bg5\" (UID: \"2adae216-4db3-49d7-8d11-24bbd7d47f8f\") " pod="openstack/keystone-6bd6cfc878-65bg5" Nov 28 13:38:15 crc kubenswrapper[4631]: I1128 13:38:15.471222 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2adae216-4db3-49d7-8d11-24bbd7d47f8f-public-tls-certs\") pod \"keystone-6bd6cfc878-65bg5\" (UID: \"2adae216-4db3-49d7-8d11-24bbd7d47f8f\") " pod="openstack/keystone-6bd6cfc878-65bg5" Nov 28 13:38:15 crc kubenswrapper[4631]: I1128 13:38:15.471727 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/2adae216-4db3-49d7-8d11-24bbd7d47f8f-fernet-keys\") pod \"keystone-6bd6cfc878-65bg5\" (UID: \"2adae216-4db3-49d7-8d11-24bbd7d47f8f\") " pod="openstack/keystone-6bd6cfc878-65bg5" Nov 28 13:38:15 crc kubenswrapper[4631]: I1128 13:38:15.472195 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2adae216-4db3-49d7-8d11-24bbd7d47f8f-combined-ca-bundle\") pod \"keystone-6bd6cfc878-65bg5\" (UID: \"2adae216-4db3-49d7-8d11-24bbd7d47f8f\") " pod="openstack/keystone-6bd6cfc878-65bg5" Nov 28 13:38:15 crc kubenswrapper[4631]: I1128 13:38:15.472464 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2adae216-4db3-49d7-8d11-24bbd7d47f8f-scripts\") pod \"keystone-6bd6cfc878-65bg5\" (UID: \"2adae216-4db3-49d7-8d11-24bbd7d47f8f\") " pod="openstack/keystone-6bd6cfc878-65bg5" Nov 28 13:38:15 crc kubenswrapper[4631]: I1128 13:38:15.500252 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2adae216-4db3-49d7-8d11-24bbd7d47f8f-config-data\") pod \"keystone-6bd6cfc878-65bg5\" (UID: \"2adae216-4db3-49d7-8d11-24bbd7d47f8f\") " pod="openstack/keystone-6bd6cfc878-65bg5" Nov 28 13:38:15 crc kubenswrapper[4631]: I1128 13:38:15.504603 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-btg7k\" (UniqueName: \"kubernetes.io/projected/2adae216-4db3-49d7-8d11-24bbd7d47f8f-kube-api-access-btg7k\") pod \"keystone-6bd6cfc878-65bg5\" (UID: \"2adae216-4db3-49d7-8d11-24bbd7d47f8f\") " pod="openstack/keystone-6bd6cfc878-65bg5" Nov 28 13:38:15 crc kubenswrapper[4631]: I1128 13:38:15.558255 4631 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-544496c768-rzhsw" podUID="cd607dae-568a-4b81-af81-3310c0e95854" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.147:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.147:8443: connect: connection refused" Nov 28 13:38:15 crc kubenswrapper[4631]: I1128 13:38:15.585229 4631 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8d7961e8-77d2-4b92-8758-f06c6d216f73" path="/var/lib/kubelet/pods/8d7961e8-77d2-4b92-8758-f06c6d216f73/volumes" Nov 28 13:38:15 crc kubenswrapper[4631]: I1128 13:38:15.613032 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-6bd6cfc878-65bg5" Nov 28 13:38:15 crc kubenswrapper[4631]: I1128 13:38:15.621065 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-769746444-8zhp8" Nov 28 13:38:16 crc kubenswrapper[4631]: I1128 13:38:16.277153 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-z6tgb" podStartSLOduration=5.395154889 podStartE2EDuration="1m2.277127616s" podCreationTimestamp="2025-11-28 13:37:14 +0000 UTC" firstStartedPulling="2025-11-28 13:37:17.185426049 +0000 UTC m=+993.992729393" lastFinishedPulling="2025-11-28 13:38:14.067398776 +0000 UTC m=+1050.874702120" observedRunningTime="2025-11-28 13:38:15.410967512 +0000 UTC m=+1052.218270856" watchObservedRunningTime="2025-11-28 13:38:16.277127616 +0000 UTC m=+1053.084430960" Nov 28 13:38:16 crc kubenswrapper[4631]: I1128 13:38:16.284775 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-769746444-8zhp8"] Nov 28 13:38:16 crc kubenswrapper[4631]: I1128 13:38:16.460532 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-gprj7" event={"ID":"af357d67-237c-468a-81ac-990964e8e8b2","Type":"ContainerStarted","Data":"e71efaf4f5015e3c6b71bcb5af4c2d9dab05f04b0454ced34b4bd97d9f353272"} Nov 28 13:38:16 crc kubenswrapper[4631]: I1128 13:38:16.469398 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-769746444-8zhp8" event={"ID":"176c2dec-4940-449b-ab1a-95cb11cdfeff","Type":"ContainerStarted","Data":"6f6864d7a7c74d1a6f4d68d33638ca89e6967633407e60be14cb486da76f8f76"} Nov 28 13:38:16 crc kubenswrapper[4631]: I1128 13:38:16.492078 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-gprj7" podStartSLOduration=5.083979241 podStartE2EDuration="1m2.492057637s" podCreationTimestamp="2025-11-28 13:37:14 +0000 UTC" firstStartedPulling="2025-11-28 13:37:16.673882309 +0000 UTC m=+993.481185653" lastFinishedPulling="2025-11-28 13:38:14.081960705 +0000 UTC m=+1050.889264049" observedRunningTime="2025-11-28 13:38:16.481946329 +0000 UTC m=+1053.289249663" watchObservedRunningTime="2025-11-28 13:38:16.492057637 +0000 UTC m=+1053.299360981" Nov 28 13:38:16 crc kubenswrapper[4631]: I1128 13:38:16.579072 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-6bd6cfc878-65bg5"] Nov 28 13:38:16 crc kubenswrapper[4631]: W1128 13:38:16.600121 4631 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2adae216_4db3_49d7_8d11_24bbd7d47f8f.slice/crio-9c683f54a805cbe9687271fceb2c062531da930809011d4933959773bae05679 WatchSource:0}: Error finding container 9c683f54a805cbe9687271fceb2c062531da930809011d4933959773bae05679: Status 404 returned error can't find the container with id 9c683f54a805cbe9687271fceb2c062531da930809011d4933959773bae05679 Nov 28 13:38:17 crc kubenswrapper[4631]: I1128 13:38:17.501514 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-6bd6cfc878-65bg5" event={"ID":"2adae216-4db3-49d7-8d11-24bbd7d47f8f","Type":"ContainerStarted","Data":"c2be83eb6d2f4b055f5d72d94fc91b621bbcecd322a92767bb149ece3b083e48"} Nov 28 13:38:17 crc kubenswrapper[4631]: I1128 13:38:17.502025 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-6bd6cfc878-65bg5" Nov 28 13:38:17 crc kubenswrapper[4631]: I1128 13:38:17.502045 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-6bd6cfc878-65bg5" event={"ID":"2adae216-4db3-49d7-8d11-24bbd7d47f8f","Type":"ContainerStarted","Data":"9c683f54a805cbe9687271fceb2c062531da930809011d4933959773bae05679"} Nov 28 13:38:17 crc kubenswrapper[4631]: I1128 13:38:17.532394 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-6bd6cfc878-65bg5" podStartSLOduration=2.532373761 podStartE2EDuration="2.532373761s" podCreationTimestamp="2025-11-28 13:38:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 13:38:17.529535891 +0000 UTC m=+1054.336839235" watchObservedRunningTime="2025-11-28 13:38:17.532373761 +0000 UTC m=+1054.339677105" Nov 28 13:38:17 crc kubenswrapper[4631]: I1128 13:38:17.565086 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-769746444-8zhp8" event={"ID":"176c2dec-4940-449b-ab1a-95cb11cdfeff","Type":"ContainerStarted","Data":"0b344152ea66a916e2414e10fecec3097af6528fc205793b394329e8b6867251"} Nov 28 13:38:18 crc kubenswrapper[4631]: I1128 13:38:18.585355 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-769746444-8zhp8" event={"ID":"176c2dec-4940-449b-ab1a-95cb11cdfeff","Type":"ContainerStarted","Data":"fe80911df90ae1e7ca6c60d87861082ea2dea837f7d00516e4be4dd1cff1b478"} Nov 28 13:38:18 crc kubenswrapper[4631]: I1128 13:38:18.585853 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-769746444-8zhp8" Nov 28 13:38:18 crc kubenswrapper[4631]: I1128 13:38:18.585870 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-769746444-8zhp8" Nov 28 13:38:18 crc kubenswrapper[4631]: I1128 13:38:18.630896 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-769746444-8zhp8" podStartSLOduration=4.630873544 podStartE2EDuration="4.630873544s" podCreationTimestamp="2025-11-28 13:38:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 13:38:18.612773389 +0000 UTC m=+1055.420076733" watchObservedRunningTime="2025-11-28 13:38:18.630873544 +0000 UTC m=+1055.438176888" Nov 28 13:38:20 crc kubenswrapper[4631]: I1128 13:38:20.612725 4631 generic.go:334] "Generic (PLEG): container finished" podID="3050013d-2db4-4801-8626-1fb6398b714e" containerID="8659b5cf744dfcf5eb7c3313fb4013f0f03ffb7680a0047f33139bfff995c446" exitCode=0 Nov 28 13:38:20 crc kubenswrapper[4631]: I1128 13:38:20.612811 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-z6tgb" event={"ID":"3050013d-2db4-4801-8626-1fb6398b714e","Type":"ContainerDied","Data":"8659b5cf744dfcf5eb7c3313fb4013f0f03ffb7680a0047f33139bfff995c446"} Nov 28 13:38:23 crc kubenswrapper[4631]: I1128 13:38:23.664054 4631 generic.go:334] "Generic (PLEG): container finished" podID="af357d67-237c-468a-81ac-990964e8e8b2" containerID="e71efaf4f5015e3c6b71bcb5af4c2d9dab05f04b0454ced34b4bd97d9f353272" exitCode=0 Nov 28 13:38:23 crc kubenswrapper[4631]: I1128 13:38:23.664437 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-gprj7" event={"ID":"af357d67-237c-468a-81ac-990964e8e8b2","Type":"ContainerDied","Data":"e71efaf4f5015e3c6b71bcb5af4c2d9dab05f04b0454ced34b4bd97d9f353272"} Nov 28 13:38:23 crc kubenswrapper[4631]: I1128 13:38:23.859172 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-5fb67d6644-dbmzf" Nov 28 13:38:24 crc kubenswrapper[4631]: I1128 13:38:24.685393 4631 generic.go:334] "Generic (PLEG): container finished" podID="c3bcfdda-21af-4504-806c-50e6bc0f8770" containerID="1ac446f19faea84032fc117fe1a86106333262fc6b9f72897a52070eea7a431c" exitCode=137 Nov 28 13:38:24 crc kubenswrapper[4631]: I1128 13:38:24.688053 4631 generic.go:334] "Generic (PLEG): container finished" podID="c3bcfdda-21af-4504-806c-50e6bc0f8770" containerID="f15628a29d87012b07a748a7da86e986a7407dc5babc30b157d6e5e5ac4c694c" exitCode=137 Nov 28 13:38:24 crc kubenswrapper[4631]: I1128 13:38:24.688006 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6d58995c-nzh9k" event={"ID":"c3bcfdda-21af-4504-806c-50e6bc0f8770","Type":"ContainerDied","Data":"1ac446f19faea84032fc117fe1a86106333262fc6b9f72897a52070eea7a431c"} Nov 28 13:38:24 crc kubenswrapper[4631]: I1128 13:38:24.688481 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6d58995c-nzh9k" event={"ID":"c3bcfdda-21af-4504-806c-50e6bc0f8770","Type":"ContainerDied","Data":"f15628a29d87012b07a748a7da86e986a7407dc5babc30b157d6e5e5ac4c694c"} Nov 28 13:38:25 crc kubenswrapper[4631]: I1128 13:38:25.332330 4631 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-64f8fdf97b-jn5mg" podUID="111c5349-528d-4856-a3e4-1277e3b10889" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.146:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.146:8443: connect: connection refused" Nov 28 13:38:25 crc kubenswrapper[4631]: I1128 13:38:25.332717 4631 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-64f8fdf97b-jn5mg" Nov 28 13:38:25 crc kubenswrapper[4631]: I1128 13:38:25.333474 4631 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="horizon" containerStatusID={"Type":"cri-o","ID":"ffafadc828320f83436d72910a3912cda7f6bb048d8c4c01528f4529bf55b772"} pod="openstack/horizon-64f8fdf97b-jn5mg" containerMessage="Container horizon failed startup probe, will be restarted" Nov 28 13:38:25 crc kubenswrapper[4631]: I1128 13:38:25.333509 4631 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-64f8fdf97b-jn5mg" podUID="111c5349-528d-4856-a3e4-1277e3b10889" containerName="horizon" containerID="cri-o://ffafadc828320f83436d72910a3912cda7f6bb048d8c4c01528f4529bf55b772" gracePeriod=30 Nov 28 13:38:25 crc kubenswrapper[4631]: I1128 13:38:25.554128 4631 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-544496c768-rzhsw" podUID="cd607dae-568a-4b81-af81-3310c0e95854" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.147:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.147:8443: connect: connection refused" Nov 28 13:38:25 crc kubenswrapper[4631]: I1128 13:38:25.554233 4631 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-544496c768-rzhsw" Nov 28 13:38:25 crc kubenswrapper[4631]: I1128 13:38:25.555114 4631 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="horizon" containerStatusID={"Type":"cri-o","ID":"bd4617d23aba6bbc47167ded3acd30280726cfe208e8bfc920d75fb684ec48e8"} pod="openstack/horizon-544496c768-rzhsw" containerMessage="Container horizon failed startup probe, will be restarted" Nov 28 13:38:25 crc kubenswrapper[4631]: I1128 13:38:25.555158 4631 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-544496c768-rzhsw" podUID="cd607dae-568a-4b81-af81-3310c0e95854" containerName="horizon" containerID="cri-o://bd4617d23aba6bbc47167ded3acd30280726cfe208e8bfc920d75fb684ec48e8" gracePeriod=30 Nov 28 13:38:27 crc kubenswrapper[4631]: I1128 13:38:27.057180 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-gprj7" Nov 28 13:38:27 crc kubenswrapper[4631]: I1128 13:38:27.075656 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-z6tgb" Nov 28 13:38:27 crc kubenswrapper[4631]: I1128 13:38:27.108728 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/af357d67-237c-468a-81ac-990964e8e8b2-etc-machine-id\") pod \"af357d67-237c-468a-81ac-990964e8e8b2\" (UID: \"af357d67-237c-468a-81ac-990964e8e8b2\") " Nov 28 13:38:27 crc kubenswrapper[4631]: I1128 13:38:27.109074 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/af357d67-237c-468a-81ac-990964e8e8b2-combined-ca-bundle\") pod \"af357d67-237c-468a-81ac-990964e8e8b2\" (UID: \"af357d67-237c-468a-81ac-990964e8e8b2\") " Nov 28 13:38:27 crc kubenswrapper[4631]: I1128 13:38:27.109114 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/af357d67-237c-468a-81ac-990964e8e8b2-config-data\") pod \"af357d67-237c-468a-81ac-990964e8e8b2\" (UID: \"af357d67-237c-468a-81ac-990964e8e8b2\") " Nov 28 13:38:27 crc kubenswrapper[4631]: I1128 13:38:27.109138 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/af357d67-237c-468a-81ac-990964e8e8b2-scripts\") pod \"af357d67-237c-468a-81ac-990964e8e8b2\" (UID: \"af357d67-237c-468a-81ac-990964e8e8b2\") " Nov 28 13:38:27 crc kubenswrapper[4631]: I1128 13:38:27.109182 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lrt64\" (UniqueName: \"kubernetes.io/projected/af357d67-237c-468a-81ac-990964e8e8b2-kube-api-access-lrt64\") pod \"af357d67-237c-468a-81ac-990964e8e8b2\" (UID: \"af357d67-237c-468a-81ac-990964e8e8b2\") " Nov 28 13:38:27 crc kubenswrapper[4631]: I1128 13:38:27.109211 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/af357d67-237c-468a-81ac-990964e8e8b2-db-sync-config-data\") pod \"af357d67-237c-468a-81ac-990964e8e8b2\" (UID: \"af357d67-237c-468a-81ac-990964e8e8b2\") " Nov 28 13:38:27 crc kubenswrapper[4631]: I1128 13:38:27.122385 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/af357d67-237c-468a-81ac-990964e8e8b2-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "af357d67-237c-468a-81ac-990964e8e8b2" (UID: "af357d67-237c-468a-81ac-990964e8e8b2"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 13:38:27 crc kubenswrapper[4631]: I1128 13:38:27.124778 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/af357d67-237c-468a-81ac-990964e8e8b2-scripts" (OuterVolumeSpecName: "scripts") pod "af357d67-237c-468a-81ac-990964e8e8b2" (UID: "af357d67-237c-468a-81ac-990964e8e8b2"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:38:27 crc kubenswrapper[4631]: I1128 13:38:27.133543 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/af357d67-237c-468a-81ac-990964e8e8b2-kube-api-access-lrt64" (OuterVolumeSpecName: "kube-api-access-lrt64") pod "af357d67-237c-468a-81ac-990964e8e8b2" (UID: "af357d67-237c-468a-81ac-990964e8e8b2"). InnerVolumeSpecName "kube-api-access-lrt64". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:38:27 crc kubenswrapper[4631]: I1128 13:38:27.146919 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/af357d67-237c-468a-81ac-990964e8e8b2-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "af357d67-237c-468a-81ac-990964e8e8b2" (UID: "af357d67-237c-468a-81ac-990964e8e8b2"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:38:27 crc kubenswrapper[4631]: I1128 13:38:27.214495 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3050013d-2db4-4801-8626-1fb6398b714e-combined-ca-bundle\") pod \"3050013d-2db4-4801-8626-1fb6398b714e\" (UID: \"3050013d-2db4-4801-8626-1fb6398b714e\") " Nov 28 13:38:27 crc kubenswrapper[4631]: I1128 13:38:27.215611 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/3050013d-2db4-4801-8626-1fb6398b714e-db-sync-config-data\") pod \"3050013d-2db4-4801-8626-1fb6398b714e\" (UID: \"3050013d-2db4-4801-8626-1fb6398b714e\") " Nov 28 13:38:27 crc kubenswrapper[4631]: I1128 13:38:27.221431 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qjbrq\" (UniqueName: \"kubernetes.io/projected/3050013d-2db4-4801-8626-1fb6398b714e-kube-api-access-qjbrq\") pod \"3050013d-2db4-4801-8626-1fb6398b714e\" (UID: \"3050013d-2db4-4801-8626-1fb6398b714e\") " Nov 28 13:38:27 crc kubenswrapper[4631]: I1128 13:38:27.222487 4631 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/af357d67-237c-468a-81ac-990964e8e8b2-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 13:38:27 crc kubenswrapper[4631]: I1128 13:38:27.222529 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lrt64\" (UniqueName: \"kubernetes.io/projected/af357d67-237c-468a-81ac-990964e8e8b2-kube-api-access-lrt64\") on node \"crc\" DevicePath \"\"" Nov 28 13:38:27 crc kubenswrapper[4631]: I1128 13:38:27.222546 4631 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/af357d67-237c-468a-81ac-990964e8e8b2-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 13:38:27 crc kubenswrapper[4631]: I1128 13:38:27.222561 4631 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/af357d67-237c-468a-81ac-990964e8e8b2-etc-machine-id\") on node \"crc\" DevicePath \"\"" Nov 28 13:38:27 crc kubenswrapper[4631]: I1128 13:38:27.247265 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/af357d67-237c-468a-81ac-990964e8e8b2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "af357d67-237c-468a-81ac-990964e8e8b2" (UID: "af357d67-237c-468a-81ac-990964e8e8b2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:38:27 crc kubenswrapper[4631]: I1128 13:38:27.247457 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3050013d-2db4-4801-8626-1fb6398b714e-kube-api-access-qjbrq" (OuterVolumeSpecName: "kube-api-access-qjbrq") pod "3050013d-2db4-4801-8626-1fb6398b714e" (UID: "3050013d-2db4-4801-8626-1fb6398b714e"). InnerVolumeSpecName "kube-api-access-qjbrq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:38:27 crc kubenswrapper[4631]: I1128 13:38:27.256226 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/af357d67-237c-468a-81ac-990964e8e8b2-config-data" (OuterVolumeSpecName: "config-data") pod "af357d67-237c-468a-81ac-990964e8e8b2" (UID: "af357d67-237c-468a-81ac-990964e8e8b2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:38:27 crc kubenswrapper[4631]: I1128 13:38:27.258318 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3050013d-2db4-4801-8626-1fb6398b714e-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "3050013d-2db4-4801-8626-1fb6398b714e" (UID: "3050013d-2db4-4801-8626-1fb6398b714e"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:38:27 crc kubenswrapper[4631]: I1128 13:38:27.289368 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3050013d-2db4-4801-8626-1fb6398b714e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3050013d-2db4-4801-8626-1fb6398b714e" (UID: "3050013d-2db4-4801-8626-1fb6398b714e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:38:27 crc kubenswrapper[4631]: I1128 13:38:27.325735 4631 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/3050013d-2db4-4801-8626-1fb6398b714e-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 13:38:27 crc kubenswrapper[4631]: I1128 13:38:27.325778 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qjbrq\" (UniqueName: \"kubernetes.io/projected/3050013d-2db4-4801-8626-1fb6398b714e-kube-api-access-qjbrq\") on node \"crc\" DevicePath \"\"" Nov 28 13:38:27 crc kubenswrapper[4631]: I1128 13:38:27.325790 4631 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3050013d-2db4-4801-8626-1fb6398b714e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 13:38:27 crc kubenswrapper[4631]: I1128 13:38:27.325799 4631 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/af357d67-237c-468a-81ac-990964e8e8b2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 13:38:27 crc kubenswrapper[4631]: I1128 13:38:27.325810 4631 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/af357d67-237c-468a-81ac-990964e8e8b2-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 13:38:27 crc kubenswrapper[4631]: I1128 13:38:27.555422 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-7f7bb9bf8f-42ssw" Nov 28 13:38:27 crc kubenswrapper[4631]: I1128 13:38:27.644309 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-5fb67d6644-dbmzf"] Nov 28 13:38:27 crc kubenswrapper[4631]: I1128 13:38:27.644577 4631 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-5fb67d6644-dbmzf" podUID="6a0a46ba-e48f-45a6-801e-7cb0ececabee" containerName="neutron-api" containerID="cri-o://821a3f2847ee539065f14edad868b3980fc7ccbc996dbb6dbf3a771daf646f34" gracePeriod=30 Nov 28 13:38:27 crc kubenswrapper[4631]: I1128 13:38:27.645044 4631 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-5fb67d6644-dbmzf" podUID="6a0a46ba-e48f-45a6-801e-7cb0ececabee" containerName="neutron-httpd" containerID="cri-o://42641fee468bc31b2584b9a1325c1fdc14069913633b3c778d6d1079d6e2d587" gracePeriod=30 Nov 28 13:38:27 crc kubenswrapper[4631]: I1128 13:38:27.765196 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-gprj7" event={"ID":"af357d67-237c-468a-81ac-990964e8e8b2","Type":"ContainerDied","Data":"7a93010954a947c18f242d29b0ac230e7ee364211ff12c2bc539d2f8e230404d"} Nov 28 13:38:27 crc kubenswrapper[4631]: I1128 13:38:27.765240 4631 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7a93010954a947c18f242d29b0ac230e7ee364211ff12c2bc539d2f8e230404d" Nov 28 13:38:27 crc kubenswrapper[4631]: I1128 13:38:27.765334 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-gprj7" Nov 28 13:38:27 crc kubenswrapper[4631]: I1128 13:38:27.768276 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-z6tgb" event={"ID":"3050013d-2db4-4801-8626-1fb6398b714e","Type":"ContainerDied","Data":"2a49f5b0bb46e154c3c017389154cdcab6938ea4fe61c8d06e0c4be2c2069ad8"} Nov 28 13:38:27 crc kubenswrapper[4631]: I1128 13:38:27.768348 4631 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2a49f5b0bb46e154c3c017389154cdcab6938ea4fe61c8d06e0c4be2c2069ad8" Nov 28 13:38:27 crc kubenswrapper[4631]: I1128 13:38:27.768428 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-z6tgb" Nov 28 13:38:28 crc kubenswrapper[4631]: I1128 13:38:28.023779 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6d58995c-nzh9k" Nov 28 13:38:28 crc kubenswrapper[4631]: I1128 13:38:28.160335 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c3bcfdda-21af-4504-806c-50e6bc0f8770-scripts\") pod \"c3bcfdda-21af-4504-806c-50e6bc0f8770\" (UID: \"c3bcfdda-21af-4504-806c-50e6bc0f8770\") " Nov 28 13:38:28 crc kubenswrapper[4631]: I1128 13:38:28.161455 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/c3bcfdda-21af-4504-806c-50e6bc0f8770-horizon-secret-key\") pod \"c3bcfdda-21af-4504-806c-50e6bc0f8770\" (UID: \"c3bcfdda-21af-4504-806c-50e6bc0f8770\") " Nov 28 13:38:28 crc kubenswrapper[4631]: I1128 13:38:28.165444 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c3bcfdda-21af-4504-806c-50e6bc0f8770-logs\") pod \"c3bcfdda-21af-4504-806c-50e6bc0f8770\" (UID: \"c3bcfdda-21af-4504-806c-50e6bc0f8770\") " Nov 28 13:38:28 crc kubenswrapper[4631]: I1128 13:38:28.165609 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ndw5r\" (UniqueName: \"kubernetes.io/projected/c3bcfdda-21af-4504-806c-50e6bc0f8770-kube-api-access-ndw5r\") pod \"c3bcfdda-21af-4504-806c-50e6bc0f8770\" (UID: \"c3bcfdda-21af-4504-806c-50e6bc0f8770\") " Nov 28 13:38:28 crc kubenswrapper[4631]: I1128 13:38:28.165980 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c3bcfdda-21af-4504-806c-50e6bc0f8770-config-data\") pod \"c3bcfdda-21af-4504-806c-50e6bc0f8770\" (UID: \"c3bcfdda-21af-4504-806c-50e6bc0f8770\") " Nov 28 13:38:28 crc kubenswrapper[4631]: I1128 13:38:28.168637 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c3bcfdda-21af-4504-806c-50e6bc0f8770-logs" (OuterVolumeSpecName: "logs") pod "c3bcfdda-21af-4504-806c-50e6bc0f8770" (UID: "c3bcfdda-21af-4504-806c-50e6bc0f8770"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 13:38:28 crc kubenswrapper[4631]: I1128 13:38:28.176772 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c3bcfdda-21af-4504-806c-50e6bc0f8770-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "c3bcfdda-21af-4504-806c-50e6bc0f8770" (UID: "c3bcfdda-21af-4504-806c-50e6bc0f8770"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:38:28 crc kubenswrapper[4631]: I1128 13:38:28.190623 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c3bcfdda-21af-4504-806c-50e6bc0f8770-kube-api-access-ndw5r" (OuterVolumeSpecName: "kube-api-access-ndw5r") pod "c3bcfdda-21af-4504-806c-50e6bc0f8770" (UID: "c3bcfdda-21af-4504-806c-50e6bc0f8770"). InnerVolumeSpecName "kube-api-access-ndw5r". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:38:28 crc kubenswrapper[4631]: I1128 13:38:28.247335 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c3bcfdda-21af-4504-806c-50e6bc0f8770-scripts" (OuterVolumeSpecName: "scripts") pod "c3bcfdda-21af-4504-806c-50e6bc0f8770" (UID: "c3bcfdda-21af-4504-806c-50e6bc0f8770"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 13:38:28 crc kubenswrapper[4631]: I1128 13:38:28.249537 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c3bcfdda-21af-4504-806c-50e6bc0f8770-config-data" (OuterVolumeSpecName: "config-data") pod "c3bcfdda-21af-4504-806c-50e6bc0f8770" (UID: "c3bcfdda-21af-4504-806c-50e6bc0f8770"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 13:38:28 crc kubenswrapper[4631]: I1128 13:38:28.290939 4631 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/c3bcfdda-21af-4504-806c-50e6bc0f8770-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Nov 28 13:38:28 crc kubenswrapper[4631]: I1128 13:38:28.291317 4631 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c3bcfdda-21af-4504-806c-50e6bc0f8770-logs\") on node \"crc\" DevicePath \"\"" Nov 28 13:38:28 crc kubenswrapper[4631]: I1128 13:38:28.291400 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ndw5r\" (UniqueName: \"kubernetes.io/projected/c3bcfdda-21af-4504-806c-50e6bc0f8770-kube-api-access-ndw5r\") on node \"crc\" DevicePath \"\"" Nov 28 13:38:28 crc kubenswrapper[4631]: I1128 13:38:28.291467 4631 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c3bcfdda-21af-4504-806c-50e6bc0f8770-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 13:38:28 crc kubenswrapper[4631]: I1128 13:38:28.291539 4631 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c3bcfdda-21af-4504-806c-50e6bc0f8770-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 13:38:28 crc kubenswrapper[4631]: I1128 13:38:28.787563 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-59c4c4dc74-qt87t"] Nov 28 13:38:28 crc kubenswrapper[4631]: E1128 13:38:28.790683 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c3bcfdda-21af-4504-806c-50e6bc0f8770" containerName="horizon" Nov 28 13:38:28 crc kubenswrapper[4631]: I1128 13:38:28.790827 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="c3bcfdda-21af-4504-806c-50e6bc0f8770" containerName="horizon" Nov 28 13:38:28 crc kubenswrapper[4631]: E1128 13:38:28.790907 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af357d67-237c-468a-81ac-990964e8e8b2" containerName="cinder-db-sync" Nov 28 13:38:28 crc kubenswrapper[4631]: I1128 13:38:28.790975 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="af357d67-237c-468a-81ac-990964e8e8b2" containerName="cinder-db-sync" Nov 28 13:38:28 crc kubenswrapper[4631]: E1128 13:38:28.791067 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c3bcfdda-21af-4504-806c-50e6bc0f8770" containerName="horizon-log" Nov 28 13:38:28 crc kubenswrapper[4631]: I1128 13:38:28.791136 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="c3bcfdda-21af-4504-806c-50e6bc0f8770" containerName="horizon-log" Nov 28 13:38:28 crc kubenswrapper[4631]: E1128 13:38:28.791189 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3050013d-2db4-4801-8626-1fb6398b714e" containerName="barbican-db-sync" Nov 28 13:38:28 crc kubenswrapper[4631]: I1128 13:38:28.791244 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="3050013d-2db4-4801-8626-1fb6398b714e" containerName="barbican-db-sync" Nov 28 13:38:28 crc kubenswrapper[4631]: I1128 13:38:28.791558 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="c3bcfdda-21af-4504-806c-50e6bc0f8770" containerName="horizon" Nov 28 13:38:28 crc kubenswrapper[4631]: I1128 13:38:28.791659 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="c3bcfdda-21af-4504-806c-50e6bc0f8770" containerName="horizon-log" Nov 28 13:38:28 crc kubenswrapper[4631]: I1128 13:38:28.791746 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="3050013d-2db4-4801-8626-1fb6398b714e" containerName="barbican-db-sync" Nov 28 13:38:28 crc kubenswrapper[4631]: I1128 13:38:28.791806 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="af357d67-237c-468a-81ac-990964e8e8b2" containerName="cinder-db-sync" Nov 28 13:38:28 crc kubenswrapper[4631]: I1128 13:38:28.791816 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6d58995c-nzh9k" Nov 28 13:38:28 crc kubenswrapper[4631]: I1128 13:38:28.793060 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6d58995c-nzh9k" event={"ID":"c3bcfdda-21af-4504-806c-50e6bc0f8770","Type":"ContainerDied","Data":"04724a4cd88466cc6fc93aaeaed95d494fd770437900d2ca5b18383ea381e17a"} Nov 28 13:38:28 crc kubenswrapper[4631]: I1128 13:38:28.793201 4631 scope.go:117] "RemoveContainer" containerID="1ac446f19faea84032fc117fe1a86106333262fc6b9f72897a52070eea7a431c" Nov 28 13:38:28 crc kubenswrapper[4631]: I1128 13:38:28.793573 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-59c4c4dc74-qt87t" Nov 28 13:38:28 crc kubenswrapper[4631]: I1128 13:38:28.814310 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-5b8f56c4bc-cntw9"] Nov 28 13:38:28 crc kubenswrapper[4631]: I1128 13:38:28.831131 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d95d3536-1392-4576-8ee5-390c09fe7a49","Type":"ContainerStarted","Data":"756cbe01466c40de3773fa6a449d3e360e15094ae83297664fbe80c0bd2d91ba"} Nov 28 13:38:28 crc kubenswrapper[4631]: I1128 13:38:28.834112 4631 generic.go:334] "Generic (PLEG): container finished" podID="6a0a46ba-e48f-45a6-801e-7cb0ececabee" containerID="42641fee468bc31b2584b9a1325c1fdc14069913633b3c778d6d1079d6e2d587" exitCode=0 Nov 28 13:38:28 crc kubenswrapper[4631]: I1128 13:38:28.834156 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5fb67d6644-dbmzf" event={"ID":"6a0a46ba-e48f-45a6-801e-7cb0ececabee","Type":"ContainerDied","Data":"42641fee468bc31b2584b9a1325c1fdc14069913633b3c778d6d1079d6e2d587"} Nov 28 13:38:28 crc kubenswrapper[4631]: I1128 13:38:28.840881 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Nov 28 13:38:28 crc kubenswrapper[4631]: I1128 13:38:28.841211 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-mxvzt" Nov 28 13:38:28 crc kubenswrapper[4631]: I1128 13:38:28.841375 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Nov 28 13:38:28 crc kubenswrapper[4631]: I1128 13:38:28.843432 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-5b8f56c4bc-cntw9" Nov 28 13:38:28 crc kubenswrapper[4631]: I1128 13:38:28.843606 4631 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d95d3536-1392-4576-8ee5-390c09fe7a49" containerName="ceilometer-central-agent" containerID="cri-o://42664776e24655262092eae45a6b6b94e6ed84e5b6614b7c0abe60d7e38df756" gracePeriod=30 Nov 28 13:38:28 crc kubenswrapper[4631]: I1128 13:38:28.843766 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 28 13:38:28 crc kubenswrapper[4631]: I1128 13:38:28.844510 4631 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d95d3536-1392-4576-8ee5-390c09fe7a49" containerName="proxy-httpd" containerID="cri-o://756cbe01466c40de3773fa6a449d3e360e15094ae83297664fbe80c0bd2d91ba" gracePeriod=30 Nov 28 13:38:28 crc kubenswrapper[4631]: I1128 13:38:28.844571 4631 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d95d3536-1392-4576-8ee5-390c09fe7a49" containerName="sg-core" containerID="cri-o://aed54e68e791750f2a7b381406c45fd57ec0e772fa0e75496b255dcade224d18" gracePeriod=30 Nov 28 13:38:28 crc kubenswrapper[4631]: I1128 13:38:28.844609 4631 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d95d3536-1392-4576-8ee5-390c09fe7a49" containerName="ceilometer-notification-agent" containerID="cri-o://2141eebe891186cff2651a86531e3f94d1b78cbeeead6dde52101d7a46457ff5" gracePeriod=30 Nov 28 13:38:28 crc kubenswrapper[4631]: I1128 13:38:28.856744 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Nov 28 13:38:28 crc kubenswrapper[4631]: I1128 13:38:28.871206 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Nov 28 13:38:28 crc kubenswrapper[4631]: I1128 13:38:28.873428 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 28 13:38:28 crc kubenswrapper[4631]: I1128 13:38:28.910960 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1b8b4a3c-2e24-4a9b-b18d-0fa09f70a155-config-data-custom\") pod \"barbican-worker-5b8f56c4bc-cntw9\" (UID: \"1b8b4a3c-2e24-4a9b-b18d-0fa09f70a155\") " pod="openstack/barbican-worker-5b8f56c4bc-cntw9" Nov 28 13:38:28 crc kubenswrapper[4631]: I1128 13:38:28.911039 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/177f3e63-a6ce-4512-80b3-f229e3ace62a-logs\") pod \"barbican-keystone-listener-59c4c4dc74-qt87t\" (UID: \"177f3e63-a6ce-4512-80b3-f229e3ace62a\") " pod="openstack/barbican-keystone-listener-59c4c4dc74-qt87t" Nov 28 13:38:28 crc kubenswrapper[4631]: I1128 13:38:28.911103 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/177f3e63-a6ce-4512-80b3-f229e3ace62a-combined-ca-bundle\") pod \"barbican-keystone-listener-59c4c4dc74-qt87t\" (UID: \"177f3e63-a6ce-4512-80b3-f229e3ace62a\") " pod="openstack/barbican-keystone-listener-59c4c4dc74-qt87t" Nov 28 13:38:28 crc kubenswrapper[4631]: I1128 13:38:28.911179 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tt4jg\" (UniqueName: \"kubernetes.io/projected/1b8b4a3c-2e24-4a9b-b18d-0fa09f70a155-kube-api-access-tt4jg\") pod \"barbican-worker-5b8f56c4bc-cntw9\" (UID: \"1b8b4a3c-2e24-4a9b-b18d-0fa09f70a155\") " pod="openstack/barbican-worker-5b8f56c4bc-cntw9" Nov 28 13:38:28 crc kubenswrapper[4631]: I1128 13:38:28.911206 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1b8b4a3c-2e24-4a9b-b18d-0fa09f70a155-logs\") pod \"barbican-worker-5b8f56c4bc-cntw9\" (UID: \"1b8b4a3c-2e24-4a9b-b18d-0fa09f70a155\") " pod="openstack/barbican-worker-5b8f56c4bc-cntw9" Nov 28 13:38:28 crc kubenswrapper[4631]: I1128 13:38:28.911241 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/177f3e63-a6ce-4512-80b3-f229e3ace62a-config-data-custom\") pod \"barbican-keystone-listener-59c4c4dc74-qt87t\" (UID: \"177f3e63-a6ce-4512-80b3-f229e3ace62a\") " pod="openstack/barbican-keystone-listener-59c4c4dc74-qt87t" Nov 28 13:38:28 crc kubenswrapper[4631]: I1128 13:38:28.911265 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1b8b4a3c-2e24-4a9b-b18d-0fa09f70a155-config-data\") pod \"barbican-worker-5b8f56c4bc-cntw9\" (UID: \"1b8b4a3c-2e24-4a9b-b18d-0fa09f70a155\") " pod="openstack/barbican-worker-5b8f56c4bc-cntw9" Nov 28 13:38:28 crc kubenswrapper[4631]: I1128 13:38:28.911317 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/177f3e63-a6ce-4512-80b3-f229e3ace62a-config-data\") pod \"barbican-keystone-listener-59c4c4dc74-qt87t\" (UID: \"177f3e63-a6ce-4512-80b3-f229e3ace62a\") " pod="openstack/barbican-keystone-listener-59c4c4dc74-qt87t" Nov 28 13:38:28 crc kubenswrapper[4631]: I1128 13:38:28.911335 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kpn5w\" (UniqueName: \"kubernetes.io/projected/177f3e63-a6ce-4512-80b3-f229e3ace62a-kube-api-access-kpn5w\") pod \"barbican-keystone-listener-59c4c4dc74-qt87t\" (UID: \"177f3e63-a6ce-4512-80b3-f229e3ace62a\") " pod="openstack/barbican-keystone-listener-59c4c4dc74-qt87t" Nov 28 13:38:28 crc kubenswrapper[4631]: I1128 13:38:28.927902 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1b8b4a3c-2e24-4a9b-b18d-0fa09f70a155-combined-ca-bundle\") pod \"barbican-worker-5b8f56c4bc-cntw9\" (UID: \"1b8b4a3c-2e24-4a9b-b18d-0fa09f70a155\") " pod="openstack/barbican-worker-5b8f56c4bc-cntw9" Nov 28 13:38:28 crc kubenswrapper[4631]: I1128 13:38:28.932626 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-59c4c4dc74-qt87t"] Nov 28 13:38:28 crc kubenswrapper[4631]: I1128 13:38:28.963661 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Nov 28 13:38:28 crc kubenswrapper[4631]: I1128 13:38:28.963828 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-t2s6l" Nov 28 13:38:28 crc kubenswrapper[4631]: I1128 13:38:28.963942 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Nov 28 13:38:28 crc kubenswrapper[4631]: I1128 13:38:28.964041 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Nov 28 13:38:28 crc kubenswrapper[4631]: I1128 13:38:28.969762 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-5b8f56c4bc-cntw9"] Nov 28 13:38:29 crc kubenswrapper[4631]: I1128 13:38:29.025281 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 28 13:38:29 crc kubenswrapper[4631]: I1128 13:38:29.048195 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8f2270d7-d4a8-4cf5-a0b4-b5c29f4fb817-scripts\") pod \"cinder-scheduler-0\" (UID: \"8f2270d7-d4a8-4cf5-a0b4-b5c29f4fb817\") " pod="openstack/cinder-scheduler-0" Nov 28 13:38:29 crc kubenswrapper[4631]: I1128 13:38:29.048316 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kpkhv\" (UniqueName: \"kubernetes.io/projected/8f2270d7-d4a8-4cf5-a0b4-b5c29f4fb817-kube-api-access-kpkhv\") pod \"cinder-scheduler-0\" (UID: \"8f2270d7-d4a8-4cf5-a0b4-b5c29f4fb817\") " pod="openstack/cinder-scheduler-0" Nov 28 13:38:29 crc kubenswrapper[4631]: I1128 13:38:29.048393 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/177f3e63-a6ce-4512-80b3-f229e3ace62a-combined-ca-bundle\") pod \"barbican-keystone-listener-59c4c4dc74-qt87t\" (UID: \"177f3e63-a6ce-4512-80b3-f229e3ace62a\") " pod="openstack/barbican-keystone-listener-59c4c4dc74-qt87t" Nov 28 13:38:29 crc kubenswrapper[4631]: I1128 13:38:29.048525 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8f2270d7-d4a8-4cf5-a0b4-b5c29f4fb817-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"8f2270d7-d4a8-4cf5-a0b4-b5c29f4fb817\") " pod="openstack/cinder-scheduler-0" Nov 28 13:38:29 crc kubenswrapper[4631]: I1128 13:38:29.048597 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tt4jg\" (UniqueName: \"kubernetes.io/projected/1b8b4a3c-2e24-4a9b-b18d-0fa09f70a155-kube-api-access-tt4jg\") pod \"barbican-worker-5b8f56c4bc-cntw9\" (UID: \"1b8b4a3c-2e24-4a9b-b18d-0fa09f70a155\") " pod="openstack/barbican-worker-5b8f56c4bc-cntw9" Nov 28 13:38:29 crc kubenswrapper[4631]: I1128 13:38:29.048654 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1b8b4a3c-2e24-4a9b-b18d-0fa09f70a155-logs\") pod \"barbican-worker-5b8f56c4bc-cntw9\" (UID: \"1b8b4a3c-2e24-4a9b-b18d-0fa09f70a155\") " pod="openstack/barbican-worker-5b8f56c4bc-cntw9" Nov 28 13:38:29 crc kubenswrapper[4631]: I1128 13:38:29.048680 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/177f3e63-a6ce-4512-80b3-f229e3ace62a-config-data-custom\") pod \"barbican-keystone-listener-59c4c4dc74-qt87t\" (UID: \"177f3e63-a6ce-4512-80b3-f229e3ace62a\") " pod="openstack/barbican-keystone-listener-59c4c4dc74-qt87t" Nov 28 13:38:29 crc kubenswrapper[4631]: I1128 13:38:29.048739 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1b8b4a3c-2e24-4a9b-b18d-0fa09f70a155-config-data\") pod \"barbican-worker-5b8f56c4bc-cntw9\" (UID: \"1b8b4a3c-2e24-4a9b-b18d-0fa09f70a155\") " pod="openstack/barbican-worker-5b8f56c4bc-cntw9" Nov 28 13:38:29 crc kubenswrapper[4631]: I1128 13:38:29.048829 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/177f3e63-a6ce-4512-80b3-f229e3ace62a-config-data\") pod \"barbican-keystone-listener-59c4c4dc74-qt87t\" (UID: \"177f3e63-a6ce-4512-80b3-f229e3ace62a\") " pod="openstack/barbican-keystone-listener-59c4c4dc74-qt87t" Nov 28 13:38:29 crc kubenswrapper[4631]: I1128 13:38:29.048853 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kpn5w\" (UniqueName: \"kubernetes.io/projected/177f3e63-a6ce-4512-80b3-f229e3ace62a-kube-api-access-kpn5w\") pod \"barbican-keystone-listener-59c4c4dc74-qt87t\" (UID: \"177f3e63-a6ce-4512-80b3-f229e3ace62a\") " pod="openstack/barbican-keystone-listener-59c4c4dc74-qt87t" Nov 28 13:38:29 crc kubenswrapper[4631]: I1128 13:38:29.048915 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1b8b4a3c-2e24-4a9b-b18d-0fa09f70a155-combined-ca-bundle\") pod \"barbican-worker-5b8f56c4bc-cntw9\" (UID: \"1b8b4a3c-2e24-4a9b-b18d-0fa09f70a155\") " pod="openstack/barbican-worker-5b8f56c4bc-cntw9" Nov 28 13:38:29 crc kubenswrapper[4631]: I1128 13:38:29.048960 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8f2270d7-d4a8-4cf5-a0b4-b5c29f4fb817-config-data\") pod \"cinder-scheduler-0\" (UID: \"8f2270d7-d4a8-4cf5-a0b4-b5c29f4fb817\") " pod="openstack/cinder-scheduler-0" Nov 28 13:38:29 crc kubenswrapper[4631]: I1128 13:38:29.049002 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8f2270d7-d4a8-4cf5-a0b4-b5c29f4fb817-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"8f2270d7-d4a8-4cf5-a0b4-b5c29f4fb817\") " pod="openstack/cinder-scheduler-0" Nov 28 13:38:29 crc kubenswrapper[4631]: I1128 13:38:29.049042 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8f2270d7-d4a8-4cf5-a0b4-b5c29f4fb817-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"8f2270d7-d4a8-4cf5-a0b4-b5c29f4fb817\") " pod="openstack/cinder-scheduler-0" Nov 28 13:38:29 crc kubenswrapper[4631]: I1128 13:38:29.049082 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1b8b4a3c-2e24-4a9b-b18d-0fa09f70a155-config-data-custom\") pod \"barbican-worker-5b8f56c4bc-cntw9\" (UID: \"1b8b4a3c-2e24-4a9b-b18d-0fa09f70a155\") " pod="openstack/barbican-worker-5b8f56c4bc-cntw9" Nov 28 13:38:29 crc kubenswrapper[4631]: I1128 13:38:29.049184 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/177f3e63-a6ce-4512-80b3-f229e3ace62a-logs\") pod \"barbican-keystone-listener-59c4c4dc74-qt87t\" (UID: \"177f3e63-a6ce-4512-80b3-f229e3ace62a\") " pod="openstack/barbican-keystone-listener-59c4c4dc74-qt87t" Nov 28 13:38:29 crc kubenswrapper[4631]: I1128 13:38:29.049818 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/177f3e63-a6ce-4512-80b3-f229e3ace62a-logs\") pod \"barbican-keystone-listener-59c4c4dc74-qt87t\" (UID: \"177f3e63-a6ce-4512-80b3-f229e3ace62a\") " pod="openstack/barbican-keystone-listener-59c4c4dc74-qt87t" Nov 28 13:38:29 crc kubenswrapper[4631]: I1128 13:38:29.084014 4631 scope.go:117] "RemoveContainer" containerID="f15628a29d87012b07a748a7da86e986a7407dc5babc30b157d6e5e5ac4c694c" Nov 28 13:38:29 crc kubenswrapper[4631]: I1128 13:38:29.084654 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1b8b4a3c-2e24-4a9b-b18d-0fa09f70a155-logs\") pod \"barbican-worker-5b8f56c4bc-cntw9\" (UID: \"1b8b4a3c-2e24-4a9b-b18d-0fa09f70a155\") " pod="openstack/barbican-worker-5b8f56c4bc-cntw9" Nov 28 13:38:29 crc kubenswrapper[4631]: I1128 13:38:29.123673 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/177f3e63-a6ce-4512-80b3-f229e3ace62a-config-data-custom\") pod \"barbican-keystone-listener-59c4c4dc74-qt87t\" (UID: \"177f3e63-a6ce-4512-80b3-f229e3ace62a\") " pod="openstack/barbican-keystone-listener-59c4c4dc74-qt87t" Nov 28 13:38:29 crc kubenswrapper[4631]: I1128 13:38:29.173848 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8f2270d7-d4a8-4cf5-a0b4-b5c29f4fb817-config-data\") pod \"cinder-scheduler-0\" (UID: \"8f2270d7-d4a8-4cf5-a0b4-b5c29f4fb817\") " pod="openstack/cinder-scheduler-0" Nov 28 13:38:29 crc kubenswrapper[4631]: I1128 13:38:29.173922 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8f2270d7-d4a8-4cf5-a0b4-b5c29f4fb817-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"8f2270d7-d4a8-4cf5-a0b4-b5c29f4fb817\") " pod="openstack/cinder-scheduler-0" Nov 28 13:38:29 crc kubenswrapper[4631]: I1128 13:38:29.173980 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8f2270d7-d4a8-4cf5-a0b4-b5c29f4fb817-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"8f2270d7-d4a8-4cf5-a0b4-b5c29f4fb817\") " pod="openstack/cinder-scheduler-0" Nov 28 13:38:29 crc kubenswrapper[4631]: I1128 13:38:29.191258 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8f2270d7-d4a8-4cf5-a0b4-b5c29f4fb817-scripts\") pod \"cinder-scheduler-0\" (UID: \"8f2270d7-d4a8-4cf5-a0b4-b5c29f4fb817\") " pod="openstack/cinder-scheduler-0" Nov 28 13:38:29 crc kubenswrapper[4631]: I1128 13:38:29.191413 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kpkhv\" (UniqueName: \"kubernetes.io/projected/8f2270d7-d4a8-4cf5-a0b4-b5c29f4fb817-kube-api-access-kpkhv\") pod \"cinder-scheduler-0\" (UID: \"8f2270d7-d4a8-4cf5-a0b4-b5c29f4fb817\") " pod="openstack/cinder-scheduler-0" Nov 28 13:38:29 crc kubenswrapper[4631]: I1128 13:38:29.191626 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8f2270d7-d4a8-4cf5-a0b4-b5c29f4fb817-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"8f2270d7-d4a8-4cf5-a0b4-b5c29f4fb817\") " pod="openstack/cinder-scheduler-0" Nov 28 13:38:29 crc kubenswrapper[4631]: I1128 13:38:29.192059 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8f2270d7-d4a8-4cf5-a0b4-b5c29f4fb817-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"8f2270d7-d4a8-4cf5-a0b4-b5c29f4fb817\") " pod="openstack/cinder-scheduler-0" Nov 28 13:38:29 crc kubenswrapper[4631]: I1128 13:38:29.212891 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kpn5w\" (UniqueName: \"kubernetes.io/projected/177f3e63-a6ce-4512-80b3-f229e3ace62a-kube-api-access-kpn5w\") pod \"barbican-keystone-listener-59c4c4dc74-qt87t\" (UID: \"177f3e63-a6ce-4512-80b3-f229e3ace62a\") " pod="openstack/barbican-keystone-listener-59c4c4dc74-qt87t" Nov 28 13:38:29 crc kubenswrapper[4631]: I1128 13:38:29.219513 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/177f3e63-a6ce-4512-80b3-f229e3ace62a-config-data\") pod \"barbican-keystone-listener-59c4c4dc74-qt87t\" (UID: \"177f3e63-a6ce-4512-80b3-f229e3ace62a\") " pod="openstack/barbican-keystone-listener-59c4c4dc74-qt87t" Nov 28 13:38:29 crc kubenswrapper[4631]: I1128 13:38:29.220571 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/177f3e63-a6ce-4512-80b3-f229e3ace62a-combined-ca-bundle\") pod \"barbican-keystone-listener-59c4c4dc74-qt87t\" (UID: \"177f3e63-a6ce-4512-80b3-f229e3ace62a\") " pod="openstack/barbican-keystone-listener-59c4c4dc74-qt87t" Nov 28 13:38:29 crc kubenswrapper[4631]: I1128 13:38:29.224048 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1b8b4a3c-2e24-4a9b-b18d-0fa09f70a155-config-data\") pod \"barbican-worker-5b8f56c4bc-cntw9\" (UID: \"1b8b4a3c-2e24-4a9b-b18d-0fa09f70a155\") " pod="openstack/barbican-worker-5b8f56c4bc-cntw9" Nov 28 13:38:29 crc kubenswrapper[4631]: I1128 13:38:29.230844 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1b8b4a3c-2e24-4a9b-b18d-0fa09f70a155-config-data-custom\") pod \"barbican-worker-5b8f56c4bc-cntw9\" (UID: \"1b8b4a3c-2e24-4a9b-b18d-0fa09f70a155\") " pod="openstack/barbican-worker-5b8f56c4bc-cntw9" Nov 28 13:38:29 crc kubenswrapper[4631]: I1128 13:38:29.249349 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1b8b4a3c-2e24-4a9b-b18d-0fa09f70a155-combined-ca-bundle\") pod \"barbican-worker-5b8f56c4bc-cntw9\" (UID: \"1b8b4a3c-2e24-4a9b-b18d-0fa09f70a155\") " pod="openstack/barbican-worker-5b8f56c4bc-cntw9" Nov 28 13:38:29 crc kubenswrapper[4631]: I1128 13:38:29.259272 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8f2270d7-d4a8-4cf5-a0b4-b5c29f4fb817-config-data\") pod \"cinder-scheduler-0\" (UID: \"8f2270d7-d4a8-4cf5-a0b4-b5c29f4fb817\") " pod="openstack/cinder-scheduler-0" Nov 28 13:38:29 crc kubenswrapper[4631]: I1128 13:38:29.291276 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8f2270d7-d4a8-4cf5-a0b4-b5c29f4fb817-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"8f2270d7-d4a8-4cf5-a0b4-b5c29f4fb817\") " pod="openstack/cinder-scheduler-0" Nov 28 13:38:29 crc kubenswrapper[4631]: I1128 13:38:29.292321 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8f2270d7-d4a8-4cf5-a0b4-b5c29f4fb817-scripts\") pod \"cinder-scheduler-0\" (UID: \"8f2270d7-d4a8-4cf5-a0b4-b5c29f4fb817\") " pod="openstack/cinder-scheduler-0" Nov 28 13:38:29 crc kubenswrapper[4631]: I1128 13:38:29.320614 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8f2270d7-d4a8-4cf5-a0b4-b5c29f4fb817-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"8f2270d7-d4a8-4cf5-a0b4-b5c29f4fb817\") " pod="openstack/cinder-scheduler-0" Nov 28 13:38:29 crc kubenswrapper[4631]: I1128 13:38:29.361688 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tt4jg\" (UniqueName: \"kubernetes.io/projected/1b8b4a3c-2e24-4a9b-b18d-0fa09f70a155-kube-api-access-tt4jg\") pod \"barbican-worker-5b8f56c4bc-cntw9\" (UID: \"1b8b4a3c-2e24-4a9b-b18d-0fa09f70a155\") " pod="openstack/barbican-worker-5b8f56c4bc-cntw9" Nov 28 13:38:29 crc kubenswrapper[4631]: I1128 13:38:29.371091 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kpkhv\" (UniqueName: \"kubernetes.io/projected/8f2270d7-d4a8-4cf5-a0b4-b5c29f4fb817-kube-api-access-kpkhv\") pod \"cinder-scheduler-0\" (UID: \"8f2270d7-d4a8-4cf5-a0b4-b5c29f4fb817\") " pod="openstack/cinder-scheduler-0" Nov 28 13:38:29 crc kubenswrapper[4631]: I1128 13:38:29.408786 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-85ff748b95-hhmtl"] Nov 28 13:38:29 crc kubenswrapper[4631]: I1128 13:38:29.423324 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-59c4c4dc74-qt87t" Nov 28 13:38:29 crc kubenswrapper[4631]: I1128 13:38:29.424052 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85ff748b95-hhmtl" Nov 28 13:38:29 crc kubenswrapper[4631]: I1128 13:38:29.435185 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=4.603508976 podStartE2EDuration="1m15.435147305s" podCreationTimestamp="2025-11-28 13:37:14 +0000 UTC" firstStartedPulling="2025-11-28 13:37:17.256752312 +0000 UTC m=+994.064055656" lastFinishedPulling="2025-11-28 13:38:28.088390641 +0000 UTC m=+1064.895693985" observedRunningTime="2025-11-28 13:38:29.264194184 +0000 UTC m=+1066.071497528" watchObservedRunningTime="2025-11-28 13:38:29.435147305 +0000 UTC m=+1066.242450649" Nov 28 13:38:29 crc kubenswrapper[4631]: I1128 13:38:29.450945 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2trlv\" (UniqueName: \"kubernetes.io/projected/abf8d962-b7ff-4e5e-8ccc-e5ffae26f822-kube-api-access-2trlv\") pod \"dnsmasq-dns-85ff748b95-hhmtl\" (UID: \"abf8d962-b7ff-4e5e-8ccc-e5ffae26f822\") " pod="openstack/dnsmasq-dns-85ff748b95-hhmtl" Nov 28 13:38:29 crc kubenswrapper[4631]: I1128 13:38:29.451050 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/abf8d962-b7ff-4e5e-8ccc-e5ffae26f822-ovsdbserver-nb\") pod \"dnsmasq-dns-85ff748b95-hhmtl\" (UID: \"abf8d962-b7ff-4e5e-8ccc-e5ffae26f822\") " pod="openstack/dnsmasq-dns-85ff748b95-hhmtl" Nov 28 13:38:29 crc kubenswrapper[4631]: I1128 13:38:29.451070 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/abf8d962-b7ff-4e5e-8ccc-e5ffae26f822-dns-swift-storage-0\") pod \"dnsmasq-dns-85ff748b95-hhmtl\" (UID: \"abf8d962-b7ff-4e5e-8ccc-e5ffae26f822\") " pod="openstack/dnsmasq-dns-85ff748b95-hhmtl" Nov 28 13:38:29 crc kubenswrapper[4631]: I1128 13:38:29.451094 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/abf8d962-b7ff-4e5e-8ccc-e5ffae26f822-dns-svc\") pod \"dnsmasq-dns-85ff748b95-hhmtl\" (UID: \"abf8d962-b7ff-4e5e-8ccc-e5ffae26f822\") " pod="openstack/dnsmasq-dns-85ff748b95-hhmtl" Nov 28 13:38:29 crc kubenswrapper[4631]: I1128 13:38:29.451218 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/abf8d962-b7ff-4e5e-8ccc-e5ffae26f822-config\") pod \"dnsmasq-dns-85ff748b95-hhmtl\" (UID: \"abf8d962-b7ff-4e5e-8ccc-e5ffae26f822\") " pod="openstack/dnsmasq-dns-85ff748b95-hhmtl" Nov 28 13:38:29 crc kubenswrapper[4631]: I1128 13:38:29.451258 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/abf8d962-b7ff-4e5e-8ccc-e5ffae26f822-ovsdbserver-sb\") pod \"dnsmasq-dns-85ff748b95-hhmtl\" (UID: \"abf8d962-b7ff-4e5e-8ccc-e5ffae26f822\") " pod="openstack/dnsmasq-dns-85ff748b95-hhmtl" Nov 28 13:38:29 crc kubenswrapper[4631]: I1128 13:38:29.479359 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-85ff748b95-hhmtl"] Nov 28 13:38:29 crc kubenswrapper[4631]: I1128 13:38:29.480230 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-5b8f56c4bc-cntw9" Nov 28 13:38:29 crc kubenswrapper[4631]: I1128 13:38:29.513113 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-6d58995c-nzh9k"] Nov 28 13:38:29 crc kubenswrapper[4631]: I1128 13:38:29.542553 4631 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-6d58995c-nzh9k"] Nov 28 13:38:29 crc kubenswrapper[4631]: I1128 13:38:29.553858 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2trlv\" (UniqueName: \"kubernetes.io/projected/abf8d962-b7ff-4e5e-8ccc-e5ffae26f822-kube-api-access-2trlv\") pod \"dnsmasq-dns-85ff748b95-hhmtl\" (UID: \"abf8d962-b7ff-4e5e-8ccc-e5ffae26f822\") " pod="openstack/dnsmasq-dns-85ff748b95-hhmtl" Nov 28 13:38:29 crc kubenswrapper[4631]: I1128 13:38:29.553939 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/abf8d962-b7ff-4e5e-8ccc-e5ffae26f822-ovsdbserver-nb\") pod \"dnsmasq-dns-85ff748b95-hhmtl\" (UID: \"abf8d962-b7ff-4e5e-8ccc-e5ffae26f822\") " pod="openstack/dnsmasq-dns-85ff748b95-hhmtl" Nov 28 13:38:29 crc kubenswrapper[4631]: I1128 13:38:29.553961 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/abf8d962-b7ff-4e5e-8ccc-e5ffae26f822-dns-swift-storage-0\") pod \"dnsmasq-dns-85ff748b95-hhmtl\" (UID: \"abf8d962-b7ff-4e5e-8ccc-e5ffae26f822\") " pod="openstack/dnsmasq-dns-85ff748b95-hhmtl" Nov 28 13:38:29 crc kubenswrapper[4631]: I1128 13:38:29.553986 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/abf8d962-b7ff-4e5e-8ccc-e5ffae26f822-dns-svc\") pod \"dnsmasq-dns-85ff748b95-hhmtl\" (UID: \"abf8d962-b7ff-4e5e-8ccc-e5ffae26f822\") " pod="openstack/dnsmasq-dns-85ff748b95-hhmtl" Nov 28 13:38:29 crc kubenswrapper[4631]: I1128 13:38:29.554057 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/abf8d962-b7ff-4e5e-8ccc-e5ffae26f822-config\") pod \"dnsmasq-dns-85ff748b95-hhmtl\" (UID: \"abf8d962-b7ff-4e5e-8ccc-e5ffae26f822\") " pod="openstack/dnsmasq-dns-85ff748b95-hhmtl" Nov 28 13:38:29 crc kubenswrapper[4631]: I1128 13:38:29.554082 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/abf8d962-b7ff-4e5e-8ccc-e5ffae26f822-ovsdbserver-sb\") pod \"dnsmasq-dns-85ff748b95-hhmtl\" (UID: \"abf8d962-b7ff-4e5e-8ccc-e5ffae26f822\") " pod="openstack/dnsmasq-dns-85ff748b95-hhmtl" Nov 28 13:38:29 crc kubenswrapper[4631]: I1128 13:38:29.570115 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/abf8d962-b7ff-4e5e-8ccc-e5ffae26f822-ovsdbserver-nb\") pod \"dnsmasq-dns-85ff748b95-hhmtl\" (UID: \"abf8d962-b7ff-4e5e-8ccc-e5ffae26f822\") " pod="openstack/dnsmasq-dns-85ff748b95-hhmtl" Nov 28 13:38:29 crc kubenswrapper[4631]: I1128 13:38:29.571097 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/abf8d962-b7ff-4e5e-8ccc-e5ffae26f822-dns-svc\") pod \"dnsmasq-dns-85ff748b95-hhmtl\" (UID: \"abf8d962-b7ff-4e5e-8ccc-e5ffae26f822\") " pod="openstack/dnsmasq-dns-85ff748b95-hhmtl" Nov 28 13:38:29 crc kubenswrapper[4631]: I1128 13:38:29.571217 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/abf8d962-b7ff-4e5e-8ccc-e5ffae26f822-ovsdbserver-sb\") pod \"dnsmasq-dns-85ff748b95-hhmtl\" (UID: \"abf8d962-b7ff-4e5e-8ccc-e5ffae26f822\") " pod="openstack/dnsmasq-dns-85ff748b95-hhmtl" Nov 28 13:38:29 crc kubenswrapper[4631]: I1128 13:38:29.571672 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/abf8d962-b7ff-4e5e-8ccc-e5ffae26f822-config\") pod \"dnsmasq-dns-85ff748b95-hhmtl\" (UID: \"abf8d962-b7ff-4e5e-8ccc-e5ffae26f822\") " pod="openstack/dnsmasq-dns-85ff748b95-hhmtl" Nov 28 13:38:29 crc kubenswrapper[4631]: I1128 13:38:29.571739 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 28 13:38:29 crc kubenswrapper[4631]: I1128 13:38:29.590810 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-85ff748b95-hhmtl"] Nov 28 13:38:29 crc kubenswrapper[4631]: I1128 13:38:29.591419 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/abf8d962-b7ff-4e5e-8ccc-e5ffae26f822-dns-swift-storage-0\") pod \"dnsmasq-dns-85ff748b95-hhmtl\" (UID: \"abf8d962-b7ff-4e5e-8ccc-e5ffae26f822\") " pod="openstack/dnsmasq-dns-85ff748b95-hhmtl" Nov 28 13:38:29 crc kubenswrapper[4631]: E1128 13:38:29.591985 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[dns-swift-storage-0 kube-api-access-2trlv], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/dnsmasq-dns-85ff748b95-hhmtl" podUID="abf8d962-b7ff-4e5e-8ccc-e5ffae26f822" Nov 28 13:38:29 crc kubenswrapper[4631]: I1128 13:38:29.627868 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2trlv\" (UniqueName: \"kubernetes.io/projected/abf8d962-b7ff-4e5e-8ccc-e5ffae26f822-kube-api-access-2trlv\") pod \"dnsmasq-dns-85ff748b95-hhmtl\" (UID: \"abf8d962-b7ff-4e5e-8ccc-e5ffae26f822\") " pod="openstack/dnsmasq-dns-85ff748b95-hhmtl" Nov 28 13:38:29 crc kubenswrapper[4631]: I1128 13:38:29.639859 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-756d64c4c8-jzvlv"] Nov 28 13:38:29 crc kubenswrapper[4631]: I1128 13:38:29.654234 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-756d64c4c8-jzvlv" Nov 28 13:38:29 crc kubenswrapper[4631]: I1128 13:38:29.669921 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Nov 28 13:38:29 crc kubenswrapper[4631]: I1128 13:38:29.696365 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7eea7f3e-fc3c-4c60-998e-79c9a31ee084-combined-ca-bundle\") pod \"barbican-api-756d64c4c8-jzvlv\" (UID: \"7eea7f3e-fc3c-4c60-998e-79c9a31ee084\") " pod="openstack/barbican-api-756d64c4c8-jzvlv" Nov 28 13:38:29 crc kubenswrapper[4631]: I1128 13:38:29.696457 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7eea7f3e-fc3c-4c60-998e-79c9a31ee084-logs\") pod \"barbican-api-756d64c4c8-jzvlv\" (UID: \"7eea7f3e-fc3c-4c60-998e-79c9a31ee084\") " pod="openstack/barbican-api-756d64c4c8-jzvlv" Nov 28 13:38:29 crc kubenswrapper[4631]: I1128 13:38:29.696699 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xsl6r\" (UniqueName: \"kubernetes.io/projected/7eea7f3e-fc3c-4c60-998e-79c9a31ee084-kube-api-access-xsl6r\") pod \"barbican-api-756d64c4c8-jzvlv\" (UID: \"7eea7f3e-fc3c-4c60-998e-79c9a31ee084\") " pod="openstack/barbican-api-756d64c4c8-jzvlv" Nov 28 13:38:29 crc kubenswrapper[4631]: I1128 13:38:29.696768 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7eea7f3e-fc3c-4c60-998e-79c9a31ee084-config-data\") pod \"barbican-api-756d64c4c8-jzvlv\" (UID: \"7eea7f3e-fc3c-4c60-998e-79c9a31ee084\") " pod="openstack/barbican-api-756d64c4c8-jzvlv" Nov 28 13:38:29 crc kubenswrapper[4631]: I1128 13:38:29.696985 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7eea7f3e-fc3c-4c60-998e-79c9a31ee084-config-data-custom\") pod \"barbican-api-756d64c4c8-jzvlv\" (UID: \"7eea7f3e-fc3c-4c60-998e-79c9a31ee084\") " pod="openstack/barbican-api-756d64c4c8-jzvlv" Nov 28 13:38:29 crc kubenswrapper[4631]: I1128 13:38:29.745523 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-756d64c4c8-jzvlv"] Nov 28 13:38:29 crc kubenswrapper[4631]: I1128 13:38:29.792916 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Nov 28 13:38:29 crc kubenswrapper[4631]: I1128 13:38:29.794920 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 28 13:38:29 crc kubenswrapper[4631]: I1128 13:38:29.803315 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Nov 28 13:38:29 crc kubenswrapper[4631]: I1128 13:38:29.805395 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xsl6r\" (UniqueName: \"kubernetes.io/projected/7eea7f3e-fc3c-4c60-998e-79c9a31ee084-kube-api-access-xsl6r\") pod \"barbican-api-756d64c4c8-jzvlv\" (UID: \"7eea7f3e-fc3c-4c60-998e-79c9a31ee084\") " pod="openstack/barbican-api-756d64c4c8-jzvlv" Nov 28 13:38:29 crc kubenswrapper[4631]: I1128 13:38:29.805445 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7eea7f3e-fc3c-4c60-998e-79c9a31ee084-config-data\") pod \"barbican-api-756d64c4c8-jzvlv\" (UID: \"7eea7f3e-fc3c-4c60-998e-79c9a31ee084\") " pod="openstack/barbican-api-756d64c4c8-jzvlv" Nov 28 13:38:29 crc kubenswrapper[4631]: I1128 13:38:29.805507 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7eea7f3e-fc3c-4c60-998e-79c9a31ee084-config-data-custom\") pod \"barbican-api-756d64c4c8-jzvlv\" (UID: \"7eea7f3e-fc3c-4c60-998e-79c9a31ee084\") " pod="openstack/barbican-api-756d64c4c8-jzvlv" Nov 28 13:38:29 crc kubenswrapper[4631]: I1128 13:38:29.805548 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7eea7f3e-fc3c-4c60-998e-79c9a31ee084-combined-ca-bundle\") pod \"barbican-api-756d64c4c8-jzvlv\" (UID: \"7eea7f3e-fc3c-4c60-998e-79c9a31ee084\") " pod="openstack/barbican-api-756d64c4c8-jzvlv" Nov 28 13:38:29 crc kubenswrapper[4631]: I1128 13:38:29.805574 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7eea7f3e-fc3c-4c60-998e-79c9a31ee084-logs\") pod \"barbican-api-756d64c4c8-jzvlv\" (UID: \"7eea7f3e-fc3c-4c60-998e-79c9a31ee084\") " pod="openstack/barbican-api-756d64c4c8-jzvlv" Nov 28 13:38:29 crc kubenswrapper[4631]: I1128 13:38:29.805984 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7eea7f3e-fc3c-4c60-998e-79c9a31ee084-logs\") pod \"barbican-api-756d64c4c8-jzvlv\" (UID: \"7eea7f3e-fc3c-4c60-998e-79c9a31ee084\") " pod="openstack/barbican-api-756d64c4c8-jzvlv" Nov 28 13:38:29 crc kubenswrapper[4631]: I1128 13:38:29.826039 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Nov 28 13:38:29 crc kubenswrapper[4631]: I1128 13:38:29.840136 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7eea7f3e-fc3c-4c60-998e-79c9a31ee084-config-data\") pod \"barbican-api-756d64c4c8-jzvlv\" (UID: \"7eea7f3e-fc3c-4c60-998e-79c9a31ee084\") " pod="openstack/barbican-api-756d64c4c8-jzvlv" Nov 28 13:38:29 crc kubenswrapper[4631]: I1128 13:38:29.841807 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7eea7f3e-fc3c-4c60-998e-79c9a31ee084-combined-ca-bundle\") pod \"barbican-api-756d64c4c8-jzvlv\" (UID: \"7eea7f3e-fc3c-4c60-998e-79c9a31ee084\") " pod="openstack/barbican-api-756d64c4c8-jzvlv" Nov 28 13:38:29 crc kubenswrapper[4631]: I1128 13:38:29.851613 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7eea7f3e-fc3c-4c60-998e-79c9a31ee084-config-data-custom\") pod \"barbican-api-756d64c4c8-jzvlv\" (UID: \"7eea7f3e-fc3c-4c60-998e-79c9a31ee084\") " pod="openstack/barbican-api-756d64c4c8-jzvlv" Nov 28 13:38:29 crc kubenswrapper[4631]: I1128 13:38:29.880133 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-9ctzb"] Nov 28 13:38:29 crc kubenswrapper[4631]: I1128 13:38:29.880505 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xsl6r\" (UniqueName: \"kubernetes.io/projected/7eea7f3e-fc3c-4c60-998e-79c9a31ee084-kube-api-access-xsl6r\") pod \"barbican-api-756d64c4c8-jzvlv\" (UID: \"7eea7f3e-fc3c-4c60-998e-79c9a31ee084\") " pod="openstack/barbican-api-756d64c4c8-jzvlv" Nov 28 13:38:29 crc kubenswrapper[4631]: I1128 13:38:29.882258 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9776ccc5-9ctzb" Nov 28 13:38:29 crc kubenswrapper[4631]: I1128 13:38:29.914983 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x6plr\" (UniqueName: \"kubernetes.io/projected/bef6bb99-5262-4f84-a2f7-0908a080f3a0-kube-api-access-x6plr\") pod \"cinder-api-0\" (UID: \"bef6bb99-5262-4f84-a2f7-0908a080f3a0\") " pod="openstack/cinder-api-0" Nov 28 13:38:29 crc kubenswrapper[4631]: I1128 13:38:29.915266 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/bef6bb99-5262-4f84-a2f7-0908a080f3a0-etc-machine-id\") pod \"cinder-api-0\" (UID: \"bef6bb99-5262-4f84-a2f7-0908a080f3a0\") " pod="openstack/cinder-api-0" Nov 28 13:38:29 crc kubenswrapper[4631]: I1128 13:38:29.915637 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bef6bb99-5262-4f84-a2f7-0908a080f3a0-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"bef6bb99-5262-4f84-a2f7-0908a080f3a0\") " pod="openstack/cinder-api-0" Nov 28 13:38:29 crc kubenswrapper[4631]: I1128 13:38:29.915787 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/bef6bb99-5262-4f84-a2f7-0908a080f3a0-config-data-custom\") pod \"cinder-api-0\" (UID: \"bef6bb99-5262-4f84-a2f7-0908a080f3a0\") " pod="openstack/cinder-api-0" Nov 28 13:38:29 crc kubenswrapper[4631]: I1128 13:38:29.915913 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bef6bb99-5262-4f84-a2f7-0908a080f3a0-config-data\") pod \"cinder-api-0\" (UID: \"bef6bb99-5262-4f84-a2f7-0908a080f3a0\") " pod="openstack/cinder-api-0" Nov 28 13:38:29 crc kubenswrapper[4631]: I1128 13:38:29.916121 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bef6bb99-5262-4f84-a2f7-0908a080f3a0-logs\") pod \"cinder-api-0\" (UID: \"bef6bb99-5262-4f84-a2f7-0908a080f3a0\") " pod="openstack/cinder-api-0" Nov 28 13:38:29 crc kubenswrapper[4631]: I1128 13:38:29.916213 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bef6bb99-5262-4f84-a2f7-0908a080f3a0-scripts\") pod \"cinder-api-0\" (UID: \"bef6bb99-5262-4f84-a2f7-0908a080f3a0\") " pod="openstack/cinder-api-0" Nov 28 13:38:29 crc kubenswrapper[4631]: I1128 13:38:29.981215 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-9ctzb"] Nov 28 13:38:30 crc kubenswrapper[4631]: I1128 13:38:30.029411 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bef6bb99-5262-4f84-a2f7-0908a080f3a0-config-data\") pod \"cinder-api-0\" (UID: \"bef6bb99-5262-4f84-a2f7-0908a080f3a0\") " pod="openstack/cinder-api-0" Nov 28 13:38:30 crc kubenswrapper[4631]: I1128 13:38:30.029571 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/dc4f7c59-e6c6-4659-8d2d-8e584afbe2ab-dns-svc\") pod \"dnsmasq-dns-5c9776ccc5-9ctzb\" (UID: \"dc4f7c59-e6c6-4659-8d2d-8e584afbe2ab\") " pod="openstack/dnsmasq-dns-5c9776ccc5-9ctzb" Nov 28 13:38:30 crc kubenswrapper[4631]: I1128 13:38:30.029598 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/dc4f7c59-e6c6-4659-8d2d-8e584afbe2ab-ovsdbserver-nb\") pod \"dnsmasq-dns-5c9776ccc5-9ctzb\" (UID: \"dc4f7c59-e6c6-4659-8d2d-8e584afbe2ab\") " pod="openstack/dnsmasq-dns-5c9776ccc5-9ctzb" Nov 28 13:38:30 crc kubenswrapper[4631]: I1128 13:38:30.029730 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bef6bb99-5262-4f84-a2f7-0908a080f3a0-logs\") pod \"cinder-api-0\" (UID: \"bef6bb99-5262-4f84-a2f7-0908a080f3a0\") " pod="openstack/cinder-api-0" Nov 28 13:38:30 crc kubenswrapper[4631]: I1128 13:38:30.029765 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bef6bb99-5262-4f84-a2f7-0908a080f3a0-scripts\") pod \"cinder-api-0\" (UID: \"bef6bb99-5262-4f84-a2f7-0908a080f3a0\") " pod="openstack/cinder-api-0" Nov 28 13:38:30 crc kubenswrapper[4631]: I1128 13:38:30.029790 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/dc4f7c59-e6c6-4659-8d2d-8e584afbe2ab-dns-swift-storage-0\") pod \"dnsmasq-dns-5c9776ccc5-9ctzb\" (UID: \"dc4f7c59-e6c6-4659-8d2d-8e584afbe2ab\") " pod="openstack/dnsmasq-dns-5c9776ccc5-9ctzb" Nov 28 13:38:30 crc kubenswrapper[4631]: I1128 13:38:30.029895 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x6plr\" (UniqueName: \"kubernetes.io/projected/bef6bb99-5262-4f84-a2f7-0908a080f3a0-kube-api-access-x6plr\") pod \"cinder-api-0\" (UID: \"bef6bb99-5262-4f84-a2f7-0908a080f3a0\") " pod="openstack/cinder-api-0" Nov 28 13:38:30 crc kubenswrapper[4631]: I1128 13:38:30.029918 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/bef6bb99-5262-4f84-a2f7-0908a080f3a0-etc-machine-id\") pod \"cinder-api-0\" (UID: \"bef6bb99-5262-4f84-a2f7-0908a080f3a0\") " pod="openstack/cinder-api-0" Nov 28 13:38:30 crc kubenswrapper[4631]: I1128 13:38:30.030026 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dc4f7c59-e6c6-4659-8d2d-8e584afbe2ab-config\") pod \"dnsmasq-dns-5c9776ccc5-9ctzb\" (UID: \"dc4f7c59-e6c6-4659-8d2d-8e584afbe2ab\") " pod="openstack/dnsmasq-dns-5c9776ccc5-9ctzb" Nov 28 13:38:30 crc kubenswrapper[4631]: I1128 13:38:30.030094 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nc9pv\" (UniqueName: \"kubernetes.io/projected/dc4f7c59-e6c6-4659-8d2d-8e584afbe2ab-kube-api-access-nc9pv\") pod \"dnsmasq-dns-5c9776ccc5-9ctzb\" (UID: \"dc4f7c59-e6c6-4659-8d2d-8e584afbe2ab\") " pod="openstack/dnsmasq-dns-5c9776ccc5-9ctzb" Nov 28 13:38:30 crc kubenswrapper[4631]: I1128 13:38:30.030174 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bef6bb99-5262-4f84-a2f7-0908a080f3a0-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"bef6bb99-5262-4f84-a2f7-0908a080f3a0\") " pod="openstack/cinder-api-0" Nov 28 13:38:30 crc kubenswrapper[4631]: I1128 13:38:30.030231 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/dc4f7c59-e6c6-4659-8d2d-8e584afbe2ab-ovsdbserver-sb\") pod \"dnsmasq-dns-5c9776ccc5-9ctzb\" (UID: \"dc4f7c59-e6c6-4659-8d2d-8e584afbe2ab\") " pod="openstack/dnsmasq-dns-5c9776ccc5-9ctzb" Nov 28 13:38:30 crc kubenswrapper[4631]: I1128 13:38:30.030275 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/bef6bb99-5262-4f84-a2f7-0908a080f3a0-config-data-custom\") pod \"cinder-api-0\" (UID: \"bef6bb99-5262-4f84-a2f7-0908a080f3a0\") " pod="openstack/cinder-api-0" Nov 28 13:38:30 crc kubenswrapper[4631]: I1128 13:38:30.031695 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/bef6bb99-5262-4f84-a2f7-0908a080f3a0-etc-machine-id\") pod \"cinder-api-0\" (UID: \"bef6bb99-5262-4f84-a2f7-0908a080f3a0\") " pod="openstack/cinder-api-0" Nov 28 13:38:30 crc kubenswrapper[4631]: I1128 13:38:30.037767 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bef6bb99-5262-4f84-a2f7-0908a080f3a0-logs\") pod \"cinder-api-0\" (UID: \"bef6bb99-5262-4f84-a2f7-0908a080f3a0\") " pod="openstack/cinder-api-0" Nov 28 13:38:30 crc kubenswrapper[4631]: I1128 13:38:30.039754 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-756d64c4c8-jzvlv" Nov 28 13:38:30 crc kubenswrapper[4631]: I1128 13:38:30.055875 4631 generic.go:334] "Generic (PLEG): container finished" podID="d95d3536-1392-4576-8ee5-390c09fe7a49" containerID="aed54e68e791750f2a7b381406c45fd57ec0e772fa0e75496b255dcade224d18" exitCode=2 Nov 28 13:38:30 crc kubenswrapper[4631]: I1128 13:38:30.055958 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85ff748b95-hhmtl" Nov 28 13:38:30 crc kubenswrapper[4631]: I1128 13:38:30.056563 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d95d3536-1392-4576-8ee5-390c09fe7a49","Type":"ContainerDied","Data":"aed54e68e791750f2a7b381406c45fd57ec0e772fa0e75496b255dcade224d18"} Nov 28 13:38:30 crc kubenswrapper[4631]: I1128 13:38:30.062603 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/bef6bb99-5262-4f84-a2f7-0908a080f3a0-config-data-custom\") pod \"cinder-api-0\" (UID: \"bef6bb99-5262-4f84-a2f7-0908a080f3a0\") " pod="openstack/cinder-api-0" Nov 28 13:38:30 crc kubenswrapper[4631]: I1128 13:38:30.064378 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bef6bb99-5262-4f84-a2f7-0908a080f3a0-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"bef6bb99-5262-4f84-a2f7-0908a080f3a0\") " pod="openstack/cinder-api-0" Nov 28 13:38:30 crc kubenswrapper[4631]: I1128 13:38:30.065040 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bef6bb99-5262-4f84-a2f7-0908a080f3a0-config-data\") pod \"cinder-api-0\" (UID: \"bef6bb99-5262-4f84-a2f7-0908a080f3a0\") " pod="openstack/cinder-api-0" Nov 28 13:38:30 crc kubenswrapper[4631]: I1128 13:38:30.091998 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bef6bb99-5262-4f84-a2f7-0908a080f3a0-scripts\") pod \"cinder-api-0\" (UID: \"bef6bb99-5262-4f84-a2f7-0908a080f3a0\") " pod="openstack/cinder-api-0" Nov 28 13:38:30 crc kubenswrapper[4631]: I1128 13:38:30.127439 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x6plr\" (UniqueName: \"kubernetes.io/projected/bef6bb99-5262-4f84-a2f7-0908a080f3a0-kube-api-access-x6plr\") pod \"cinder-api-0\" (UID: \"bef6bb99-5262-4f84-a2f7-0908a080f3a0\") " pod="openstack/cinder-api-0" Nov 28 13:38:30 crc kubenswrapper[4631]: I1128 13:38:30.143922 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/dc4f7c59-e6c6-4659-8d2d-8e584afbe2ab-ovsdbserver-sb\") pod \"dnsmasq-dns-5c9776ccc5-9ctzb\" (UID: \"dc4f7c59-e6c6-4659-8d2d-8e584afbe2ab\") " pod="openstack/dnsmasq-dns-5c9776ccc5-9ctzb" Nov 28 13:38:30 crc kubenswrapper[4631]: I1128 13:38:30.144108 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/dc4f7c59-e6c6-4659-8d2d-8e584afbe2ab-dns-svc\") pod \"dnsmasq-dns-5c9776ccc5-9ctzb\" (UID: \"dc4f7c59-e6c6-4659-8d2d-8e584afbe2ab\") " pod="openstack/dnsmasq-dns-5c9776ccc5-9ctzb" Nov 28 13:38:30 crc kubenswrapper[4631]: I1128 13:38:30.144137 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/dc4f7c59-e6c6-4659-8d2d-8e584afbe2ab-ovsdbserver-nb\") pod \"dnsmasq-dns-5c9776ccc5-9ctzb\" (UID: \"dc4f7c59-e6c6-4659-8d2d-8e584afbe2ab\") " pod="openstack/dnsmasq-dns-5c9776ccc5-9ctzb" Nov 28 13:38:30 crc kubenswrapper[4631]: I1128 13:38:30.144303 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/dc4f7c59-e6c6-4659-8d2d-8e584afbe2ab-dns-swift-storage-0\") pod \"dnsmasq-dns-5c9776ccc5-9ctzb\" (UID: \"dc4f7c59-e6c6-4659-8d2d-8e584afbe2ab\") " pod="openstack/dnsmasq-dns-5c9776ccc5-9ctzb" Nov 28 13:38:30 crc kubenswrapper[4631]: I1128 13:38:30.144428 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dc4f7c59-e6c6-4659-8d2d-8e584afbe2ab-config\") pod \"dnsmasq-dns-5c9776ccc5-9ctzb\" (UID: \"dc4f7c59-e6c6-4659-8d2d-8e584afbe2ab\") " pod="openstack/dnsmasq-dns-5c9776ccc5-9ctzb" Nov 28 13:38:30 crc kubenswrapper[4631]: I1128 13:38:30.144498 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nc9pv\" (UniqueName: \"kubernetes.io/projected/dc4f7c59-e6c6-4659-8d2d-8e584afbe2ab-kube-api-access-nc9pv\") pod \"dnsmasq-dns-5c9776ccc5-9ctzb\" (UID: \"dc4f7c59-e6c6-4659-8d2d-8e584afbe2ab\") " pod="openstack/dnsmasq-dns-5c9776ccc5-9ctzb" Nov 28 13:38:30 crc kubenswrapper[4631]: I1128 13:38:30.145799 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/dc4f7c59-e6c6-4659-8d2d-8e584afbe2ab-ovsdbserver-sb\") pod \"dnsmasq-dns-5c9776ccc5-9ctzb\" (UID: \"dc4f7c59-e6c6-4659-8d2d-8e584afbe2ab\") " pod="openstack/dnsmasq-dns-5c9776ccc5-9ctzb" Nov 28 13:38:30 crc kubenswrapper[4631]: I1128 13:38:30.150149 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/dc4f7c59-e6c6-4659-8d2d-8e584afbe2ab-dns-swift-storage-0\") pod \"dnsmasq-dns-5c9776ccc5-9ctzb\" (UID: \"dc4f7c59-e6c6-4659-8d2d-8e584afbe2ab\") " pod="openstack/dnsmasq-dns-5c9776ccc5-9ctzb" Nov 28 13:38:30 crc kubenswrapper[4631]: I1128 13:38:30.151256 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/dc4f7c59-e6c6-4659-8d2d-8e584afbe2ab-ovsdbserver-nb\") pod \"dnsmasq-dns-5c9776ccc5-9ctzb\" (UID: \"dc4f7c59-e6c6-4659-8d2d-8e584afbe2ab\") " pod="openstack/dnsmasq-dns-5c9776ccc5-9ctzb" Nov 28 13:38:30 crc kubenswrapper[4631]: I1128 13:38:30.152617 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dc4f7c59-e6c6-4659-8d2d-8e584afbe2ab-config\") pod \"dnsmasq-dns-5c9776ccc5-9ctzb\" (UID: \"dc4f7c59-e6c6-4659-8d2d-8e584afbe2ab\") " pod="openstack/dnsmasq-dns-5c9776ccc5-9ctzb" Nov 28 13:38:30 crc kubenswrapper[4631]: I1128 13:38:30.154369 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/dc4f7c59-e6c6-4659-8d2d-8e584afbe2ab-dns-svc\") pod \"dnsmasq-dns-5c9776ccc5-9ctzb\" (UID: \"dc4f7c59-e6c6-4659-8d2d-8e584afbe2ab\") " pod="openstack/dnsmasq-dns-5c9776ccc5-9ctzb" Nov 28 13:38:30 crc kubenswrapper[4631]: I1128 13:38:30.182729 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 28 13:38:30 crc kubenswrapper[4631]: I1128 13:38:30.196242 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nc9pv\" (UniqueName: \"kubernetes.io/projected/dc4f7c59-e6c6-4659-8d2d-8e584afbe2ab-kube-api-access-nc9pv\") pod \"dnsmasq-dns-5c9776ccc5-9ctzb\" (UID: \"dc4f7c59-e6c6-4659-8d2d-8e584afbe2ab\") " pod="openstack/dnsmasq-dns-5c9776ccc5-9ctzb" Nov 28 13:38:30 crc kubenswrapper[4631]: I1128 13:38:30.276801 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85ff748b95-hhmtl" Nov 28 13:38:30 crc kubenswrapper[4631]: I1128 13:38:30.342203 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9776ccc5-9ctzb" Nov 28 13:38:30 crc kubenswrapper[4631]: I1128 13:38:30.348621 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/abf8d962-b7ff-4e5e-8ccc-e5ffae26f822-config\") pod \"abf8d962-b7ff-4e5e-8ccc-e5ffae26f822\" (UID: \"abf8d962-b7ff-4e5e-8ccc-e5ffae26f822\") " Nov 28 13:38:30 crc kubenswrapper[4631]: I1128 13:38:30.348679 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/abf8d962-b7ff-4e5e-8ccc-e5ffae26f822-ovsdbserver-sb\") pod \"abf8d962-b7ff-4e5e-8ccc-e5ffae26f822\" (UID: \"abf8d962-b7ff-4e5e-8ccc-e5ffae26f822\") " Nov 28 13:38:30 crc kubenswrapper[4631]: I1128 13:38:30.348765 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/abf8d962-b7ff-4e5e-8ccc-e5ffae26f822-ovsdbserver-nb\") pod \"abf8d962-b7ff-4e5e-8ccc-e5ffae26f822\" (UID: \"abf8d962-b7ff-4e5e-8ccc-e5ffae26f822\") " Nov 28 13:38:30 crc kubenswrapper[4631]: I1128 13:38:30.348878 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/abf8d962-b7ff-4e5e-8ccc-e5ffae26f822-dns-swift-storage-0\") pod \"abf8d962-b7ff-4e5e-8ccc-e5ffae26f822\" (UID: \"abf8d962-b7ff-4e5e-8ccc-e5ffae26f822\") " Nov 28 13:38:30 crc kubenswrapper[4631]: I1128 13:38:30.349016 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2trlv\" (UniqueName: \"kubernetes.io/projected/abf8d962-b7ff-4e5e-8ccc-e5ffae26f822-kube-api-access-2trlv\") pod \"abf8d962-b7ff-4e5e-8ccc-e5ffae26f822\" (UID: \"abf8d962-b7ff-4e5e-8ccc-e5ffae26f822\") " Nov 28 13:38:30 crc kubenswrapper[4631]: I1128 13:38:30.349085 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/abf8d962-b7ff-4e5e-8ccc-e5ffae26f822-dns-svc\") pod \"abf8d962-b7ff-4e5e-8ccc-e5ffae26f822\" (UID: \"abf8d962-b7ff-4e5e-8ccc-e5ffae26f822\") " Nov 28 13:38:30 crc kubenswrapper[4631]: I1128 13:38:30.349797 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/abf8d962-b7ff-4e5e-8ccc-e5ffae26f822-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "abf8d962-b7ff-4e5e-8ccc-e5ffae26f822" (UID: "abf8d962-b7ff-4e5e-8ccc-e5ffae26f822"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 13:38:30 crc kubenswrapper[4631]: I1128 13:38:30.350052 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/abf8d962-b7ff-4e5e-8ccc-e5ffae26f822-config" (OuterVolumeSpecName: "config") pod "abf8d962-b7ff-4e5e-8ccc-e5ffae26f822" (UID: "abf8d962-b7ff-4e5e-8ccc-e5ffae26f822"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 13:38:30 crc kubenswrapper[4631]: I1128 13:38:30.350305 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/abf8d962-b7ff-4e5e-8ccc-e5ffae26f822-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "abf8d962-b7ff-4e5e-8ccc-e5ffae26f822" (UID: "abf8d962-b7ff-4e5e-8ccc-e5ffae26f822"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 13:38:30 crc kubenswrapper[4631]: I1128 13:38:30.350539 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/abf8d962-b7ff-4e5e-8ccc-e5ffae26f822-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "abf8d962-b7ff-4e5e-8ccc-e5ffae26f822" (UID: "abf8d962-b7ff-4e5e-8ccc-e5ffae26f822"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 13:38:30 crc kubenswrapper[4631]: I1128 13:38:30.357450 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/abf8d962-b7ff-4e5e-8ccc-e5ffae26f822-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "abf8d962-b7ff-4e5e-8ccc-e5ffae26f822" (UID: "abf8d962-b7ff-4e5e-8ccc-e5ffae26f822"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 13:38:30 crc kubenswrapper[4631]: I1128 13:38:30.372551 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/abf8d962-b7ff-4e5e-8ccc-e5ffae26f822-kube-api-access-2trlv" (OuterVolumeSpecName: "kube-api-access-2trlv") pod "abf8d962-b7ff-4e5e-8ccc-e5ffae26f822" (UID: "abf8d962-b7ff-4e5e-8ccc-e5ffae26f822"). InnerVolumeSpecName "kube-api-access-2trlv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:38:30 crc kubenswrapper[4631]: I1128 13:38:30.451307 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2trlv\" (UniqueName: \"kubernetes.io/projected/abf8d962-b7ff-4e5e-8ccc-e5ffae26f822-kube-api-access-2trlv\") on node \"crc\" DevicePath \"\"" Nov 28 13:38:30 crc kubenswrapper[4631]: I1128 13:38:30.451642 4631 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/abf8d962-b7ff-4e5e-8ccc-e5ffae26f822-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 28 13:38:30 crc kubenswrapper[4631]: I1128 13:38:30.451662 4631 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/abf8d962-b7ff-4e5e-8ccc-e5ffae26f822-config\") on node \"crc\" DevicePath \"\"" Nov 28 13:38:30 crc kubenswrapper[4631]: I1128 13:38:30.451682 4631 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/abf8d962-b7ff-4e5e-8ccc-e5ffae26f822-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 28 13:38:30 crc kubenswrapper[4631]: I1128 13:38:30.451700 4631 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/abf8d962-b7ff-4e5e-8ccc-e5ffae26f822-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 28 13:38:30 crc kubenswrapper[4631]: I1128 13:38:30.451718 4631 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/abf8d962-b7ff-4e5e-8ccc-e5ffae26f822-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 28 13:38:30 crc kubenswrapper[4631]: I1128 13:38:30.645713 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-59c4c4dc74-qt87t"] Nov 28 13:38:30 crc kubenswrapper[4631]: I1128 13:38:30.796042 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-5b8f56c4bc-cntw9"] Nov 28 13:38:30 crc kubenswrapper[4631]: I1128 13:38:30.806355 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 28 13:38:30 crc kubenswrapper[4631]: I1128 13:38:30.929365 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-756d64c4c8-jzvlv"] Nov 28 13:38:31 crc kubenswrapper[4631]: I1128 13:38:31.116560 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-756d64c4c8-jzvlv" event={"ID":"7eea7f3e-fc3c-4c60-998e-79c9a31ee084","Type":"ContainerStarted","Data":"4122a24fb417b9c8665b47ec5c5eda95ae81959c07dcb5d83737b2f293b86f13"} Nov 28 13:38:31 crc kubenswrapper[4631]: I1128 13:38:31.117442 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Nov 28 13:38:31 crc kubenswrapper[4631]: I1128 13:38:31.119588 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-59c4c4dc74-qt87t" event={"ID":"177f3e63-a6ce-4512-80b3-f229e3ace62a","Type":"ContainerStarted","Data":"dd5bd3e75d46e29bdc47acdb134e15cd23727a1b212d9b1fc58e1fdd882adbf5"} Nov 28 13:38:31 crc kubenswrapper[4631]: I1128 13:38:31.125071 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"8f2270d7-d4a8-4cf5-a0b4-b5c29f4fb817","Type":"ContainerStarted","Data":"2749c4cccac47c4ccc8000027598e38875d27778aaf92fd30117a5f0021ed5c4"} Nov 28 13:38:31 crc kubenswrapper[4631]: I1128 13:38:31.131955 4631 generic.go:334] "Generic (PLEG): container finished" podID="d95d3536-1392-4576-8ee5-390c09fe7a49" containerID="42664776e24655262092eae45a6b6b94e6ed84e5b6614b7c0abe60d7e38df756" exitCode=0 Nov 28 13:38:31 crc kubenswrapper[4631]: I1128 13:38:31.132020 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d95d3536-1392-4576-8ee5-390c09fe7a49","Type":"ContainerDied","Data":"42664776e24655262092eae45a6b6b94e6ed84e5b6614b7c0abe60d7e38df756"} Nov 28 13:38:31 crc kubenswrapper[4631]: I1128 13:38:31.155327 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-5b8f56c4bc-cntw9" event={"ID":"1b8b4a3c-2e24-4a9b-b18d-0fa09f70a155","Type":"ContainerStarted","Data":"51192b08376958b6a1a6a6a2762acc3e7869429b311a07b74d9d652369662fd7"} Nov 28 13:38:31 crc kubenswrapper[4631]: I1128 13:38:31.155380 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85ff748b95-hhmtl" Nov 28 13:38:31 crc kubenswrapper[4631]: I1128 13:38:31.342117 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-9ctzb"] Nov 28 13:38:31 crc kubenswrapper[4631]: I1128 13:38:31.463025 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-85ff748b95-hhmtl"] Nov 28 13:38:31 crc kubenswrapper[4631]: I1128 13:38:31.484795 4631 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-85ff748b95-hhmtl"] Nov 28 13:38:31 crc kubenswrapper[4631]: I1128 13:38:31.535613 4631 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="abf8d962-b7ff-4e5e-8ccc-e5ffae26f822" path="/var/lib/kubelet/pods/abf8d962-b7ff-4e5e-8ccc-e5ffae26f822/volumes" Nov 28 13:38:31 crc kubenswrapper[4631]: I1128 13:38:31.536487 4631 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c3bcfdda-21af-4504-806c-50e6bc0f8770" path="/var/lib/kubelet/pods/c3bcfdda-21af-4504-806c-50e6bc0f8770/volumes" Nov 28 13:38:32 crc kubenswrapper[4631]: I1128 13:38:32.191737 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"bef6bb99-5262-4f84-a2f7-0908a080f3a0","Type":"ContainerStarted","Data":"a63af8a28bdb74f133cd9aba1c8b756b01dccc91b42f05e57625c4b57feba93b"} Nov 28 13:38:32 crc kubenswrapper[4631]: I1128 13:38:32.199193 4631 generic.go:334] "Generic (PLEG): container finished" podID="dc4f7c59-e6c6-4659-8d2d-8e584afbe2ab" containerID="8162b6a5a69f910fbfd625bb8c521567d39c2ebd96fe6a3aec4958d8543d33bf" exitCode=0 Nov 28 13:38:32 crc kubenswrapper[4631]: I1128 13:38:32.199280 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-9ctzb" event={"ID":"dc4f7c59-e6c6-4659-8d2d-8e584afbe2ab","Type":"ContainerDied","Data":"8162b6a5a69f910fbfd625bb8c521567d39c2ebd96fe6a3aec4958d8543d33bf"} Nov 28 13:38:32 crc kubenswrapper[4631]: I1128 13:38:32.199341 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-9ctzb" event={"ID":"dc4f7c59-e6c6-4659-8d2d-8e584afbe2ab","Type":"ContainerStarted","Data":"98ed656c4e7b1e35f2223f91dbc22e1e584d50bbde7da6b5daf1caed1d8eb034"} Nov 28 13:38:32 crc kubenswrapper[4631]: I1128 13:38:32.238455 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-756d64c4c8-jzvlv" event={"ID":"7eea7f3e-fc3c-4c60-998e-79c9a31ee084","Type":"ContainerStarted","Data":"39f6239d34a7e4114e98f79691b15de6581a312f79450b6aa821ddc52f868ed0"} Nov 28 13:38:32 crc kubenswrapper[4631]: I1128 13:38:32.238783 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-756d64c4c8-jzvlv" Nov 28 13:38:32 crc kubenswrapper[4631]: I1128 13:38:32.279375 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-756d64c4c8-jzvlv" podStartSLOduration=3.279354176 podStartE2EDuration="3.279354176s" podCreationTimestamp="2025-11-28 13:38:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 13:38:32.275938871 +0000 UTC m=+1069.083242215" watchObservedRunningTime="2025-11-28 13:38:32.279354176 +0000 UTC m=+1069.086657520" Nov 28 13:38:33 crc kubenswrapper[4631]: I1128 13:38:33.228398 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Nov 28 13:38:33 crc kubenswrapper[4631]: I1128 13:38:33.300264 4631 generic.go:334] "Generic (PLEG): container finished" podID="d95d3536-1392-4576-8ee5-390c09fe7a49" containerID="2141eebe891186cff2651a86531e3f94d1b78cbeeead6dde52101d7a46457ff5" exitCode=0 Nov 28 13:38:33 crc kubenswrapper[4631]: I1128 13:38:33.300418 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d95d3536-1392-4576-8ee5-390c09fe7a49","Type":"ContainerDied","Data":"2141eebe891186cff2651a86531e3f94d1b78cbeeead6dde52101d7a46457ff5"} Nov 28 13:38:33 crc kubenswrapper[4631]: I1128 13:38:33.317581 4631 generic.go:334] "Generic (PLEG): container finished" podID="6a0a46ba-e48f-45a6-801e-7cb0ececabee" containerID="821a3f2847ee539065f14edad868b3980fc7ccbc996dbb6dbf3a771daf646f34" exitCode=0 Nov 28 13:38:33 crc kubenswrapper[4631]: I1128 13:38:33.317664 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5fb67d6644-dbmzf" event={"ID":"6a0a46ba-e48f-45a6-801e-7cb0ececabee","Type":"ContainerDied","Data":"821a3f2847ee539065f14edad868b3980fc7ccbc996dbb6dbf3a771daf646f34"} Nov 28 13:38:33 crc kubenswrapper[4631]: I1128 13:38:33.329522 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-756d64c4c8-jzvlv" event={"ID":"7eea7f3e-fc3c-4c60-998e-79c9a31ee084","Type":"ContainerStarted","Data":"c9851fdf444e8d0158c79114f4288e3ea2c1a53294a47a6cfe8d648dde4e2e5f"} Nov 28 13:38:33 crc kubenswrapper[4631]: I1128 13:38:33.330569 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-756d64c4c8-jzvlv" Nov 28 13:38:33 crc kubenswrapper[4631]: I1128 13:38:33.342879 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"bef6bb99-5262-4f84-a2f7-0908a080f3a0","Type":"ContainerStarted","Data":"d9367482c870497f738de7f33c554511341371eb34004d4968e56132c4122076"} Nov 28 13:38:33 crc kubenswrapper[4631]: I1128 13:38:33.346539 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-9ctzb" event={"ID":"dc4f7c59-e6c6-4659-8d2d-8e584afbe2ab","Type":"ContainerStarted","Data":"c7c4bb142691d554c4385d0a03fd36470b904945d17f19500072f4ec681c4907"} Nov 28 13:38:33 crc kubenswrapper[4631]: I1128 13:38:33.346855 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5c9776ccc5-9ctzb" Nov 28 13:38:33 crc kubenswrapper[4631]: I1128 13:38:33.389940 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5c9776ccc5-9ctzb" podStartSLOduration=4.389915225 podStartE2EDuration="4.389915225s" podCreationTimestamp="2025-11-28 13:38:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 13:38:33.380049973 +0000 UTC m=+1070.187353317" watchObservedRunningTime="2025-11-28 13:38:33.389915225 +0000 UTC m=+1070.197218569" Nov 28 13:38:34 crc kubenswrapper[4631]: I1128 13:38:34.390117 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"8f2270d7-d4a8-4cf5-a0b4-b5c29f4fb817","Type":"ContainerStarted","Data":"598d394db8cefb7d6686ceb6bdb8cc64bd0da8e045b347b40219b4b81fa90967"} Nov 28 13:38:35 crc kubenswrapper[4631]: I1128 13:38:35.221011 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5fb67d6644-dbmzf" Nov 28 13:38:35 crc kubenswrapper[4631]: I1128 13:38:35.305521 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a0a46ba-e48f-45a6-801e-7cb0ececabee-combined-ca-bundle\") pod \"6a0a46ba-e48f-45a6-801e-7cb0ececabee\" (UID: \"6a0a46ba-e48f-45a6-801e-7cb0ececabee\") " Nov 28 13:38:35 crc kubenswrapper[4631]: I1128 13:38:35.305650 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/6a0a46ba-e48f-45a6-801e-7cb0ececabee-config\") pod \"6a0a46ba-e48f-45a6-801e-7cb0ececabee\" (UID: \"6a0a46ba-e48f-45a6-801e-7cb0ececabee\") " Nov 28 13:38:35 crc kubenswrapper[4631]: I1128 13:38:35.305872 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/6a0a46ba-e48f-45a6-801e-7cb0ececabee-ovndb-tls-certs\") pod \"6a0a46ba-e48f-45a6-801e-7cb0ececabee\" (UID: \"6a0a46ba-e48f-45a6-801e-7cb0ececabee\") " Nov 28 13:38:35 crc kubenswrapper[4631]: I1128 13:38:35.305957 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cblkg\" (UniqueName: \"kubernetes.io/projected/6a0a46ba-e48f-45a6-801e-7cb0ececabee-kube-api-access-cblkg\") pod \"6a0a46ba-e48f-45a6-801e-7cb0ececabee\" (UID: \"6a0a46ba-e48f-45a6-801e-7cb0ececabee\") " Nov 28 13:38:35 crc kubenswrapper[4631]: I1128 13:38:35.305992 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/6a0a46ba-e48f-45a6-801e-7cb0ececabee-httpd-config\") pod \"6a0a46ba-e48f-45a6-801e-7cb0ececabee\" (UID: \"6a0a46ba-e48f-45a6-801e-7cb0ececabee\") " Nov 28 13:38:35 crc kubenswrapper[4631]: I1128 13:38:35.313669 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6a0a46ba-e48f-45a6-801e-7cb0ececabee-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "6a0a46ba-e48f-45a6-801e-7cb0ececabee" (UID: "6a0a46ba-e48f-45a6-801e-7cb0ececabee"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:38:35 crc kubenswrapper[4631]: I1128 13:38:35.320973 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6a0a46ba-e48f-45a6-801e-7cb0ececabee-kube-api-access-cblkg" (OuterVolumeSpecName: "kube-api-access-cblkg") pod "6a0a46ba-e48f-45a6-801e-7cb0ececabee" (UID: "6a0a46ba-e48f-45a6-801e-7cb0ececabee"). InnerVolumeSpecName "kube-api-access-cblkg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:38:35 crc kubenswrapper[4631]: I1128 13:38:35.384879 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6a0a46ba-e48f-45a6-801e-7cb0ececabee-config" (OuterVolumeSpecName: "config") pod "6a0a46ba-e48f-45a6-801e-7cb0ececabee" (UID: "6a0a46ba-e48f-45a6-801e-7cb0ececabee"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:38:35 crc kubenswrapper[4631]: I1128 13:38:35.408809 4631 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/6a0a46ba-e48f-45a6-801e-7cb0ececabee-httpd-config\") on node \"crc\" DevicePath \"\"" Nov 28 13:38:35 crc kubenswrapper[4631]: I1128 13:38:35.408842 4631 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/6a0a46ba-e48f-45a6-801e-7cb0ececabee-config\") on node \"crc\" DevicePath \"\"" Nov 28 13:38:35 crc kubenswrapper[4631]: I1128 13:38:35.408854 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cblkg\" (UniqueName: \"kubernetes.io/projected/6a0a46ba-e48f-45a6-801e-7cb0ececabee-kube-api-access-cblkg\") on node \"crc\" DevicePath \"\"" Nov 28 13:38:35 crc kubenswrapper[4631]: I1128 13:38:35.414822 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6a0a46ba-e48f-45a6-801e-7cb0ececabee-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6a0a46ba-e48f-45a6-801e-7cb0ececabee" (UID: "6a0a46ba-e48f-45a6-801e-7cb0ececabee"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:38:35 crc kubenswrapper[4631]: I1128 13:38:35.430757 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"bef6bb99-5262-4f84-a2f7-0908a080f3a0","Type":"ContainerStarted","Data":"2fae4b9d182ba1a2f4c1ae5abd2caf92fbcae61387ec588c09af0ad1460aaf0a"} Nov 28 13:38:35 crc kubenswrapper[4631]: I1128 13:38:35.431123 4631 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="bef6bb99-5262-4f84-a2f7-0908a080f3a0" containerName="cinder-api-log" containerID="cri-o://d9367482c870497f738de7f33c554511341371eb34004d4968e56132c4122076" gracePeriod=30 Nov 28 13:38:35 crc kubenswrapper[4631]: I1128 13:38:35.431464 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Nov 28 13:38:35 crc kubenswrapper[4631]: I1128 13:38:35.431493 4631 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="bef6bb99-5262-4f84-a2f7-0908a080f3a0" containerName="cinder-api" containerID="cri-o://2fae4b9d182ba1a2f4c1ae5abd2caf92fbcae61387ec588c09af0ad1460aaf0a" gracePeriod=30 Nov 28 13:38:35 crc kubenswrapper[4631]: I1128 13:38:35.440042 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5fb67d6644-dbmzf" event={"ID":"6a0a46ba-e48f-45a6-801e-7cb0ececabee","Type":"ContainerDied","Data":"630b1d6a6a9df29e837276658f20fa9c41ae2a7730d5d0daf69e9bd2a2b1eb75"} Nov 28 13:38:35 crc kubenswrapper[4631]: I1128 13:38:35.440634 4631 scope.go:117] "RemoveContainer" containerID="42641fee468bc31b2584b9a1325c1fdc14069913633b3c778d6d1079d6e2d587" Nov 28 13:38:35 crc kubenswrapper[4631]: I1128 13:38:35.440834 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5fb67d6644-dbmzf" Nov 28 13:38:35 crc kubenswrapper[4631]: I1128 13:38:35.461944 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=6.46192128 podStartE2EDuration="6.46192128s" podCreationTimestamp="2025-11-28 13:38:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 13:38:35.45379326 +0000 UTC m=+1072.261096604" watchObservedRunningTime="2025-11-28 13:38:35.46192128 +0000 UTC m=+1072.269224624" Nov 28 13:38:35 crc kubenswrapper[4631]: I1128 13:38:35.473405 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6a0a46ba-e48f-45a6-801e-7cb0ececabee-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "6a0a46ba-e48f-45a6-801e-7cb0ececabee" (UID: "6a0a46ba-e48f-45a6-801e-7cb0ececabee"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:38:35 crc kubenswrapper[4631]: I1128 13:38:35.511063 4631 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/6a0a46ba-e48f-45a6-801e-7cb0ececabee-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 28 13:38:35 crc kubenswrapper[4631]: I1128 13:38:35.511093 4631 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a0a46ba-e48f-45a6-801e-7cb0ececabee-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 13:38:35 crc kubenswrapper[4631]: I1128 13:38:35.579596 4631 scope.go:117] "RemoveContainer" containerID="821a3f2847ee539065f14edad868b3980fc7ccbc996dbb6dbf3a771daf646f34" Nov 28 13:38:35 crc kubenswrapper[4631]: I1128 13:38:35.781389 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-5fb67d6644-dbmzf"] Nov 28 13:38:35 crc kubenswrapper[4631]: I1128 13:38:35.792573 4631 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-5fb67d6644-dbmzf"] Nov 28 13:38:36 crc kubenswrapper[4631]: I1128 13:38:36.463227 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"8f2270d7-d4a8-4cf5-a0b4-b5c29f4fb817","Type":"ContainerStarted","Data":"c79f1d7d0649f621429d5f1b6d86425221cf1a427688fb118f898b252947cec4"} Nov 28 13:38:36 crc kubenswrapper[4631]: I1128 13:38:36.465378 4631 generic.go:334] "Generic (PLEG): container finished" podID="bef6bb99-5262-4f84-a2f7-0908a080f3a0" containerID="d9367482c870497f738de7f33c554511341371eb34004d4968e56132c4122076" exitCode=143 Nov 28 13:38:36 crc kubenswrapper[4631]: I1128 13:38:36.465462 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"bef6bb99-5262-4f84-a2f7-0908a080f3a0","Type":"ContainerDied","Data":"d9367482c870497f738de7f33c554511341371eb34004d4968e56132c4122076"} Nov 28 13:38:36 crc kubenswrapper[4631]: I1128 13:38:36.468715 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-5b8f56c4bc-cntw9" event={"ID":"1b8b4a3c-2e24-4a9b-b18d-0fa09f70a155","Type":"ContainerStarted","Data":"8cfdf33a1d3e4b99b957a85bc40f34adc97479616139e436c51646bad450f042"} Nov 28 13:38:36 crc kubenswrapper[4631]: I1128 13:38:36.468749 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-5b8f56c4bc-cntw9" event={"ID":"1b8b4a3c-2e24-4a9b-b18d-0fa09f70a155","Type":"ContainerStarted","Data":"496ecef4248096516cd71640f86389b1bcd23beb5c6a127ce2f519cb7bbf9ce9"} Nov 28 13:38:36 crc kubenswrapper[4631]: I1128 13:38:36.471880 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-59c4c4dc74-qt87t" event={"ID":"177f3e63-a6ce-4512-80b3-f229e3ace62a","Type":"ContainerStarted","Data":"9f04e14ecd224ff44580cdf899cd134b2bfd6fb8a0150051e5e17cb402601212"} Nov 28 13:38:36 crc kubenswrapper[4631]: I1128 13:38:36.471919 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-59c4c4dc74-qt87t" event={"ID":"177f3e63-a6ce-4512-80b3-f229e3ace62a","Type":"ContainerStarted","Data":"c1b7a2636934458121763ae6ccfa58f2ca911b9dc4caf646896dafaf13501f92"} Nov 28 13:38:36 crc kubenswrapper[4631]: I1128 13:38:36.558067 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-5b8f56c4bc-cntw9" podStartSLOduration=4.253670044 podStartE2EDuration="8.558042064s" podCreationTimestamp="2025-11-28 13:38:28 +0000 UTC" firstStartedPulling="2025-11-28 13:38:30.825183782 +0000 UTC m=+1067.632487126" lastFinishedPulling="2025-11-28 13:38:35.129555802 +0000 UTC m=+1071.936859146" observedRunningTime="2025-11-28 13:38:36.518119833 +0000 UTC m=+1073.325423177" watchObservedRunningTime="2025-11-28 13:38:36.558042064 +0000 UTC m=+1073.365345408" Nov 28 13:38:36 crc kubenswrapper[4631]: I1128 13:38:36.560882 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=7.542383587 podStartE2EDuration="8.560869754s" podCreationTimestamp="2025-11-28 13:38:28 +0000 UTC" firstStartedPulling="2025-11-28 13:38:30.84748696 +0000 UTC m=+1067.654790304" lastFinishedPulling="2025-11-28 13:38:31.865973127 +0000 UTC m=+1068.673276471" observedRunningTime="2025-11-28 13:38:36.491798676 +0000 UTC m=+1073.299102030" watchObservedRunningTime="2025-11-28 13:38:36.560869754 +0000 UTC m=+1073.368173098" Nov 28 13:38:36 crc kubenswrapper[4631]: I1128 13:38:36.649446 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-54885d4b84-55wlk"] Nov 28 13:38:36 crc kubenswrapper[4631]: E1128 13:38:36.650433 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a0a46ba-e48f-45a6-801e-7cb0ececabee" containerName="neutron-api" Nov 28 13:38:36 crc kubenswrapper[4631]: I1128 13:38:36.650460 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a0a46ba-e48f-45a6-801e-7cb0ececabee" containerName="neutron-api" Nov 28 13:38:36 crc kubenswrapper[4631]: E1128 13:38:36.650514 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a0a46ba-e48f-45a6-801e-7cb0ececabee" containerName="neutron-httpd" Nov 28 13:38:36 crc kubenswrapper[4631]: I1128 13:38:36.650525 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a0a46ba-e48f-45a6-801e-7cb0ececabee" containerName="neutron-httpd" Nov 28 13:38:36 crc kubenswrapper[4631]: I1128 13:38:36.650913 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="6a0a46ba-e48f-45a6-801e-7cb0ececabee" containerName="neutron-api" Nov 28 13:38:36 crc kubenswrapper[4631]: I1128 13:38:36.650951 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="6a0a46ba-e48f-45a6-801e-7cb0ececabee" containerName="neutron-httpd" Nov 28 13:38:36 crc kubenswrapper[4631]: I1128 13:38:36.654010 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-54885d4b84-55wlk" Nov 28 13:38:36 crc kubenswrapper[4631]: I1128 13:38:36.664207 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Nov 28 13:38:36 crc kubenswrapper[4631]: I1128 13:38:36.664256 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Nov 28 13:38:36 crc kubenswrapper[4631]: I1128 13:38:36.676546 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-59c4c4dc74-qt87t" podStartSLOduration=4.229961661 podStartE2EDuration="8.676511985s" podCreationTimestamp="2025-11-28 13:38:28 +0000 UTC" firstStartedPulling="2025-11-28 13:38:30.666134534 +0000 UTC m=+1067.473437878" lastFinishedPulling="2025-11-28 13:38:35.112684858 +0000 UTC m=+1071.919988202" observedRunningTime="2025-11-28 13:38:36.573994906 +0000 UTC m=+1073.381298250" watchObservedRunningTime="2025-11-28 13:38:36.676511985 +0000 UTC m=+1073.483815329" Nov 28 13:38:36 crc kubenswrapper[4631]: I1128 13:38:36.752368 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-54885d4b84-55wlk"] Nov 28 13:38:36 crc kubenswrapper[4631]: I1128 13:38:36.755503 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1a3cfa34-17ae-42d6-8ea7-46691bb10eaa-config-data\") pod \"barbican-api-54885d4b84-55wlk\" (UID: \"1a3cfa34-17ae-42d6-8ea7-46691bb10eaa\") " pod="openstack/barbican-api-54885d4b84-55wlk" Nov 28 13:38:36 crc kubenswrapper[4631]: I1128 13:38:36.755613 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1a3cfa34-17ae-42d6-8ea7-46691bb10eaa-config-data-custom\") pod \"barbican-api-54885d4b84-55wlk\" (UID: \"1a3cfa34-17ae-42d6-8ea7-46691bb10eaa\") " pod="openstack/barbican-api-54885d4b84-55wlk" Nov 28 13:38:36 crc kubenswrapper[4631]: I1128 13:38:36.755894 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1a3cfa34-17ae-42d6-8ea7-46691bb10eaa-logs\") pod \"barbican-api-54885d4b84-55wlk\" (UID: \"1a3cfa34-17ae-42d6-8ea7-46691bb10eaa\") " pod="openstack/barbican-api-54885d4b84-55wlk" Nov 28 13:38:36 crc kubenswrapper[4631]: I1128 13:38:36.756098 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1a3cfa34-17ae-42d6-8ea7-46691bb10eaa-internal-tls-certs\") pod \"barbican-api-54885d4b84-55wlk\" (UID: \"1a3cfa34-17ae-42d6-8ea7-46691bb10eaa\") " pod="openstack/barbican-api-54885d4b84-55wlk" Nov 28 13:38:36 crc kubenswrapper[4631]: I1128 13:38:36.756203 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1a3cfa34-17ae-42d6-8ea7-46691bb10eaa-public-tls-certs\") pod \"barbican-api-54885d4b84-55wlk\" (UID: \"1a3cfa34-17ae-42d6-8ea7-46691bb10eaa\") " pod="openstack/barbican-api-54885d4b84-55wlk" Nov 28 13:38:36 crc kubenswrapper[4631]: I1128 13:38:36.756258 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dpqjn\" (UniqueName: \"kubernetes.io/projected/1a3cfa34-17ae-42d6-8ea7-46691bb10eaa-kube-api-access-dpqjn\") pod \"barbican-api-54885d4b84-55wlk\" (UID: \"1a3cfa34-17ae-42d6-8ea7-46691bb10eaa\") " pod="openstack/barbican-api-54885d4b84-55wlk" Nov 28 13:38:36 crc kubenswrapper[4631]: I1128 13:38:36.756405 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a3cfa34-17ae-42d6-8ea7-46691bb10eaa-combined-ca-bundle\") pod \"barbican-api-54885d4b84-55wlk\" (UID: \"1a3cfa34-17ae-42d6-8ea7-46691bb10eaa\") " pod="openstack/barbican-api-54885d4b84-55wlk" Nov 28 13:38:36 crc kubenswrapper[4631]: I1128 13:38:36.858499 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1a3cfa34-17ae-42d6-8ea7-46691bb10eaa-logs\") pod \"barbican-api-54885d4b84-55wlk\" (UID: \"1a3cfa34-17ae-42d6-8ea7-46691bb10eaa\") " pod="openstack/barbican-api-54885d4b84-55wlk" Nov 28 13:38:36 crc kubenswrapper[4631]: I1128 13:38:36.858982 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1a3cfa34-17ae-42d6-8ea7-46691bb10eaa-logs\") pod \"barbican-api-54885d4b84-55wlk\" (UID: \"1a3cfa34-17ae-42d6-8ea7-46691bb10eaa\") " pod="openstack/barbican-api-54885d4b84-55wlk" Nov 28 13:38:36 crc kubenswrapper[4631]: I1128 13:38:36.860238 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1a3cfa34-17ae-42d6-8ea7-46691bb10eaa-internal-tls-certs\") pod \"barbican-api-54885d4b84-55wlk\" (UID: \"1a3cfa34-17ae-42d6-8ea7-46691bb10eaa\") " pod="openstack/barbican-api-54885d4b84-55wlk" Nov 28 13:38:36 crc kubenswrapper[4631]: I1128 13:38:36.860331 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1a3cfa34-17ae-42d6-8ea7-46691bb10eaa-public-tls-certs\") pod \"barbican-api-54885d4b84-55wlk\" (UID: \"1a3cfa34-17ae-42d6-8ea7-46691bb10eaa\") " pod="openstack/barbican-api-54885d4b84-55wlk" Nov 28 13:38:36 crc kubenswrapper[4631]: I1128 13:38:36.860366 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dpqjn\" (UniqueName: \"kubernetes.io/projected/1a3cfa34-17ae-42d6-8ea7-46691bb10eaa-kube-api-access-dpqjn\") pod \"barbican-api-54885d4b84-55wlk\" (UID: \"1a3cfa34-17ae-42d6-8ea7-46691bb10eaa\") " pod="openstack/barbican-api-54885d4b84-55wlk" Nov 28 13:38:36 crc kubenswrapper[4631]: I1128 13:38:36.860408 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a3cfa34-17ae-42d6-8ea7-46691bb10eaa-combined-ca-bundle\") pod \"barbican-api-54885d4b84-55wlk\" (UID: \"1a3cfa34-17ae-42d6-8ea7-46691bb10eaa\") " pod="openstack/barbican-api-54885d4b84-55wlk" Nov 28 13:38:36 crc kubenswrapper[4631]: I1128 13:38:36.860450 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1a3cfa34-17ae-42d6-8ea7-46691bb10eaa-config-data\") pod \"barbican-api-54885d4b84-55wlk\" (UID: \"1a3cfa34-17ae-42d6-8ea7-46691bb10eaa\") " pod="openstack/barbican-api-54885d4b84-55wlk" Nov 28 13:38:36 crc kubenswrapper[4631]: I1128 13:38:36.860485 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1a3cfa34-17ae-42d6-8ea7-46691bb10eaa-config-data-custom\") pod \"barbican-api-54885d4b84-55wlk\" (UID: \"1a3cfa34-17ae-42d6-8ea7-46691bb10eaa\") " pod="openstack/barbican-api-54885d4b84-55wlk" Nov 28 13:38:36 crc kubenswrapper[4631]: I1128 13:38:36.868656 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a3cfa34-17ae-42d6-8ea7-46691bb10eaa-combined-ca-bundle\") pod \"barbican-api-54885d4b84-55wlk\" (UID: \"1a3cfa34-17ae-42d6-8ea7-46691bb10eaa\") " pod="openstack/barbican-api-54885d4b84-55wlk" Nov 28 13:38:36 crc kubenswrapper[4631]: I1128 13:38:36.869401 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1a3cfa34-17ae-42d6-8ea7-46691bb10eaa-internal-tls-certs\") pod \"barbican-api-54885d4b84-55wlk\" (UID: \"1a3cfa34-17ae-42d6-8ea7-46691bb10eaa\") " pod="openstack/barbican-api-54885d4b84-55wlk" Nov 28 13:38:36 crc kubenswrapper[4631]: I1128 13:38:36.869950 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1a3cfa34-17ae-42d6-8ea7-46691bb10eaa-config-data-custom\") pod \"barbican-api-54885d4b84-55wlk\" (UID: \"1a3cfa34-17ae-42d6-8ea7-46691bb10eaa\") " pod="openstack/barbican-api-54885d4b84-55wlk" Nov 28 13:38:36 crc kubenswrapper[4631]: I1128 13:38:36.879184 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1a3cfa34-17ae-42d6-8ea7-46691bb10eaa-config-data\") pod \"barbican-api-54885d4b84-55wlk\" (UID: \"1a3cfa34-17ae-42d6-8ea7-46691bb10eaa\") " pod="openstack/barbican-api-54885d4b84-55wlk" Nov 28 13:38:36 crc kubenswrapper[4631]: I1128 13:38:36.871358 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1a3cfa34-17ae-42d6-8ea7-46691bb10eaa-public-tls-certs\") pod \"barbican-api-54885d4b84-55wlk\" (UID: \"1a3cfa34-17ae-42d6-8ea7-46691bb10eaa\") " pod="openstack/barbican-api-54885d4b84-55wlk" Nov 28 13:38:36 crc kubenswrapper[4631]: I1128 13:38:36.893071 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dpqjn\" (UniqueName: \"kubernetes.io/projected/1a3cfa34-17ae-42d6-8ea7-46691bb10eaa-kube-api-access-dpqjn\") pod \"barbican-api-54885d4b84-55wlk\" (UID: \"1a3cfa34-17ae-42d6-8ea7-46691bb10eaa\") " pod="openstack/barbican-api-54885d4b84-55wlk" Nov 28 13:38:36 crc kubenswrapper[4631]: I1128 13:38:36.984979 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-54885d4b84-55wlk" Nov 28 13:38:37 crc kubenswrapper[4631]: I1128 13:38:37.395577 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-54885d4b84-55wlk"] Nov 28 13:38:37 crc kubenswrapper[4631]: I1128 13:38:37.512771 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-54885d4b84-55wlk" event={"ID":"1a3cfa34-17ae-42d6-8ea7-46691bb10eaa","Type":"ContainerStarted","Data":"1dbd42f0a8fc97248ce0df485ad85d03022ff3326a0a5bf1f3c6e4b93531f66a"} Nov 28 13:38:37 crc kubenswrapper[4631]: I1128 13:38:37.545914 4631 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6a0a46ba-e48f-45a6-801e-7cb0ececabee" path="/var/lib/kubelet/pods/6a0a46ba-e48f-45a6-801e-7cb0ececabee/volumes" Nov 28 13:38:38 crc kubenswrapper[4631]: I1128 13:38:38.522707 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-54885d4b84-55wlk" event={"ID":"1a3cfa34-17ae-42d6-8ea7-46691bb10eaa","Type":"ContainerStarted","Data":"d04a5d47eb9a86f8f29a04fdaa6307e4bc8d0265876b8cea77475871dd3587f0"} Nov 28 13:38:38 crc kubenswrapper[4631]: I1128 13:38:38.523410 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-54885d4b84-55wlk" Nov 28 13:38:38 crc kubenswrapper[4631]: I1128 13:38:38.523428 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-54885d4b84-55wlk" Nov 28 13:38:38 crc kubenswrapper[4631]: I1128 13:38:38.523436 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-54885d4b84-55wlk" event={"ID":"1a3cfa34-17ae-42d6-8ea7-46691bb10eaa","Type":"ContainerStarted","Data":"32bccec78982ee37a407fae7ee05d8fd50876c783b9e79d021ada5ffc1fe7eaf"} Nov 28 13:38:38 crc kubenswrapper[4631]: I1128 13:38:38.544110 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-54885d4b84-55wlk" podStartSLOduration=2.544082157 podStartE2EDuration="2.544082157s" podCreationTimestamp="2025-11-28 13:38:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 13:38:38.542403196 +0000 UTC m=+1075.349706540" watchObservedRunningTime="2025-11-28 13:38:38.544082157 +0000 UTC m=+1075.351385501" Nov 28 13:38:39 crc kubenswrapper[4631]: I1128 13:38:39.572853 4631 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Nov 28 13:38:39 crc kubenswrapper[4631]: I1128 13:38:39.830043 4631 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Nov 28 13:38:40 crc kubenswrapper[4631]: I1128 13:38:40.344633 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5c9776ccc5-9ctzb" Nov 28 13:38:40 crc kubenswrapper[4631]: I1128 13:38:40.474277 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-l4h89"] Nov 28 13:38:40 crc kubenswrapper[4631]: I1128 13:38:40.475846 4631 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-55f844cf75-l4h89" podUID="83f364c4-bdf6-48ba-a5dc-31bb4fbb6f66" containerName="dnsmasq-dns" containerID="cri-o://ea19af2e351b50e1211a9c464d27d1c84979bdf113d275a5aed20975b3bff54b" gracePeriod=10 Nov 28 13:38:40 crc kubenswrapper[4631]: I1128 13:38:40.616499 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 28 13:38:41 crc kubenswrapper[4631]: I1128 13:38:41.090101 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55f844cf75-l4h89" Nov 28 13:38:41 crc kubenswrapper[4631]: I1128 13:38:41.278664 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fgjjx\" (UniqueName: \"kubernetes.io/projected/83f364c4-bdf6-48ba-a5dc-31bb4fbb6f66-kube-api-access-fgjjx\") pod \"83f364c4-bdf6-48ba-a5dc-31bb4fbb6f66\" (UID: \"83f364c4-bdf6-48ba-a5dc-31bb4fbb6f66\") " Nov 28 13:38:41 crc kubenswrapper[4631]: I1128 13:38:41.278855 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/83f364c4-bdf6-48ba-a5dc-31bb4fbb6f66-dns-svc\") pod \"83f364c4-bdf6-48ba-a5dc-31bb4fbb6f66\" (UID: \"83f364c4-bdf6-48ba-a5dc-31bb4fbb6f66\") " Nov 28 13:38:41 crc kubenswrapper[4631]: I1128 13:38:41.278942 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/83f364c4-bdf6-48ba-a5dc-31bb4fbb6f66-config\") pod \"83f364c4-bdf6-48ba-a5dc-31bb4fbb6f66\" (UID: \"83f364c4-bdf6-48ba-a5dc-31bb4fbb6f66\") " Nov 28 13:38:41 crc kubenswrapper[4631]: I1128 13:38:41.278991 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/83f364c4-bdf6-48ba-a5dc-31bb4fbb6f66-ovsdbserver-sb\") pod \"83f364c4-bdf6-48ba-a5dc-31bb4fbb6f66\" (UID: \"83f364c4-bdf6-48ba-a5dc-31bb4fbb6f66\") " Nov 28 13:38:41 crc kubenswrapper[4631]: I1128 13:38:41.279020 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/83f364c4-bdf6-48ba-a5dc-31bb4fbb6f66-ovsdbserver-nb\") pod \"83f364c4-bdf6-48ba-a5dc-31bb4fbb6f66\" (UID: \"83f364c4-bdf6-48ba-a5dc-31bb4fbb6f66\") " Nov 28 13:38:41 crc kubenswrapper[4631]: I1128 13:38:41.279065 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/83f364c4-bdf6-48ba-a5dc-31bb4fbb6f66-dns-swift-storage-0\") pod \"83f364c4-bdf6-48ba-a5dc-31bb4fbb6f66\" (UID: \"83f364c4-bdf6-48ba-a5dc-31bb4fbb6f66\") " Nov 28 13:38:41 crc kubenswrapper[4631]: I1128 13:38:41.307243 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/83f364c4-bdf6-48ba-a5dc-31bb4fbb6f66-kube-api-access-fgjjx" (OuterVolumeSpecName: "kube-api-access-fgjjx") pod "83f364c4-bdf6-48ba-a5dc-31bb4fbb6f66" (UID: "83f364c4-bdf6-48ba-a5dc-31bb4fbb6f66"). InnerVolumeSpecName "kube-api-access-fgjjx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:38:41 crc kubenswrapper[4631]: I1128 13:38:41.369022 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/83f364c4-bdf6-48ba-a5dc-31bb4fbb6f66-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "83f364c4-bdf6-48ba-a5dc-31bb4fbb6f66" (UID: "83f364c4-bdf6-48ba-a5dc-31bb4fbb6f66"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 13:38:41 crc kubenswrapper[4631]: I1128 13:38:41.378690 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/83f364c4-bdf6-48ba-a5dc-31bb4fbb6f66-config" (OuterVolumeSpecName: "config") pod "83f364c4-bdf6-48ba-a5dc-31bb4fbb6f66" (UID: "83f364c4-bdf6-48ba-a5dc-31bb4fbb6f66"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 13:38:41 crc kubenswrapper[4631]: I1128 13:38:41.381646 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fgjjx\" (UniqueName: \"kubernetes.io/projected/83f364c4-bdf6-48ba-a5dc-31bb4fbb6f66-kube-api-access-fgjjx\") on node \"crc\" DevicePath \"\"" Nov 28 13:38:41 crc kubenswrapper[4631]: I1128 13:38:41.381682 4631 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/83f364c4-bdf6-48ba-a5dc-31bb4fbb6f66-config\") on node \"crc\" DevicePath \"\"" Nov 28 13:38:41 crc kubenswrapper[4631]: I1128 13:38:41.381694 4631 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/83f364c4-bdf6-48ba-a5dc-31bb4fbb6f66-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 28 13:38:41 crc kubenswrapper[4631]: I1128 13:38:41.393428 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/83f364c4-bdf6-48ba-a5dc-31bb4fbb6f66-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "83f364c4-bdf6-48ba-a5dc-31bb4fbb6f66" (UID: "83f364c4-bdf6-48ba-a5dc-31bb4fbb6f66"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 13:38:41 crc kubenswrapper[4631]: I1128 13:38:41.411129 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/83f364c4-bdf6-48ba-a5dc-31bb4fbb6f66-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "83f364c4-bdf6-48ba-a5dc-31bb4fbb6f66" (UID: "83f364c4-bdf6-48ba-a5dc-31bb4fbb6f66"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 13:38:41 crc kubenswrapper[4631]: I1128 13:38:41.418917 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/83f364c4-bdf6-48ba-a5dc-31bb4fbb6f66-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "83f364c4-bdf6-48ba-a5dc-31bb4fbb6f66" (UID: "83f364c4-bdf6-48ba-a5dc-31bb4fbb6f66"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 13:38:41 crc kubenswrapper[4631]: I1128 13:38:41.484768 4631 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/83f364c4-bdf6-48ba-a5dc-31bb4fbb6f66-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 28 13:38:41 crc kubenswrapper[4631]: I1128 13:38:41.484819 4631 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/83f364c4-bdf6-48ba-a5dc-31bb4fbb6f66-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 28 13:38:41 crc kubenswrapper[4631]: I1128 13:38:41.484838 4631 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/83f364c4-bdf6-48ba-a5dc-31bb4fbb6f66-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 28 13:38:41 crc kubenswrapper[4631]: I1128 13:38:41.555122 4631 generic.go:334] "Generic (PLEG): container finished" podID="83f364c4-bdf6-48ba-a5dc-31bb4fbb6f66" containerID="ea19af2e351b50e1211a9c464d27d1c84979bdf113d275a5aed20975b3bff54b" exitCode=0 Nov 28 13:38:41 crc kubenswrapper[4631]: I1128 13:38:41.555247 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55f844cf75-l4h89" Nov 28 13:38:41 crc kubenswrapper[4631]: I1128 13:38:41.555245 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55f844cf75-l4h89" event={"ID":"83f364c4-bdf6-48ba-a5dc-31bb4fbb6f66","Type":"ContainerDied","Data":"ea19af2e351b50e1211a9c464d27d1c84979bdf113d275a5aed20975b3bff54b"} Nov 28 13:38:41 crc kubenswrapper[4631]: I1128 13:38:41.555341 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55f844cf75-l4h89" event={"ID":"83f364c4-bdf6-48ba-a5dc-31bb4fbb6f66","Type":"ContainerDied","Data":"794b95d779176e79e82c45e7d2b377761a40a354cfd5048bb8e41b6c642b84d7"} Nov 28 13:38:41 crc kubenswrapper[4631]: I1128 13:38:41.555372 4631 scope.go:117] "RemoveContainer" containerID="ea19af2e351b50e1211a9c464d27d1c84979bdf113d275a5aed20975b3bff54b" Nov 28 13:38:41 crc kubenswrapper[4631]: I1128 13:38:41.555397 4631 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="8f2270d7-d4a8-4cf5-a0b4-b5c29f4fb817" containerName="cinder-scheduler" containerID="cri-o://598d394db8cefb7d6686ceb6bdb8cc64bd0da8e045b347b40219b4b81fa90967" gracePeriod=30 Nov 28 13:38:41 crc kubenswrapper[4631]: I1128 13:38:41.555519 4631 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="8f2270d7-d4a8-4cf5-a0b4-b5c29f4fb817" containerName="probe" containerID="cri-o://c79f1d7d0649f621429d5f1b6d86425221cf1a427688fb118f898b252947cec4" gracePeriod=30 Nov 28 13:38:41 crc kubenswrapper[4631]: I1128 13:38:41.597942 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-l4h89"] Nov 28 13:38:41 crc kubenswrapper[4631]: I1128 13:38:41.600035 4631 scope.go:117] "RemoveContainer" containerID="2f977a596e6188962f616acbe2f9777643926d465cba41a2749970823e0c0dd1" Nov 28 13:38:41 crc kubenswrapper[4631]: I1128 13:38:41.608455 4631 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-l4h89"] Nov 28 13:38:41 crc kubenswrapper[4631]: I1128 13:38:41.670349 4631 scope.go:117] "RemoveContainer" containerID="ea19af2e351b50e1211a9c464d27d1c84979bdf113d275a5aed20975b3bff54b" Nov 28 13:38:41 crc kubenswrapper[4631]: E1128 13:38:41.675224 4631 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ea19af2e351b50e1211a9c464d27d1c84979bdf113d275a5aed20975b3bff54b\": container with ID starting with ea19af2e351b50e1211a9c464d27d1c84979bdf113d275a5aed20975b3bff54b not found: ID does not exist" containerID="ea19af2e351b50e1211a9c464d27d1c84979bdf113d275a5aed20975b3bff54b" Nov 28 13:38:41 crc kubenswrapper[4631]: I1128 13:38:41.675295 4631 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ea19af2e351b50e1211a9c464d27d1c84979bdf113d275a5aed20975b3bff54b"} err="failed to get container status \"ea19af2e351b50e1211a9c464d27d1c84979bdf113d275a5aed20975b3bff54b\": rpc error: code = NotFound desc = could not find container \"ea19af2e351b50e1211a9c464d27d1c84979bdf113d275a5aed20975b3bff54b\": container with ID starting with ea19af2e351b50e1211a9c464d27d1c84979bdf113d275a5aed20975b3bff54b not found: ID does not exist" Nov 28 13:38:41 crc kubenswrapper[4631]: I1128 13:38:41.675321 4631 scope.go:117] "RemoveContainer" containerID="2f977a596e6188962f616acbe2f9777643926d465cba41a2749970823e0c0dd1" Nov 28 13:38:41 crc kubenswrapper[4631]: E1128 13:38:41.679069 4631 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2f977a596e6188962f616acbe2f9777643926d465cba41a2749970823e0c0dd1\": container with ID starting with 2f977a596e6188962f616acbe2f9777643926d465cba41a2749970823e0c0dd1 not found: ID does not exist" containerID="2f977a596e6188962f616acbe2f9777643926d465cba41a2749970823e0c0dd1" Nov 28 13:38:41 crc kubenswrapper[4631]: I1128 13:38:41.679110 4631 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2f977a596e6188962f616acbe2f9777643926d465cba41a2749970823e0c0dd1"} err="failed to get container status \"2f977a596e6188962f616acbe2f9777643926d465cba41a2749970823e0c0dd1\": rpc error: code = NotFound desc = could not find container \"2f977a596e6188962f616acbe2f9777643926d465cba41a2749970823e0c0dd1\": container with ID starting with 2f977a596e6188962f616acbe2f9777643926d465cba41a2749970823e0c0dd1 not found: ID does not exist" Nov 28 13:38:42 crc kubenswrapper[4631]: I1128 13:38:42.155130 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-756d64c4c8-jzvlv" Nov 28 13:38:42 crc kubenswrapper[4631]: I1128 13:38:42.326693 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-756d64c4c8-jzvlv" Nov 28 13:38:43 crc kubenswrapper[4631]: I1128 13:38:43.562768 4631 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="83f364c4-bdf6-48ba-a5dc-31bb4fbb6f66" path="/var/lib/kubelet/pods/83f364c4-bdf6-48ba-a5dc-31bb4fbb6f66/volumes" Nov 28 13:38:43 crc kubenswrapper[4631]: I1128 13:38:43.606568 4631 generic.go:334] "Generic (PLEG): container finished" podID="8f2270d7-d4a8-4cf5-a0b4-b5c29f4fb817" containerID="c79f1d7d0649f621429d5f1b6d86425221cf1a427688fb118f898b252947cec4" exitCode=0 Nov 28 13:38:43 crc kubenswrapper[4631]: I1128 13:38:43.606630 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"8f2270d7-d4a8-4cf5-a0b4-b5c29f4fb817","Type":"ContainerDied","Data":"c79f1d7d0649f621429d5f1b6d86425221cf1a427688fb118f898b252947cec4"} Nov 28 13:38:43 crc kubenswrapper[4631]: I1128 13:38:43.934426 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Nov 28 13:38:44 crc kubenswrapper[4631]: I1128 13:38:44.431559 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 28 13:38:44 crc kubenswrapper[4631]: I1128 13:38:44.573255 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8f2270d7-d4a8-4cf5-a0b4-b5c29f4fb817-scripts\") pod \"8f2270d7-d4a8-4cf5-a0b4-b5c29f4fb817\" (UID: \"8f2270d7-d4a8-4cf5-a0b4-b5c29f4fb817\") " Nov 28 13:38:44 crc kubenswrapper[4631]: I1128 13:38:44.573381 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8f2270d7-d4a8-4cf5-a0b4-b5c29f4fb817-config-data-custom\") pod \"8f2270d7-d4a8-4cf5-a0b4-b5c29f4fb817\" (UID: \"8f2270d7-d4a8-4cf5-a0b4-b5c29f4fb817\") " Nov 28 13:38:44 crc kubenswrapper[4631]: I1128 13:38:44.573478 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8f2270d7-d4a8-4cf5-a0b4-b5c29f4fb817-etc-machine-id\") pod \"8f2270d7-d4a8-4cf5-a0b4-b5c29f4fb817\" (UID: \"8f2270d7-d4a8-4cf5-a0b4-b5c29f4fb817\") " Nov 28 13:38:44 crc kubenswrapper[4631]: I1128 13:38:44.573545 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kpkhv\" (UniqueName: \"kubernetes.io/projected/8f2270d7-d4a8-4cf5-a0b4-b5c29f4fb817-kube-api-access-kpkhv\") pod \"8f2270d7-d4a8-4cf5-a0b4-b5c29f4fb817\" (UID: \"8f2270d7-d4a8-4cf5-a0b4-b5c29f4fb817\") " Nov 28 13:38:44 crc kubenswrapper[4631]: I1128 13:38:44.573567 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8f2270d7-d4a8-4cf5-a0b4-b5c29f4fb817-combined-ca-bundle\") pod \"8f2270d7-d4a8-4cf5-a0b4-b5c29f4fb817\" (UID: \"8f2270d7-d4a8-4cf5-a0b4-b5c29f4fb817\") " Nov 28 13:38:44 crc kubenswrapper[4631]: I1128 13:38:44.573735 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8f2270d7-d4a8-4cf5-a0b4-b5c29f4fb817-config-data\") pod \"8f2270d7-d4a8-4cf5-a0b4-b5c29f4fb817\" (UID: \"8f2270d7-d4a8-4cf5-a0b4-b5c29f4fb817\") " Nov 28 13:38:44 crc kubenswrapper[4631]: I1128 13:38:44.574330 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8f2270d7-d4a8-4cf5-a0b4-b5c29f4fb817-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "8f2270d7-d4a8-4cf5-a0b4-b5c29f4fb817" (UID: "8f2270d7-d4a8-4cf5-a0b4-b5c29f4fb817"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 13:38:44 crc kubenswrapper[4631]: I1128 13:38:44.574923 4631 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8f2270d7-d4a8-4cf5-a0b4-b5c29f4fb817-etc-machine-id\") on node \"crc\" DevicePath \"\"" Nov 28 13:38:44 crc kubenswrapper[4631]: I1128 13:38:44.581930 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f2270d7-d4a8-4cf5-a0b4-b5c29f4fb817-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "8f2270d7-d4a8-4cf5-a0b4-b5c29f4fb817" (UID: "8f2270d7-d4a8-4cf5-a0b4-b5c29f4fb817"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:38:44 crc kubenswrapper[4631]: I1128 13:38:44.584426 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f2270d7-d4a8-4cf5-a0b4-b5c29f4fb817-kube-api-access-kpkhv" (OuterVolumeSpecName: "kube-api-access-kpkhv") pod "8f2270d7-d4a8-4cf5-a0b4-b5c29f4fb817" (UID: "8f2270d7-d4a8-4cf5-a0b4-b5c29f4fb817"). InnerVolumeSpecName "kube-api-access-kpkhv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:38:44 crc kubenswrapper[4631]: I1128 13:38:44.584478 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f2270d7-d4a8-4cf5-a0b4-b5c29f4fb817-scripts" (OuterVolumeSpecName: "scripts") pod "8f2270d7-d4a8-4cf5-a0b4-b5c29f4fb817" (UID: "8f2270d7-d4a8-4cf5-a0b4-b5c29f4fb817"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:38:44 crc kubenswrapper[4631]: I1128 13:38:44.621022 4631 generic.go:334] "Generic (PLEG): container finished" podID="8f2270d7-d4a8-4cf5-a0b4-b5c29f4fb817" containerID="598d394db8cefb7d6686ceb6bdb8cc64bd0da8e045b347b40219b4b81fa90967" exitCode=0 Nov 28 13:38:44 crc kubenswrapper[4631]: I1128 13:38:44.621123 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 28 13:38:44 crc kubenswrapper[4631]: I1128 13:38:44.621113 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"8f2270d7-d4a8-4cf5-a0b4-b5c29f4fb817","Type":"ContainerDied","Data":"598d394db8cefb7d6686ceb6bdb8cc64bd0da8e045b347b40219b4b81fa90967"} Nov 28 13:38:44 crc kubenswrapper[4631]: I1128 13:38:44.621262 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"8f2270d7-d4a8-4cf5-a0b4-b5c29f4fb817","Type":"ContainerDied","Data":"2749c4cccac47c4ccc8000027598e38875d27778aaf92fd30117a5f0021ed5c4"} Nov 28 13:38:44 crc kubenswrapper[4631]: I1128 13:38:44.621436 4631 scope.go:117] "RemoveContainer" containerID="c79f1d7d0649f621429d5f1b6d86425221cf1a427688fb118f898b252947cec4" Nov 28 13:38:44 crc kubenswrapper[4631]: I1128 13:38:44.646359 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f2270d7-d4a8-4cf5-a0b4-b5c29f4fb817-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8f2270d7-d4a8-4cf5-a0b4-b5c29f4fb817" (UID: "8f2270d7-d4a8-4cf5-a0b4-b5c29f4fb817"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:38:44 crc kubenswrapper[4631]: I1128 13:38:44.677584 4631 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8f2270d7-d4a8-4cf5-a0b4-b5c29f4fb817-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 13:38:44 crc kubenswrapper[4631]: I1128 13:38:44.677637 4631 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8f2270d7-d4a8-4cf5-a0b4-b5c29f4fb817-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 28 13:38:44 crc kubenswrapper[4631]: I1128 13:38:44.677656 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kpkhv\" (UniqueName: \"kubernetes.io/projected/8f2270d7-d4a8-4cf5-a0b4-b5c29f4fb817-kube-api-access-kpkhv\") on node \"crc\" DevicePath \"\"" Nov 28 13:38:44 crc kubenswrapper[4631]: I1128 13:38:44.677668 4631 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8f2270d7-d4a8-4cf5-a0b4-b5c29f4fb817-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 13:38:44 crc kubenswrapper[4631]: I1128 13:38:44.687279 4631 scope.go:117] "RemoveContainer" containerID="598d394db8cefb7d6686ceb6bdb8cc64bd0da8e045b347b40219b4b81fa90967" Nov 28 13:38:44 crc kubenswrapper[4631]: I1128 13:38:44.726714 4631 scope.go:117] "RemoveContainer" containerID="c79f1d7d0649f621429d5f1b6d86425221cf1a427688fb118f898b252947cec4" Nov 28 13:38:44 crc kubenswrapper[4631]: E1128 13:38:44.727387 4631 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c79f1d7d0649f621429d5f1b6d86425221cf1a427688fb118f898b252947cec4\": container with ID starting with c79f1d7d0649f621429d5f1b6d86425221cf1a427688fb118f898b252947cec4 not found: ID does not exist" containerID="c79f1d7d0649f621429d5f1b6d86425221cf1a427688fb118f898b252947cec4" Nov 28 13:38:44 crc kubenswrapper[4631]: I1128 13:38:44.727459 4631 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c79f1d7d0649f621429d5f1b6d86425221cf1a427688fb118f898b252947cec4"} err="failed to get container status \"c79f1d7d0649f621429d5f1b6d86425221cf1a427688fb118f898b252947cec4\": rpc error: code = NotFound desc = could not find container \"c79f1d7d0649f621429d5f1b6d86425221cf1a427688fb118f898b252947cec4\": container with ID starting with c79f1d7d0649f621429d5f1b6d86425221cf1a427688fb118f898b252947cec4 not found: ID does not exist" Nov 28 13:38:44 crc kubenswrapper[4631]: I1128 13:38:44.727504 4631 scope.go:117] "RemoveContainer" containerID="598d394db8cefb7d6686ceb6bdb8cc64bd0da8e045b347b40219b4b81fa90967" Nov 28 13:38:44 crc kubenswrapper[4631]: E1128 13:38:44.728067 4631 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"598d394db8cefb7d6686ceb6bdb8cc64bd0da8e045b347b40219b4b81fa90967\": container with ID starting with 598d394db8cefb7d6686ceb6bdb8cc64bd0da8e045b347b40219b4b81fa90967 not found: ID does not exist" containerID="598d394db8cefb7d6686ceb6bdb8cc64bd0da8e045b347b40219b4b81fa90967" Nov 28 13:38:44 crc kubenswrapper[4631]: I1128 13:38:44.728105 4631 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"598d394db8cefb7d6686ceb6bdb8cc64bd0da8e045b347b40219b4b81fa90967"} err="failed to get container status \"598d394db8cefb7d6686ceb6bdb8cc64bd0da8e045b347b40219b4b81fa90967\": rpc error: code = NotFound desc = could not find container \"598d394db8cefb7d6686ceb6bdb8cc64bd0da8e045b347b40219b4b81fa90967\": container with ID starting with 598d394db8cefb7d6686ceb6bdb8cc64bd0da8e045b347b40219b4b81fa90967 not found: ID does not exist" Nov 28 13:38:44 crc kubenswrapper[4631]: I1128 13:38:44.729349 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f2270d7-d4a8-4cf5-a0b4-b5c29f4fb817-config-data" (OuterVolumeSpecName: "config-data") pod "8f2270d7-d4a8-4cf5-a0b4-b5c29f4fb817" (UID: "8f2270d7-d4a8-4cf5-a0b4-b5c29f4fb817"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:38:44 crc kubenswrapper[4631]: I1128 13:38:44.779886 4631 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8f2270d7-d4a8-4cf5-a0b4-b5c29f4fb817-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 13:38:44 crc kubenswrapper[4631]: I1128 13:38:44.967077 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 28 13:38:44 crc kubenswrapper[4631]: I1128 13:38:44.978436 4631 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 28 13:38:44 crc kubenswrapper[4631]: I1128 13:38:44.990816 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Nov 28 13:38:44 crc kubenswrapper[4631]: E1128 13:38:44.991370 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8f2270d7-d4a8-4cf5-a0b4-b5c29f4fb817" containerName="probe" Nov 28 13:38:44 crc kubenswrapper[4631]: I1128 13:38:44.991396 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="8f2270d7-d4a8-4cf5-a0b4-b5c29f4fb817" containerName="probe" Nov 28 13:38:44 crc kubenswrapper[4631]: E1128 13:38:44.991427 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="83f364c4-bdf6-48ba-a5dc-31bb4fbb6f66" containerName="init" Nov 28 13:38:44 crc kubenswrapper[4631]: I1128 13:38:44.991436 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="83f364c4-bdf6-48ba-a5dc-31bb4fbb6f66" containerName="init" Nov 28 13:38:44 crc kubenswrapper[4631]: E1128 13:38:44.991461 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="83f364c4-bdf6-48ba-a5dc-31bb4fbb6f66" containerName="dnsmasq-dns" Nov 28 13:38:44 crc kubenswrapper[4631]: I1128 13:38:44.991472 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="83f364c4-bdf6-48ba-a5dc-31bb4fbb6f66" containerName="dnsmasq-dns" Nov 28 13:38:44 crc kubenswrapper[4631]: E1128 13:38:44.991495 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8f2270d7-d4a8-4cf5-a0b4-b5c29f4fb817" containerName="cinder-scheduler" Nov 28 13:38:44 crc kubenswrapper[4631]: I1128 13:38:44.991506 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="8f2270d7-d4a8-4cf5-a0b4-b5c29f4fb817" containerName="cinder-scheduler" Nov 28 13:38:44 crc kubenswrapper[4631]: I1128 13:38:44.991732 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="8f2270d7-d4a8-4cf5-a0b4-b5c29f4fb817" containerName="probe" Nov 28 13:38:44 crc kubenswrapper[4631]: I1128 13:38:44.991769 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="8f2270d7-d4a8-4cf5-a0b4-b5c29f4fb817" containerName="cinder-scheduler" Nov 28 13:38:44 crc kubenswrapper[4631]: I1128 13:38:44.991796 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="83f364c4-bdf6-48ba-a5dc-31bb4fbb6f66" containerName="dnsmasq-dns" Nov 28 13:38:44 crc kubenswrapper[4631]: I1128 13:38:44.993938 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 28 13:38:44 crc kubenswrapper[4631]: I1128 13:38:44.997845 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Nov 28 13:38:45 crc kubenswrapper[4631]: I1128 13:38:45.001599 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 28 13:38:45 crc kubenswrapper[4631]: I1128 13:38:45.188148 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c4be14a0-f1a1-46f3-86ea-48e7dfb8da00-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"c4be14a0-f1a1-46f3-86ea-48e7dfb8da00\") " pod="openstack/cinder-scheduler-0" Nov 28 13:38:45 crc kubenswrapper[4631]: I1128 13:38:45.188207 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c4be14a0-f1a1-46f3-86ea-48e7dfb8da00-config-data\") pod \"cinder-scheduler-0\" (UID: \"c4be14a0-f1a1-46f3-86ea-48e7dfb8da00\") " pod="openstack/cinder-scheduler-0" Nov 28 13:38:45 crc kubenswrapper[4631]: I1128 13:38:45.188306 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-plrpj\" (UniqueName: \"kubernetes.io/projected/c4be14a0-f1a1-46f3-86ea-48e7dfb8da00-kube-api-access-plrpj\") pod \"cinder-scheduler-0\" (UID: \"c4be14a0-f1a1-46f3-86ea-48e7dfb8da00\") " pod="openstack/cinder-scheduler-0" Nov 28 13:38:45 crc kubenswrapper[4631]: I1128 13:38:45.188367 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c4be14a0-f1a1-46f3-86ea-48e7dfb8da00-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"c4be14a0-f1a1-46f3-86ea-48e7dfb8da00\") " pod="openstack/cinder-scheduler-0" Nov 28 13:38:45 crc kubenswrapper[4631]: I1128 13:38:45.188403 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c4be14a0-f1a1-46f3-86ea-48e7dfb8da00-scripts\") pod \"cinder-scheduler-0\" (UID: \"c4be14a0-f1a1-46f3-86ea-48e7dfb8da00\") " pod="openstack/cinder-scheduler-0" Nov 28 13:38:45 crc kubenswrapper[4631]: I1128 13:38:45.188439 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c4be14a0-f1a1-46f3-86ea-48e7dfb8da00-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"c4be14a0-f1a1-46f3-86ea-48e7dfb8da00\") " pod="openstack/cinder-scheduler-0" Nov 28 13:38:45 crc kubenswrapper[4631]: I1128 13:38:45.290917 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-plrpj\" (UniqueName: \"kubernetes.io/projected/c4be14a0-f1a1-46f3-86ea-48e7dfb8da00-kube-api-access-plrpj\") pod \"cinder-scheduler-0\" (UID: \"c4be14a0-f1a1-46f3-86ea-48e7dfb8da00\") " pod="openstack/cinder-scheduler-0" Nov 28 13:38:45 crc kubenswrapper[4631]: I1128 13:38:45.291221 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c4be14a0-f1a1-46f3-86ea-48e7dfb8da00-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"c4be14a0-f1a1-46f3-86ea-48e7dfb8da00\") " pod="openstack/cinder-scheduler-0" Nov 28 13:38:45 crc kubenswrapper[4631]: I1128 13:38:45.291358 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c4be14a0-f1a1-46f3-86ea-48e7dfb8da00-scripts\") pod \"cinder-scheduler-0\" (UID: \"c4be14a0-f1a1-46f3-86ea-48e7dfb8da00\") " pod="openstack/cinder-scheduler-0" Nov 28 13:38:45 crc kubenswrapper[4631]: I1128 13:38:45.291447 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c4be14a0-f1a1-46f3-86ea-48e7dfb8da00-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"c4be14a0-f1a1-46f3-86ea-48e7dfb8da00\") " pod="openstack/cinder-scheduler-0" Nov 28 13:38:45 crc kubenswrapper[4631]: I1128 13:38:45.291549 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c4be14a0-f1a1-46f3-86ea-48e7dfb8da00-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"c4be14a0-f1a1-46f3-86ea-48e7dfb8da00\") " pod="openstack/cinder-scheduler-0" Nov 28 13:38:45 crc kubenswrapper[4631]: I1128 13:38:45.291769 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c4be14a0-f1a1-46f3-86ea-48e7dfb8da00-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"c4be14a0-f1a1-46f3-86ea-48e7dfb8da00\") " pod="openstack/cinder-scheduler-0" Nov 28 13:38:45 crc kubenswrapper[4631]: I1128 13:38:45.291855 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c4be14a0-f1a1-46f3-86ea-48e7dfb8da00-config-data\") pod \"cinder-scheduler-0\" (UID: \"c4be14a0-f1a1-46f3-86ea-48e7dfb8da00\") " pod="openstack/cinder-scheduler-0" Nov 28 13:38:45 crc kubenswrapper[4631]: I1128 13:38:45.298191 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c4be14a0-f1a1-46f3-86ea-48e7dfb8da00-config-data\") pod \"cinder-scheduler-0\" (UID: \"c4be14a0-f1a1-46f3-86ea-48e7dfb8da00\") " pod="openstack/cinder-scheduler-0" Nov 28 13:38:45 crc kubenswrapper[4631]: I1128 13:38:45.299017 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c4be14a0-f1a1-46f3-86ea-48e7dfb8da00-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"c4be14a0-f1a1-46f3-86ea-48e7dfb8da00\") " pod="openstack/cinder-scheduler-0" Nov 28 13:38:45 crc kubenswrapper[4631]: I1128 13:38:45.300894 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c4be14a0-f1a1-46f3-86ea-48e7dfb8da00-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"c4be14a0-f1a1-46f3-86ea-48e7dfb8da00\") " pod="openstack/cinder-scheduler-0" Nov 28 13:38:45 crc kubenswrapper[4631]: I1128 13:38:45.308859 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c4be14a0-f1a1-46f3-86ea-48e7dfb8da00-scripts\") pod \"cinder-scheduler-0\" (UID: \"c4be14a0-f1a1-46f3-86ea-48e7dfb8da00\") " pod="openstack/cinder-scheduler-0" Nov 28 13:38:45 crc kubenswrapper[4631]: I1128 13:38:45.329653 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-plrpj\" (UniqueName: \"kubernetes.io/projected/c4be14a0-f1a1-46f3-86ea-48e7dfb8da00-kube-api-access-plrpj\") pod \"cinder-scheduler-0\" (UID: \"c4be14a0-f1a1-46f3-86ea-48e7dfb8da00\") " pod="openstack/cinder-scheduler-0" Nov 28 13:38:45 crc kubenswrapper[4631]: I1128 13:38:45.461096 4631 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="d95d3536-1392-4576-8ee5-390c09fe7a49" containerName="proxy-httpd" probeResult="failure" output="HTTP probe failed with statuscode: 503" Nov 28 13:38:45 crc kubenswrapper[4631]: I1128 13:38:45.526095 4631 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f2270d7-d4a8-4cf5-a0b4-b5c29f4fb817" path="/var/lib/kubelet/pods/8f2270d7-d4a8-4cf5-a0b4-b5c29f4fb817/volumes" Nov 28 13:38:45 crc kubenswrapper[4631]: I1128 13:38:45.611099 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 28 13:38:46 crc kubenswrapper[4631]: I1128 13:38:46.212271 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 28 13:38:46 crc kubenswrapper[4631]: I1128 13:38:46.666217 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"c4be14a0-f1a1-46f3-86ea-48e7dfb8da00","Type":"ContainerStarted","Data":"a67fd470f09338363999192c7e04ecdedb116e74300c41ccb227107c1abb2625"} Nov 28 13:38:47 crc kubenswrapper[4631]: I1128 13:38:47.700424 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"c4be14a0-f1a1-46f3-86ea-48e7dfb8da00","Type":"ContainerStarted","Data":"ec99d80ee4dae110e3069ca7f738df551c5ddbfed122e27c710f6aefb421b929"} Nov 28 13:38:48 crc kubenswrapper[4631]: I1128 13:38:48.202952 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-769746444-8zhp8" Nov 28 13:38:48 crc kubenswrapper[4631]: I1128 13:38:48.467199 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-769746444-8zhp8" Nov 28 13:38:48 crc kubenswrapper[4631]: I1128 13:38:48.715366 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"c4be14a0-f1a1-46f3-86ea-48e7dfb8da00","Type":"ContainerStarted","Data":"236582ef2c92786cf0c304e860fddb47fdd97af5700c0d2475bec3631af87487"} Nov 28 13:38:48 crc kubenswrapper[4631]: I1128 13:38:48.739572 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=4.739547878 podStartE2EDuration="4.739547878s" podCreationTimestamp="2025-11-28 13:38:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 13:38:48.735402286 +0000 UTC m=+1085.542705630" watchObservedRunningTime="2025-11-28 13:38:48.739547878 +0000 UTC m=+1085.546851222" Nov 28 13:38:50 crc kubenswrapper[4631]: I1128 13:38:50.018050 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-6bd6cfc878-65bg5" Nov 28 13:38:50 crc kubenswrapper[4631]: I1128 13:38:50.292900 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-54885d4b84-55wlk" Nov 28 13:38:50 crc kubenswrapper[4631]: I1128 13:38:50.611473 4631 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Nov 28 13:38:50 crc kubenswrapper[4631]: I1128 13:38:50.898089 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-54885d4b84-55wlk" Nov 28 13:38:50 crc kubenswrapper[4631]: I1128 13:38:50.987171 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-756d64c4c8-jzvlv"] Nov 28 13:38:50 crc kubenswrapper[4631]: I1128 13:38:50.987456 4631 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-756d64c4c8-jzvlv" podUID="7eea7f3e-fc3c-4c60-998e-79c9a31ee084" containerName="barbican-api-log" containerID="cri-o://39f6239d34a7e4114e98f79691b15de6581a312f79450b6aa821ddc52f868ed0" gracePeriod=30 Nov 28 13:38:50 crc kubenswrapper[4631]: I1128 13:38:50.987940 4631 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-756d64c4c8-jzvlv" podUID="7eea7f3e-fc3c-4c60-998e-79c9a31ee084" containerName="barbican-api" containerID="cri-o://c9851fdf444e8d0158c79114f4288e3ea2c1a53294a47a6cfe8d648dde4e2e5f" gracePeriod=30 Nov 28 13:38:51 crc kubenswrapper[4631]: I1128 13:38:51.746767 4631 generic.go:334] "Generic (PLEG): container finished" podID="7eea7f3e-fc3c-4c60-998e-79c9a31ee084" containerID="39f6239d34a7e4114e98f79691b15de6581a312f79450b6aa821ddc52f868ed0" exitCode=143 Nov 28 13:38:51 crc kubenswrapper[4631]: I1128 13:38:51.746827 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-756d64c4c8-jzvlv" event={"ID":"7eea7f3e-fc3c-4c60-998e-79c9a31ee084","Type":"ContainerDied","Data":"39f6239d34a7e4114e98f79691b15de6581a312f79450b6aa821ddc52f868ed0"} Nov 28 13:38:52 crc kubenswrapper[4631]: I1128 13:38:52.621448 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Nov 28 13:38:52 crc kubenswrapper[4631]: I1128 13:38:52.623599 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 28 13:38:52 crc kubenswrapper[4631]: I1128 13:38:52.627371 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Nov 28 13:38:52 crc kubenswrapper[4631]: I1128 13:38:52.627407 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-nkjzk" Nov 28 13:38:52 crc kubenswrapper[4631]: I1128 13:38:52.633155 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Nov 28 13:38:52 crc kubenswrapper[4631]: I1128 13:38:52.676738 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Nov 28 13:38:52 crc kubenswrapper[4631]: I1128 13:38:52.700502 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bszlc\" (UniqueName: \"kubernetes.io/projected/2eee8ee8-0983-43e2-85d6-7d0e274a71d5-kube-api-access-bszlc\") pod \"openstackclient\" (UID: \"2eee8ee8-0983-43e2-85d6-7d0e274a71d5\") " pod="openstack/openstackclient" Nov 28 13:38:52 crc kubenswrapper[4631]: I1128 13:38:52.700626 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/2eee8ee8-0983-43e2-85d6-7d0e274a71d5-openstack-config\") pod \"openstackclient\" (UID: \"2eee8ee8-0983-43e2-85d6-7d0e274a71d5\") " pod="openstack/openstackclient" Nov 28 13:38:52 crc kubenswrapper[4631]: I1128 13:38:52.700666 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2eee8ee8-0983-43e2-85d6-7d0e274a71d5-combined-ca-bundle\") pod \"openstackclient\" (UID: \"2eee8ee8-0983-43e2-85d6-7d0e274a71d5\") " pod="openstack/openstackclient" Nov 28 13:38:52 crc kubenswrapper[4631]: I1128 13:38:52.700752 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/2eee8ee8-0983-43e2-85d6-7d0e274a71d5-openstack-config-secret\") pod \"openstackclient\" (UID: \"2eee8ee8-0983-43e2-85d6-7d0e274a71d5\") " pod="openstack/openstackclient" Nov 28 13:38:52 crc kubenswrapper[4631]: I1128 13:38:52.802559 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bszlc\" (UniqueName: \"kubernetes.io/projected/2eee8ee8-0983-43e2-85d6-7d0e274a71d5-kube-api-access-bszlc\") pod \"openstackclient\" (UID: \"2eee8ee8-0983-43e2-85d6-7d0e274a71d5\") " pod="openstack/openstackclient" Nov 28 13:38:52 crc kubenswrapper[4631]: I1128 13:38:52.804431 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/2eee8ee8-0983-43e2-85d6-7d0e274a71d5-openstack-config\") pod \"openstackclient\" (UID: \"2eee8ee8-0983-43e2-85d6-7d0e274a71d5\") " pod="openstack/openstackclient" Nov 28 13:38:52 crc kubenswrapper[4631]: I1128 13:38:52.804490 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2eee8ee8-0983-43e2-85d6-7d0e274a71d5-combined-ca-bundle\") pod \"openstackclient\" (UID: \"2eee8ee8-0983-43e2-85d6-7d0e274a71d5\") " pod="openstack/openstackclient" Nov 28 13:38:52 crc kubenswrapper[4631]: I1128 13:38:52.804591 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/2eee8ee8-0983-43e2-85d6-7d0e274a71d5-openstack-config-secret\") pod \"openstackclient\" (UID: \"2eee8ee8-0983-43e2-85d6-7d0e274a71d5\") " pod="openstack/openstackclient" Nov 28 13:38:52 crc kubenswrapper[4631]: I1128 13:38:52.806907 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/2eee8ee8-0983-43e2-85d6-7d0e274a71d5-openstack-config\") pod \"openstackclient\" (UID: \"2eee8ee8-0983-43e2-85d6-7d0e274a71d5\") " pod="openstack/openstackclient" Nov 28 13:38:52 crc kubenswrapper[4631]: I1128 13:38:52.812202 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/2eee8ee8-0983-43e2-85d6-7d0e274a71d5-openstack-config-secret\") pod \"openstackclient\" (UID: \"2eee8ee8-0983-43e2-85d6-7d0e274a71d5\") " pod="openstack/openstackclient" Nov 28 13:38:52 crc kubenswrapper[4631]: I1128 13:38:52.828872 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2eee8ee8-0983-43e2-85d6-7d0e274a71d5-combined-ca-bundle\") pod \"openstackclient\" (UID: \"2eee8ee8-0983-43e2-85d6-7d0e274a71d5\") " pod="openstack/openstackclient" Nov 28 13:38:52 crc kubenswrapper[4631]: I1128 13:38:52.829050 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bszlc\" (UniqueName: \"kubernetes.io/projected/2eee8ee8-0983-43e2-85d6-7d0e274a71d5-kube-api-access-bszlc\") pod \"openstackclient\" (UID: \"2eee8ee8-0983-43e2-85d6-7d0e274a71d5\") " pod="openstack/openstackclient" Nov 28 13:38:52 crc kubenswrapper[4631]: I1128 13:38:52.994624 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 28 13:38:53 crc kubenswrapper[4631]: I1128 13:38:53.031016 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstackclient"] Nov 28 13:38:53 crc kubenswrapper[4631]: I1128 13:38:53.065969 4631 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstackclient"] Nov 28 13:38:53 crc kubenswrapper[4631]: I1128 13:38:53.145050 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Nov 28 13:38:53 crc kubenswrapper[4631]: I1128 13:38:53.146566 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 28 13:38:53 crc kubenswrapper[4631]: I1128 13:38:53.165182 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Nov 28 13:38:53 crc kubenswrapper[4631]: I1128 13:38:53.213966 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ea063a3c-330a-4642-a815-c013ae6b80c7-combined-ca-bundle\") pod \"openstackclient\" (UID: \"ea063a3c-330a-4642-a815-c013ae6b80c7\") " pod="openstack/openstackclient" Nov 28 13:38:53 crc kubenswrapper[4631]: I1128 13:38:53.214104 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/ea063a3c-330a-4642-a815-c013ae6b80c7-openstack-config\") pod \"openstackclient\" (UID: \"ea063a3c-330a-4642-a815-c013ae6b80c7\") " pod="openstack/openstackclient" Nov 28 13:38:53 crc kubenswrapper[4631]: I1128 13:38:53.214158 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qwv27\" (UniqueName: \"kubernetes.io/projected/ea063a3c-330a-4642-a815-c013ae6b80c7-kube-api-access-qwv27\") pod \"openstackclient\" (UID: \"ea063a3c-330a-4642-a815-c013ae6b80c7\") " pod="openstack/openstackclient" Nov 28 13:38:53 crc kubenswrapper[4631]: I1128 13:38:53.214187 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/ea063a3c-330a-4642-a815-c013ae6b80c7-openstack-config-secret\") pod \"openstackclient\" (UID: \"ea063a3c-330a-4642-a815-c013ae6b80c7\") " pod="openstack/openstackclient" Nov 28 13:38:53 crc kubenswrapper[4631]: E1128 13:38:53.249725 4631 log.go:32] "RunPodSandbox from runtime service failed" err=< Nov 28 13:38:53 crc kubenswrapper[4631]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_openstackclient_openstack_2eee8ee8-0983-43e2-85d6-7d0e274a71d5_0(605e0b2e5a3612073aa7846837c921e19f2d90002379a0953717c3b908fa7b77): error adding pod openstack_openstackclient to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"605e0b2e5a3612073aa7846837c921e19f2d90002379a0953717c3b908fa7b77" Netns:"/var/run/netns/8b54d3d6-5e7c-4b2c-901a-167809f38fb7" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openstack;K8S_POD_NAME=openstackclient;K8S_POD_INFRA_CONTAINER_ID=605e0b2e5a3612073aa7846837c921e19f2d90002379a0953717c3b908fa7b77;K8S_POD_UID=2eee8ee8-0983-43e2-85d6-7d0e274a71d5" Path:"" ERRORED: error configuring pod [openstack/openstackclient] networking: Multus: [openstack/openstackclient/2eee8ee8-0983-43e2-85d6-7d0e274a71d5]: expected pod UID "2eee8ee8-0983-43e2-85d6-7d0e274a71d5" but got "ea063a3c-330a-4642-a815-c013ae6b80c7" from Kube API Nov 28 13:38:53 crc kubenswrapper[4631]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Nov 28 13:38:53 crc kubenswrapper[4631]: > Nov 28 13:38:53 crc kubenswrapper[4631]: E1128 13:38:53.249827 4631 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err=< Nov 28 13:38:53 crc kubenswrapper[4631]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_openstackclient_openstack_2eee8ee8-0983-43e2-85d6-7d0e274a71d5_0(605e0b2e5a3612073aa7846837c921e19f2d90002379a0953717c3b908fa7b77): error adding pod openstack_openstackclient to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"605e0b2e5a3612073aa7846837c921e19f2d90002379a0953717c3b908fa7b77" Netns:"/var/run/netns/8b54d3d6-5e7c-4b2c-901a-167809f38fb7" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openstack;K8S_POD_NAME=openstackclient;K8S_POD_INFRA_CONTAINER_ID=605e0b2e5a3612073aa7846837c921e19f2d90002379a0953717c3b908fa7b77;K8S_POD_UID=2eee8ee8-0983-43e2-85d6-7d0e274a71d5" Path:"" ERRORED: error configuring pod [openstack/openstackclient] networking: Multus: [openstack/openstackclient/2eee8ee8-0983-43e2-85d6-7d0e274a71d5]: expected pod UID "2eee8ee8-0983-43e2-85d6-7d0e274a71d5" but got "ea063a3c-330a-4642-a815-c013ae6b80c7" from Kube API Nov 28 13:38:53 crc kubenswrapper[4631]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Nov 28 13:38:53 crc kubenswrapper[4631]: > pod="openstack/openstackclient" Nov 28 13:38:53 crc kubenswrapper[4631]: I1128 13:38:53.316322 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/ea063a3c-330a-4642-a815-c013ae6b80c7-openstack-config\") pod \"openstackclient\" (UID: \"ea063a3c-330a-4642-a815-c013ae6b80c7\") " pod="openstack/openstackclient" Nov 28 13:38:53 crc kubenswrapper[4631]: I1128 13:38:53.316444 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qwv27\" (UniqueName: \"kubernetes.io/projected/ea063a3c-330a-4642-a815-c013ae6b80c7-kube-api-access-qwv27\") pod \"openstackclient\" (UID: \"ea063a3c-330a-4642-a815-c013ae6b80c7\") " pod="openstack/openstackclient" Nov 28 13:38:53 crc kubenswrapper[4631]: I1128 13:38:53.316479 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/ea063a3c-330a-4642-a815-c013ae6b80c7-openstack-config-secret\") pod \"openstackclient\" (UID: \"ea063a3c-330a-4642-a815-c013ae6b80c7\") " pod="openstack/openstackclient" Nov 28 13:38:53 crc kubenswrapper[4631]: I1128 13:38:53.316555 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ea063a3c-330a-4642-a815-c013ae6b80c7-combined-ca-bundle\") pod \"openstackclient\" (UID: \"ea063a3c-330a-4642-a815-c013ae6b80c7\") " pod="openstack/openstackclient" Nov 28 13:38:53 crc kubenswrapper[4631]: I1128 13:38:53.317540 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/ea063a3c-330a-4642-a815-c013ae6b80c7-openstack-config\") pod \"openstackclient\" (UID: \"ea063a3c-330a-4642-a815-c013ae6b80c7\") " pod="openstack/openstackclient" Nov 28 13:38:53 crc kubenswrapper[4631]: I1128 13:38:53.321654 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/ea063a3c-330a-4642-a815-c013ae6b80c7-openstack-config-secret\") pod \"openstackclient\" (UID: \"ea063a3c-330a-4642-a815-c013ae6b80c7\") " pod="openstack/openstackclient" Nov 28 13:38:53 crc kubenswrapper[4631]: I1128 13:38:53.322250 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ea063a3c-330a-4642-a815-c013ae6b80c7-combined-ca-bundle\") pod \"openstackclient\" (UID: \"ea063a3c-330a-4642-a815-c013ae6b80c7\") " pod="openstack/openstackclient" Nov 28 13:38:53 crc kubenswrapper[4631]: I1128 13:38:53.336087 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qwv27\" (UniqueName: \"kubernetes.io/projected/ea063a3c-330a-4642-a815-c013ae6b80c7-kube-api-access-qwv27\") pod \"openstackclient\" (UID: \"ea063a3c-330a-4642-a815-c013ae6b80c7\") " pod="openstack/openstackclient" Nov 28 13:38:53 crc kubenswrapper[4631]: I1128 13:38:53.480225 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 28 13:38:53 crc kubenswrapper[4631]: I1128 13:38:53.770430 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 28 13:38:53 crc kubenswrapper[4631]: I1128 13:38:53.788697 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 28 13:38:53 crc kubenswrapper[4631]: I1128 13:38:53.794367 4631 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="2eee8ee8-0983-43e2-85d6-7d0e274a71d5" podUID="ea063a3c-330a-4642-a815-c013ae6b80c7" Nov 28 13:38:53 crc kubenswrapper[4631]: I1128 13:38:53.829343 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2eee8ee8-0983-43e2-85d6-7d0e274a71d5-combined-ca-bundle\") pod \"2eee8ee8-0983-43e2-85d6-7d0e274a71d5\" (UID: \"2eee8ee8-0983-43e2-85d6-7d0e274a71d5\") " Nov 28 13:38:53 crc kubenswrapper[4631]: I1128 13:38:53.829525 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/2eee8ee8-0983-43e2-85d6-7d0e274a71d5-openstack-config-secret\") pod \"2eee8ee8-0983-43e2-85d6-7d0e274a71d5\" (UID: \"2eee8ee8-0983-43e2-85d6-7d0e274a71d5\") " Nov 28 13:38:53 crc kubenswrapper[4631]: I1128 13:38:53.829603 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bszlc\" (UniqueName: \"kubernetes.io/projected/2eee8ee8-0983-43e2-85d6-7d0e274a71d5-kube-api-access-bszlc\") pod \"2eee8ee8-0983-43e2-85d6-7d0e274a71d5\" (UID: \"2eee8ee8-0983-43e2-85d6-7d0e274a71d5\") " Nov 28 13:38:53 crc kubenswrapper[4631]: I1128 13:38:53.829642 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/2eee8ee8-0983-43e2-85d6-7d0e274a71d5-openstack-config\") pod \"2eee8ee8-0983-43e2-85d6-7d0e274a71d5\" (UID: \"2eee8ee8-0983-43e2-85d6-7d0e274a71d5\") " Nov 28 13:38:53 crc kubenswrapper[4631]: I1128 13:38:53.830930 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2eee8ee8-0983-43e2-85d6-7d0e274a71d5-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "2eee8ee8-0983-43e2-85d6-7d0e274a71d5" (UID: "2eee8ee8-0983-43e2-85d6-7d0e274a71d5"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 13:38:53 crc kubenswrapper[4631]: I1128 13:38:53.837597 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2eee8ee8-0983-43e2-85d6-7d0e274a71d5-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "2eee8ee8-0983-43e2-85d6-7d0e274a71d5" (UID: "2eee8ee8-0983-43e2-85d6-7d0e274a71d5"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:38:53 crc kubenswrapper[4631]: I1128 13:38:53.838577 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2eee8ee8-0983-43e2-85d6-7d0e274a71d5-kube-api-access-bszlc" (OuterVolumeSpecName: "kube-api-access-bszlc") pod "2eee8ee8-0983-43e2-85d6-7d0e274a71d5" (UID: "2eee8ee8-0983-43e2-85d6-7d0e274a71d5"). InnerVolumeSpecName "kube-api-access-bszlc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:38:53 crc kubenswrapper[4631]: I1128 13:38:53.840550 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2eee8ee8-0983-43e2-85d6-7d0e274a71d5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2eee8ee8-0983-43e2-85d6-7d0e274a71d5" (UID: "2eee8ee8-0983-43e2-85d6-7d0e274a71d5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:38:53 crc kubenswrapper[4631]: I1128 13:38:53.932647 4631 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2eee8ee8-0983-43e2-85d6-7d0e274a71d5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 13:38:53 crc kubenswrapper[4631]: I1128 13:38:53.932685 4631 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/2eee8ee8-0983-43e2-85d6-7d0e274a71d5-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Nov 28 13:38:53 crc kubenswrapper[4631]: I1128 13:38:53.932702 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bszlc\" (UniqueName: \"kubernetes.io/projected/2eee8ee8-0983-43e2-85d6-7d0e274a71d5-kube-api-access-bszlc\") on node \"crc\" DevicePath \"\"" Nov 28 13:38:53 crc kubenswrapper[4631]: I1128 13:38:53.932717 4631 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/2eee8ee8-0983-43e2-85d6-7d0e274a71d5-openstack-config\") on node \"crc\" DevicePath \"\"" Nov 28 13:38:54 crc kubenswrapper[4631]: I1128 13:38:54.025648 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Nov 28 13:38:54 crc kubenswrapper[4631]: I1128 13:38:54.787273 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"ea063a3c-330a-4642-a815-c013ae6b80c7","Type":"ContainerStarted","Data":"b5a73317b67df79bfcdbe1e80a8cb3bce489bd647abcbfc9c333d607ac2869ac"} Nov 28 13:38:54 crc kubenswrapper[4631]: I1128 13:38:54.790117 4631 generic.go:334] "Generic (PLEG): container finished" podID="7eea7f3e-fc3c-4c60-998e-79c9a31ee084" containerID="c9851fdf444e8d0158c79114f4288e3ea2c1a53294a47a6cfe8d648dde4e2e5f" exitCode=0 Nov 28 13:38:54 crc kubenswrapper[4631]: I1128 13:38:54.790209 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 28 13:38:54 crc kubenswrapper[4631]: I1128 13:38:54.790225 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-756d64c4c8-jzvlv" event={"ID":"7eea7f3e-fc3c-4c60-998e-79c9a31ee084","Type":"ContainerDied","Data":"c9851fdf444e8d0158c79114f4288e3ea2c1a53294a47a6cfe8d648dde4e2e5f"} Nov 28 13:38:54 crc kubenswrapper[4631]: I1128 13:38:54.813052 4631 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="2eee8ee8-0983-43e2-85d6-7d0e274a71d5" podUID="ea063a3c-330a-4642-a815-c013ae6b80c7" Nov 28 13:38:54 crc kubenswrapper[4631]: I1128 13:38:54.972731 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-756d64c4c8-jzvlv" Nov 28 13:38:55 crc kubenswrapper[4631]: I1128 13:38:55.063888 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7eea7f3e-fc3c-4c60-998e-79c9a31ee084-config-data\") pod \"7eea7f3e-fc3c-4c60-998e-79c9a31ee084\" (UID: \"7eea7f3e-fc3c-4c60-998e-79c9a31ee084\") " Nov 28 13:38:55 crc kubenswrapper[4631]: I1128 13:38:55.063955 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7eea7f3e-fc3c-4c60-998e-79c9a31ee084-logs\") pod \"7eea7f3e-fc3c-4c60-998e-79c9a31ee084\" (UID: \"7eea7f3e-fc3c-4c60-998e-79c9a31ee084\") " Nov 28 13:38:55 crc kubenswrapper[4631]: I1128 13:38:55.064001 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7eea7f3e-fc3c-4c60-998e-79c9a31ee084-config-data-custom\") pod \"7eea7f3e-fc3c-4c60-998e-79c9a31ee084\" (UID: \"7eea7f3e-fc3c-4c60-998e-79c9a31ee084\") " Nov 28 13:38:55 crc kubenswrapper[4631]: I1128 13:38:55.064327 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7eea7f3e-fc3c-4c60-998e-79c9a31ee084-combined-ca-bundle\") pod \"7eea7f3e-fc3c-4c60-998e-79c9a31ee084\" (UID: \"7eea7f3e-fc3c-4c60-998e-79c9a31ee084\") " Nov 28 13:38:55 crc kubenswrapper[4631]: I1128 13:38:55.064538 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xsl6r\" (UniqueName: \"kubernetes.io/projected/7eea7f3e-fc3c-4c60-998e-79c9a31ee084-kube-api-access-xsl6r\") pod \"7eea7f3e-fc3c-4c60-998e-79c9a31ee084\" (UID: \"7eea7f3e-fc3c-4c60-998e-79c9a31ee084\") " Nov 28 13:38:55 crc kubenswrapper[4631]: I1128 13:38:55.064810 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7eea7f3e-fc3c-4c60-998e-79c9a31ee084-logs" (OuterVolumeSpecName: "logs") pod "7eea7f3e-fc3c-4c60-998e-79c9a31ee084" (UID: "7eea7f3e-fc3c-4c60-998e-79c9a31ee084"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 13:38:55 crc kubenswrapper[4631]: I1128 13:38:55.065801 4631 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7eea7f3e-fc3c-4c60-998e-79c9a31ee084-logs\") on node \"crc\" DevicePath \"\"" Nov 28 13:38:55 crc kubenswrapper[4631]: I1128 13:38:55.071154 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7eea7f3e-fc3c-4c60-998e-79c9a31ee084-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "7eea7f3e-fc3c-4c60-998e-79c9a31ee084" (UID: "7eea7f3e-fc3c-4c60-998e-79c9a31ee084"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:38:55 crc kubenswrapper[4631]: I1128 13:38:55.089691 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7eea7f3e-fc3c-4c60-998e-79c9a31ee084-kube-api-access-xsl6r" (OuterVolumeSpecName: "kube-api-access-xsl6r") pod "7eea7f3e-fc3c-4c60-998e-79c9a31ee084" (UID: "7eea7f3e-fc3c-4c60-998e-79c9a31ee084"). InnerVolumeSpecName "kube-api-access-xsl6r". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:38:55 crc kubenswrapper[4631]: I1128 13:38:55.114225 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7eea7f3e-fc3c-4c60-998e-79c9a31ee084-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7eea7f3e-fc3c-4c60-998e-79c9a31ee084" (UID: "7eea7f3e-fc3c-4c60-998e-79c9a31ee084"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:38:55 crc kubenswrapper[4631]: I1128 13:38:55.139184 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7eea7f3e-fc3c-4c60-998e-79c9a31ee084-config-data" (OuterVolumeSpecName: "config-data") pod "7eea7f3e-fc3c-4c60-998e-79c9a31ee084" (UID: "7eea7f3e-fc3c-4c60-998e-79c9a31ee084"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:38:55 crc kubenswrapper[4631]: I1128 13:38:55.167785 4631 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7eea7f3e-fc3c-4c60-998e-79c9a31ee084-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 13:38:55 crc kubenswrapper[4631]: I1128 13:38:55.168011 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xsl6r\" (UniqueName: \"kubernetes.io/projected/7eea7f3e-fc3c-4c60-998e-79c9a31ee084-kube-api-access-xsl6r\") on node \"crc\" DevicePath \"\"" Nov 28 13:38:55 crc kubenswrapper[4631]: I1128 13:38:55.168132 4631 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7eea7f3e-fc3c-4c60-998e-79c9a31ee084-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 13:38:55 crc kubenswrapper[4631]: I1128 13:38:55.168216 4631 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7eea7f3e-fc3c-4c60-998e-79c9a31ee084-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 28 13:38:55 crc kubenswrapper[4631]: I1128 13:38:55.532334 4631 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2eee8ee8-0983-43e2-85d6-7d0e274a71d5" path="/var/lib/kubelet/pods/2eee8ee8-0983-43e2-85d6-7d0e274a71d5/volumes" Nov 28 13:38:55 crc kubenswrapper[4631]: I1128 13:38:55.811768 4631 generic.go:334] "Generic (PLEG): container finished" podID="cd607dae-568a-4b81-af81-3310c0e95854" containerID="bd4617d23aba6bbc47167ded3acd30280726cfe208e8bfc920d75fb684ec48e8" exitCode=137 Nov 28 13:38:55 crc kubenswrapper[4631]: I1128 13:38:55.811951 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-544496c768-rzhsw" event={"ID":"cd607dae-568a-4b81-af81-3310c0e95854","Type":"ContainerDied","Data":"bd4617d23aba6bbc47167ded3acd30280726cfe208e8bfc920d75fb684ec48e8"} Nov 28 13:38:55 crc kubenswrapper[4631]: I1128 13:38:55.815028 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-756d64c4c8-jzvlv" event={"ID":"7eea7f3e-fc3c-4c60-998e-79c9a31ee084","Type":"ContainerDied","Data":"4122a24fb417b9c8665b47ec5c5eda95ae81959c07dcb5d83737b2f293b86f13"} Nov 28 13:38:55 crc kubenswrapper[4631]: I1128 13:38:55.815086 4631 scope.go:117] "RemoveContainer" containerID="c9851fdf444e8d0158c79114f4288e3ea2c1a53294a47a6cfe8d648dde4e2e5f" Nov 28 13:38:55 crc kubenswrapper[4631]: I1128 13:38:55.815238 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-756d64c4c8-jzvlv" Nov 28 13:38:55 crc kubenswrapper[4631]: I1128 13:38:55.828996 4631 generic.go:334] "Generic (PLEG): container finished" podID="111c5349-528d-4856-a3e4-1277e3b10889" containerID="ffafadc828320f83436d72910a3912cda7f6bb048d8c4c01528f4529bf55b772" exitCode=137 Nov 28 13:38:55 crc kubenswrapper[4631]: I1128 13:38:55.829043 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-64f8fdf97b-jn5mg" event={"ID":"111c5349-528d-4856-a3e4-1277e3b10889","Type":"ContainerDied","Data":"ffafadc828320f83436d72910a3912cda7f6bb048d8c4c01528f4529bf55b772"} Nov 28 13:38:55 crc kubenswrapper[4631]: I1128 13:38:55.829077 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-64f8fdf97b-jn5mg" event={"ID":"111c5349-528d-4856-a3e4-1277e3b10889","Type":"ContainerStarted","Data":"ddca46c54a9ba33270a5c14213a03ffc4ee751ec03afaccc3b37cfa0eac87eb7"} Nov 28 13:38:55 crc kubenswrapper[4631]: I1128 13:38:55.898018 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-756d64c4c8-jzvlv"] Nov 28 13:38:55 crc kubenswrapper[4631]: I1128 13:38:55.898124 4631 scope.go:117] "RemoveContainer" containerID="39f6239d34a7e4114e98f79691b15de6581a312f79450b6aa821ddc52f868ed0" Nov 28 13:38:55 crc kubenswrapper[4631]: I1128 13:38:55.934923 4631 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-756d64c4c8-jzvlv"] Nov 28 13:38:56 crc kubenswrapper[4631]: I1128 13:38:56.140148 4631 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Nov 28 13:38:56 crc kubenswrapper[4631]: I1128 13:38:56.841830 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-544496c768-rzhsw" event={"ID":"cd607dae-568a-4b81-af81-3310c0e95854","Type":"ContainerStarted","Data":"26b4a5d749292bcec07819e5de34fe07da049e5be75efa94d19e13fdfcfb064d"} Nov 28 13:38:57 crc kubenswrapper[4631]: I1128 13:38:57.540108 4631 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7eea7f3e-fc3c-4c60-998e-79c9a31ee084" path="/var/lib/kubelet/pods/7eea7f3e-fc3c-4c60-998e-79c9a31ee084/volumes" Nov 28 13:38:59 crc kubenswrapper[4631]: I1128 13:38:59.487455 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 28 13:38:59 crc kubenswrapper[4631]: I1128 13:38:59.584439 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d95d3536-1392-4576-8ee5-390c09fe7a49-scripts\") pod \"d95d3536-1392-4576-8ee5-390c09fe7a49\" (UID: \"d95d3536-1392-4576-8ee5-390c09fe7a49\") " Nov 28 13:38:59 crc kubenswrapper[4631]: I1128 13:38:59.584603 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d95d3536-1392-4576-8ee5-390c09fe7a49-combined-ca-bundle\") pod \"d95d3536-1392-4576-8ee5-390c09fe7a49\" (UID: \"d95d3536-1392-4576-8ee5-390c09fe7a49\") " Nov 28 13:38:59 crc kubenswrapper[4631]: I1128 13:38:59.584641 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d95d3536-1392-4576-8ee5-390c09fe7a49-config-data\") pod \"d95d3536-1392-4576-8ee5-390c09fe7a49\" (UID: \"d95d3536-1392-4576-8ee5-390c09fe7a49\") " Nov 28 13:38:59 crc kubenswrapper[4631]: I1128 13:38:59.584671 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d95d3536-1392-4576-8ee5-390c09fe7a49-run-httpd\") pod \"d95d3536-1392-4576-8ee5-390c09fe7a49\" (UID: \"d95d3536-1392-4576-8ee5-390c09fe7a49\") " Nov 28 13:38:59 crc kubenswrapper[4631]: I1128 13:38:59.584738 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d95d3536-1392-4576-8ee5-390c09fe7a49-sg-core-conf-yaml\") pod \"d95d3536-1392-4576-8ee5-390c09fe7a49\" (UID: \"d95d3536-1392-4576-8ee5-390c09fe7a49\") " Nov 28 13:38:59 crc kubenswrapper[4631]: I1128 13:38:59.584887 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fd5sn\" (UniqueName: \"kubernetes.io/projected/d95d3536-1392-4576-8ee5-390c09fe7a49-kube-api-access-fd5sn\") pod \"d95d3536-1392-4576-8ee5-390c09fe7a49\" (UID: \"d95d3536-1392-4576-8ee5-390c09fe7a49\") " Nov 28 13:38:59 crc kubenswrapper[4631]: I1128 13:38:59.584918 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d95d3536-1392-4576-8ee5-390c09fe7a49-log-httpd\") pod \"d95d3536-1392-4576-8ee5-390c09fe7a49\" (UID: \"d95d3536-1392-4576-8ee5-390c09fe7a49\") " Nov 28 13:38:59 crc kubenswrapper[4631]: I1128 13:38:59.587769 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d95d3536-1392-4576-8ee5-390c09fe7a49-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "d95d3536-1392-4576-8ee5-390c09fe7a49" (UID: "d95d3536-1392-4576-8ee5-390c09fe7a49"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 13:38:59 crc kubenswrapper[4631]: I1128 13:38:59.587883 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d95d3536-1392-4576-8ee5-390c09fe7a49-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "d95d3536-1392-4576-8ee5-390c09fe7a49" (UID: "d95d3536-1392-4576-8ee5-390c09fe7a49"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 13:38:59 crc kubenswrapper[4631]: I1128 13:38:59.591494 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d95d3536-1392-4576-8ee5-390c09fe7a49-scripts" (OuterVolumeSpecName: "scripts") pod "d95d3536-1392-4576-8ee5-390c09fe7a49" (UID: "d95d3536-1392-4576-8ee5-390c09fe7a49"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:38:59 crc kubenswrapper[4631]: I1128 13:38:59.593828 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d95d3536-1392-4576-8ee5-390c09fe7a49-kube-api-access-fd5sn" (OuterVolumeSpecName: "kube-api-access-fd5sn") pod "d95d3536-1392-4576-8ee5-390c09fe7a49" (UID: "d95d3536-1392-4576-8ee5-390c09fe7a49"). InnerVolumeSpecName "kube-api-access-fd5sn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:38:59 crc kubenswrapper[4631]: I1128 13:38:59.645899 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d95d3536-1392-4576-8ee5-390c09fe7a49-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "d95d3536-1392-4576-8ee5-390c09fe7a49" (UID: "d95d3536-1392-4576-8ee5-390c09fe7a49"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:38:59 crc kubenswrapper[4631]: I1128 13:38:59.689968 4631 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d95d3536-1392-4576-8ee5-390c09fe7a49-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 13:38:59 crc kubenswrapper[4631]: I1128 13:38:59.690034 4631 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d95d3536-1392-4576-8ee5-390c09fe7a49-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 28 13:38:59 crc kubenswrapper[4631]: I1128 13:38:59.690044 4631 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d95d3536-1392-4576-8ee5-390c09fe7a49-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 28 13:38:59 crc kubenswrapper[4631]: I1128 13:38:59.690056 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fd5sn\" (UniqueName: \"kubernetes.io/projected/d95d3536-1392-4576-8ee5-390c09fe7a49-kube-api-access-fd5sn\") on node \"crc\" DevicePath \"\"" Nov 28 13:38:59 crc kubenswrapper[4631]: I1128 13:38:59.690067 4631 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d95d3536-1392-4576-8ee5-390c09fe7a49-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 28 13:38:59 crc kubenswrapper[4631]: I1128 13:38:59.756454 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d95d3536-1392-4576-8ee5-390c09fe7a49-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d95d3536-1392-4576-8ee5-390c09fe7a49" (UID: "d95d3536-1392-4576-8ee5-390c09fe7a49"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:38:59 crc kubenswrapper[4631]: I1128 13:38:59.780602 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d95d3536-1392-4576-8ee5-390c09fe7a49-config-data" (OuterVolumeSpecName: "config-data") pod "d95d3536-1392-4576-8ee5-390c09fe7a49" (UID: "d95d3536-1392-4576-8ee5-390c09fe7a49"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:38:59 crc kubenswrapper[4631]: I1128 13:38:59.792794 4631 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d95d3536-1392-4576-8ee5-390c09fe7a49-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 13:38:59 crc kubenswrapper[4631]: I1128 13:38:59.792844 4631 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d95d3536-1392-4576-8ee5-390c09fe7a49-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 13:38:59 crc kubenswrapper[4631]: I1128 13:38:59.887875 4631 generic.go:334] "Generic (PLEG): container finished" podID="d95d3536-1392-4576-8ee5-390c09fe7a49" containerID="756cbe01466c40de3773fa6a449d3e360e15094ae83297664fbe80c0bd2d91ba" exitCode=137 Nov 28 13:38:59 crc kubenswrapper[4631]: I1128 13:38:59.887949 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d95d3536-1392-4576-8ee5-390c09fe7a49","Type":"ContainerDied","Data":"756cbe01466c40de3773fa6a449d3e360e15094ae83297664fbe80c0bd2d91ba"} Nov 28 13:38:59 crc kubenswrapper[4631]: I1128 13:38:59.887974 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 28 13:38:59 crc kubenswrapper[4631]: I1128 13:38:59.888002 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d95d3536-1392-4576-8ee5-390c09fe7a49","Type":"ContainerDied","Data":"bfec386172ed3b954e0624e2432af7b8e50aea6ad51d561d1ba0a2aad259aecd"} Nov 28 13:38:59 crc kubenswrapper[4631]: I1128 13:38:59.888031 4631 scope.go:117] "RemoveContainer" containerID="756cbe01466c40de3773fa6a449d3e360e15094ae83297664fbe80c0bd2d91ba" Nov 28 13:38:59 crc kubenswrapper[4631]: I1128 13:38:59.926961 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 28 13:38:59 crc kubenswrapper[4631]: I1128 13:38:59.932496 4631 scope.go:117] "RemoveContainer" containerID="aed54e68e791750f2a7b381406c45fd57ec0e772fa0e75496b255dcade224d18" Nov 28 13:38:59 crc kubenswrapper[4631]: I1128 13:38:59.943469 4631 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 28 13:38:59 crc kubenswrapper[4631]: I1128 13:38:59.961972 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 28 13:38:59 crc kubenswrapper[4631]: E1128 13:38:59.962586 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d95d3536-1392-4576-8ee5-390c09fe7a49" containerName="ceilometer-notification-agent" Nov 28 13:38:59 crc kubenswrapper[4631]: I1128 13:38:59.962603 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="d95d3536-1392-4576-8ee5-390c09fe7a49" containerName="ceilometer-notification-agent" Nov 28 13:38:59 crc kubenswrapper[4631]: E1128 13:38:59.962622 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d95d3536-1392-4576-8ee5-390c09fe7a49" containerName="ceilometer-central-agent" Nov 28 13:38:59 crc kubenswrapper[4631]: I1128 13:38:59.962629 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="d95d3536-1392-4576-8ee5-390c09fe7a49" containerName="ceilometer-central-agent" Nov 28 13:38:59 crc kubenswrapper[4631]: E1128 13:38:59.962661 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d95d3536-1392-4576-8ee5-390c09fe7a49" containerName="proxy-httpd" Nov 28 13:38:59 crc kubenswrapper[4631]: I1128 13:38:59.962667 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="d95d3536-1392-4576-8ee5-390c09fe7a49" containerName="proxy-httpd" Nov 28 13:38:59 crc kubenswrapper[4631]: E1128 13:38:59.962680 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d95d3536-1392-4576-8ee5-390c09fe7a49" containerName="sg-core" Nov 28 13:38:59 crc kubenswrapper[4631]: I1128 13:38:59.962686 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="d95d3536-1392-4576-8ee5-390c09fe7a49" containerName="sg-core" Nov 28 13:38:59 crc kubenswrapper[4631]: E1128 13:38:59.962693 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7eea7f3e-fc3c-4c60-998e-79c9a31ee084" containerName="barbican-api-log" Nov 28 13:38:59 crc kubenswrapper[4631]: I1128 13:38:59.962699 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="7eea7f3e-fc3c-4c60-998e-79c9a31ee084" containerName="barbican-api-log" Nov 28 13:38:59 crc kubenswrapper[4631]: E1128 13:38:59.962708 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7eea7f3e-fc3c-4c60-998e-79c9a31ee084" containerName="barbican-api" Nov 28 13:38:59 crc kubenswrapper[4631]: I1128 13:38:59.962714 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="7eea7f3e-fc3c-4c60-998e-79c9a31ee084" containerName="barbican-api" Nov 28 13:38:59 crc kubenswrapper[4631]: I1128 13:38:59.962885 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="7eea7f3e-fc3c-4c60-998e-79c9a31ee084" containerName="barbican-api-log" Nov 28 13:38:59 crc kubenswrapper[4631]: I1128 13:38:59.962901 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="7eea7f3e-fc3c-4c60-998e-79c9a31ee084" containerName="barbican-api" Nov 28 13:38:59 crc kubenswrapper[4631]: I1128 13:38:59.962910 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="d95d3536-1392-4576-8ee5-390c09fe7a49" containerName="proxy-httpd" Nov 28 13:38:59 crc kubenswrapper[4631]: I1128 13:38:59.962919 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="d95d3536-1392-4576-8ee5-390c09fe7a49" containerName="sg-core" Nov 28 13:38:59 crc kubenswrapper[4631]: I1128 13:38:59.962932 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="d95d3536-1392-4576-8ee5-390c09fe7a49" containerName="ceilometer-notification-agent" Nov 28 13:38:59 crc kubenswrapper[4631]: I1128 13:38:59.962940 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="d95d3536-1392-4576-8ee5-390c09fe7a49" containerName="ceilometer-central-agent" Nov 28 13:38:59 crc kubenswrapper[4631]: I1128 13:38:59.966398 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 28 13:38:59 crc kubenswrapper[4631]: I1128 13:38:59.969795 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 28 13:38:59 crc kubenswrapper[4631]: I1128 13:38:59.969998 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 28 13:38:59 crc kubenswrapper[4631]: I1128 13:38:59.996749 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-npcz5\" (UniqueName: \"kubernetes.io/projected/ac113fd1-daaf-4958-b779-aaee117db1ea-kube-api-access-npcz5\") pod \"ceilometer-0\" (UID: \"ac113fd1-daaf-4958-b779-aaee117db1ea\") " pod="openstack/ceilometer-0" Nov 28 13:38:59 crc kubenswrapper[4631]: I1128 13:38:59.996793 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ac113fd1-daaf-4958-b779-aaee117db1ea-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"ac113fd1-daaf-4958-b779-aaee117db1ea\") " pod="openstack/ceilometer-0" Nov 28 13:38:59 crc kubenswrapper[4631]: I1128 13:38:59.996820 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ac113fd1-daaf-4958-b779-aaee117db1ea-run-httpd\") pod \"ceilometer-0\" (UID: \"ac113fd1-daaf-4958-b779-aaee117db1ea\") " pod="openstack/ceilometer-0" Nov 28 13:38:59 crc kubenswrapper[4631]: I1128 13:38:59.996870 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ac113fd1-daaf-4958-b779-aaee117db1ea-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"ac113fd1-daaf-4958-b779-aaee117db1ea\") " pod="openstack/ceilometer-0" Nov 28 13:38:59 crc kubenswrapper[4631]: I1128 13:38:59.996894 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ac113fd1-daaf-4958-b779-aaee117db1ea-scripts\") pod \"ceilometer-0\" (UID: \"ac113fd1-daaf-4958-b779-aaee117db1ea\") " pod="openstack/ceilometer-0" Nov 28 13:38:59 crc kubenswrapper[4631]: I1128 13:38:59.996961 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ac113fd1-daaf-4958-b779-aaee117db1ea-log-httpd\") pod \"ceilometer-0\" (UID: \"ac113fd1-daaf-4958-b779-aaee117db1ea\") " pod="openstack/ceilometer-0" Nov 28 13:38:59 crc kubenswrapper[4631]: I1128 13:38:59.996982 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ac113fd1-daaf-4958-b779-aaee117db1ea-config-data\") pod \"ceilometer-0\" (UID: \"ac113fd1-daaf-4958-b779-aaee117db1ea\") " pod="openstack/ceilometer-0" Nov 28 13:38:59 crc kubenswrapper[4631]: I1128 13:38:59.997167 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 28 13:39:00 crc kubenswrapper[4631]: I1128 13:39:00.055260 4631 scope.go:117] "RemoveContainer" containerID="2141eebe891186cff2651a86531e3f94d1b78cbeeead6dde52101d7a46457ff5" Nov 28 13:39:00 crc kubenswrapper[4631]: I1128 13:39:00.092058 4631 scope.go:117] "RemoveContainer" containerID="42664776e24655262092eae45a6b6b94e6ed84e5b6614b7c0abe60d7e38df756" Nov 28 13:39:00 crc kubenswrapper[4631]: I1128 13:39:00.098787 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-npcz5\" (UniqueName: \"kubernetes.io/projected/ac113fd1-daaf-4958-b779-aaee117db1ea-kube-api-access-npcz5\") pod \"ceilometer-0\" (UID: \"ac113fd1-daaf-4958-b779-aaee117db1ea\") " pod="openstack/ceilometer-0" Nov 28 13:39:00 crc kubenswrapper[4631]: I1128 13:39:00.099155 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ac113fd1-daaf-4958-b779-aaee117db1ea-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"ac113fd1-daaf-4958-b779-aaee117db1ea\") " pod="openstack/ceilometer-0" Nov 28 13:39:00 crc kubenswrapper[4631]: I1128 13:39:00.099253 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ac113fd1-daaf-4958-b779-aaee117db1ea-run-httpd\") pod \"ceilometer-0\" (UID: \"ac113fd1-daaf-4958-b779-aaee117db1ea\") " pod="openstack/ceilometer-0" Nov 28 13:39:00 crc kubenswrapper[4631]: I1128 13:39:00.099480 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ac113fd1-daaf-4958-b779-aaee117db1ea-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"ac113fd1-daaf-4958-b779-aaee117db1ea\") " pod="openstack/ceilometer-0" Nov 28 13:39:00 crc kubenswrapper[4631]: I1128 13:39:00.099581 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ac113fd1-daaf-4958-b779-aaee117db1ea-scripts\") pod \"ceilometer-0\" (UID: \"ac113fd1-daaf-4958-b779-aaee117db1ea\") " pod="openstack/ceilometer-0" Nov 28 13:39:00 crc kubenswrapper[4631]: I1128 13:39:00.100226 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ac113fd1-daaf-4958-b779-aaee117db1ea-run-httpd\") pod \"ceilometer-0\" (UID: \"ac113fd1-daaf-4958-b779-aaee117db1ea\") " pod="openstack/ceilometer-0" Nov 28 13:39:00 crc kubenswrapper[4631]: I1128 13:39:00.099715 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ac113fd1-daaf-4958-b779-aaee117db1ea-log-httpd\") pod \"ceilometer-0\" (UID: \"ac113fd1-daaf-4958-b779-aaee117db1ea\") " pod="openstack/ceilometer-0" Nov 28 13:39:00 crc kubenswrapper[4631]: I1128 13:39:00.100374 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ac113fd1-daaf-4958-b779-aaee117db1ea-config-data\") pod \"ceilometer-0\" (UID: \"ac113fd1-daaf-4958-b779-aaee117db1ea\") " pod="openstack/ceilometer-0" Nov 28 13:39:00 crc kubenswrapper[4631]: I1128 13:39:00.100487 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ac113fd1-daaf-4958-b779-aaee117db1ea-log-httpd\") pod \"ceilometer-0\" (UID: \"ac113fd1-daaf-4958-b779-aaee117db1ea\") " pod="openstack/ceilometer-0" Nov 28 13:39:00 crc kubenswrapper[4631]: I1128 13:39:00.108656 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ac113fd1-daaf-4958-b779-aaee117db1ea-scripts\") pod \"ceilometer-0\" (UID: \"ac113fd1-daaf-4958-b779-aaee117db1ea\") " pod="openstack/ceilometer-0" Nov 28 13:39:00 crc kubenswrapper[4631]: I1128 13:39:00.114025 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ac113fd1-daaf-4958-b779-aaee117db1ea-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"ac113fd1-daaf-4958-b779-aaee117db1ea\") " pod="openstack/ceilometer-0" Nov 28 13:39:00 crc kubenswrapper[4631]: I1128 13:39:00.121219 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ac113fd1-daaf-4958-b779-aaee117db1ea-config-data\") pod \"ceilometer-0\" (UID: \"ac113fd1-daaf-4958-b779-aaee117db1ea\") " pod="openstack/ceilometer-0" Nov 28 13:39:00 crc kubenswrapper[4631]: I1128 13:39:00.124238 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-npcz5\" (UniqueName: \"kubernetes.io/projected/ac113fd1-daaf-4958-b779-aaee117db1ea-kube-api-access-npcz5\") pod \"ceilometer-0\" (UID: \"ac113fd1-daaf-4958-b779-aaee117db1ea\") " pod="openstack/ceilometer-0" Nov 28 13:39:00 crc kubenswrapper[4631]: I1128 13:39:00.128053 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ac113fd1-daaf-4958-b779-aaee117db1ea-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"ac113fd1-daaf-4958-b779-aaee117db1ea\") " pod="openstack/ceilometer-0" Nov 28 13:39:00 crc kubenswrapper[4631]: I1128 13:39:00.256707 4631 scope.go:117] "RemoveContainer" containerID="756cbe01466c40de3773fa6a449d3e360e15094ae83297664fbe80c0bd2d91ba" Nov 28 13:39:00 crc kubenswrapper[4631]: E1128 13:39:00.258161 4631 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"756cbe01466c40de3773fa6a449d3e360e15094ae83297664fbe80c0bd2d91ba\": container with ID starting with 756cbe01466c40de3773fa6a449d3e360e15094ae83297664fbe80c0bd2d91ba not found: ID does not exist" containerID="756cbe01466c40de3773fa6a449d3e360e15094ae83297664fbe80c0bd2d91ba" Nov 28 13:39:00 crc kubenswrapper[4631]: I1128 13:39:00.258204 4631 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"756cbe01466c40de3773fa6a449d3e360e15094ae83297664fbe80c0bd2d91ba"} err="failed to get container status \"756cbe01466c40de3773fa6a449d3e360e15094ae83297664fbe80c0bd2d91ba\": rpc error: code = NotFound desc = could not find container \"756cbe01466c40de3773fa6a449d3e360e15094ae83297664fbe80c0bd2d91ba\": container with ID starting with 756cbe01466c40de3773fa6a449d3e360e15094ae83297664fbe80c0bd2d91ba not found: ID does not exist" Nov 28 13:39:00 crc kubenswrapper[4631]: I1128 13:39:00.258241 4631 scope.go:117] "RemoveContainer" containerID="aed54e68e791750f2a7b381406c45fd57ec0e772fa0e75496b255dcade224d18" Nov 28 13:39:00 crc kubenswrapper[4631]: E1128 13:39:00.259872 4631 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aed54e68e791750f2a7b381406c45fd57ec0e772fa0e75496b255dcade224d18\": container with ID starting with aed54e68e791750f2a7b381406c45fd57ec0e772fa0e75496b255dcade224d18 not found: ID does not exist" containerID="aed54e68e791750f2a7b381406c45fd57ec0e772fa0e75496b255dcade224d18" Nov 28 13:39:00 crc kubenswrapper[4631]: I1128 13:39:00.259905 4631 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aed54e68e791750f2a7b381406c45fd57ec0e772fa0e75496b255dcade224d18"} err="failed to get container status \"aed54e68e791750f2a7b381406c45fd57ec0e772fa0e75496b255dcade224d18\": rpc error: code = NotFound desc = could not find container \"aed54e68e791750f2a7b381406c45fd57ec0e772fa0e75496b255dcade224d18\": container with ID starting with aed54e68e791750f2a7b381406c45fd57ec0e772fa0e75496b255dcade224d18 not found: ID does not exist" Nov 28 13:39:00 crc kubenswrapper[4631]: I1128 13:39:00.259926 4631 scope.go:117] "RemoveContainer" containerID="2141eebe891186cff2651a86531e3f94d1b78cbeeead6dde52101d7a46457ff5" Nov 28 13:39:00 crc kubenswrapper[4631]: E1128 13:39:00.260511 4631 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2141eebe891186cff2651a86531e3f94d1b78cbeeead6dde52101d7a46457ff5\": container with ID starting with 2141eebe891186cff2651a86531e3f94d1b78cbeeead6dde52101d7a46457ff5 not found: ID does not exist" containerID="2141eebe891186cff2651a86531e3f94d1b78cbeeead6dde52101d7a46457ff5" Nov 28 13:39:00 crc kubenswrapper[4631]: I1128 13:39:00.260534 4631 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2141eebe891186cff2651a86531e3f94d1b78cbeeead6dde52101d7a46457ff5"} err="failed to get container status \"2141eebe891186cff2651a86531e3f94d1b78cbeeead6dde52101d7a46457ff5\": rpc error: code = NotFound desc = could not find container \"2141eebe891186cff2651a86531e3f94d1b78cbeeead6dde52101d7a46457ff5\": container with ID starting with 2141eebe891186cff2651a86531e3f94d1b78cbeeead6dde52101d7a46457ff5 not found: ID does not exist" Nov 28 13:39:00 crc kubenswrapper[4631]: I1128 13:39:00.260589 4631 scope.go:117] "RemoveContainer" containerID="42664776e24655262092eae45a6b6b94e6ed84e5b6614b7c0abe60d7e38df756" Nov 28 13:39:00 crc kubenswrapper[4631]: E1128 13:39:00.261215 4631 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"42664776e24655262092eae45a6b6b94e6ed84e5b6614b7c0abe60d7e38df756\": container with ID starting with 42664776e24655262092eae45a6b6b94e6ed84e5b6614b7c0abe60d7e38df756 not found: ID does not exist" containerID="42664776e24655262092eae45a6b6b94e6ed84e5b6614b7c0abe60d7e38df756" Nov 28 13:39:00 crc kubenswrapper[4631]: I1128 13:39:00.261238 4631 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"42664776e24655262092eae45a6b6b94e6ed84e5b6614b7c0abe60d7e38df756"} err="failed to get container status \"42664776e24655262092eae45a6b6b94e6ed84e5b6614b7c0abe60d7e38df756\": rpc error: code = NotFound desc = could not find container \"42664776e24655262092eae45a6b6b94e6ed84e5b6614b7c0abe60d7e38df756\": container with ID starting with 42664776e24655262092eae45a6b6b94e6ed84e5b6614b7c0abe60d7e38df756 not found: ID does not exist" Nov 28 13:39:00 crc kubenswrapper[4631]: I1128 13:39:00.304579 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 28 13:39:00 crc kubenswrapper[4631]: I1128 13:39:00.305154 4631 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="cb966e50-d246-48dc-b4a6-a9de4dbb8569" containerName="glance-log" containerID="cri-o://6a153b478ff34b35b8edd10fed13a31e0f77eee95649346d1c104a57495f9a84" gracePeriod=30 Nov 28 13:39:00 crc kubenswrapper[4631]: I1128 13:39:00.306018 4631 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="cb966e50-d246-48dc-b4a6-a9de4dbb8569" containerName="glance-httpd" containerID="cri-o://a869370244f2d82eb690074a62782ef88fa1793d2d6fd2b724555af086cd397f" gracePeriod=30 Nov 28 13:39:00 crc kubenswrapper[4631]: I1128 13:39:00.341113 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 28 13:39:00 crc kubenswrapper[4631]: I1128 13:39:00.680890 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 28 13:39:00 crc kubenswrapper[4631]: I1128 13:39:00.910863 4631 generic.go:334] "Generic (PLEG): container finished" podID="cb966e50-d246-48dc-b4a6-a9de4dbb8569" containerID="6a153b478ff34b35b8edd10fed13a31e0f77eee95649346d1c104a57495f9a84" exitCode=143 Nov 28 13:39:00 crc kubenswrapper[4631]: I1128 13:39:00.910947 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"cb966e50-d246-48dc-b4a6-a9de4dbb8569","Type":"ContainerDied","Data":"6a153b478ff34b35b8edd10fed13a31e0f77eee95649346d1c104a57495f9a84"} Nov 28 13:39:00 crc kubenswrapper[4631]: W1128 13:39:00.970451 4631 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podac113fd1_daaf_4958_b779_aaee117db1ea.slice/crio-00ecb9ef637423c12a9a4d2165068b41e5935cba5464a65193c60c523a342172 WatchSource:0}: Error finding container 00ecb9ef637423c12a9a4d2165068b41e5935cba5464a65193c60c523a342172: Status 404 returned error can't find the container with id 00ecb9ef637423c12a9a4d2165068b41e5935cba5464a65193c60c523a342172 Nov 28 13:39:00 crc kubenswrapper[4631]: I1128 13:39:00.977749 4631 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 28 13:39:01 crc kubenswrapper[4631]: I1128 13:39:01.013027 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 28 13:39:01 crc kubenswrapper[4631]: I1128 13:39:01.037514 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-proxy-68757c9bcf-d8gj4"] Nov 28 13:39:01 crc kubenswrapper[4631]: I1128 13:39:01.041183 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-68757c9bcf-d8gj4" Nov 28 13:39:01 crc kubenswrapper[4631]: I1128 13:39:01.051671 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-public-svc" Nov 28 13:39:01 crc kubenswrapper[4631]: I1128 13:39:01.051671 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Nov 28 13:39:01 crc kubenswrapper[4631]: I1128 13:39:01.051805 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-internal-svc" Nov 28 13:39:01 crc kubenswrapper[4631]: I1128 13:39:01.070960 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-68757c9bcf-d8gj4"] Nov 28 13:39:01 crc kubenswrapper[4631]: I1128 13:39:01.133126 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7303e352-744c-4c93-ac72-0c434384f090-log-httpd\") pod \"swift-proxy-68757c9bcf-d8gj4\" (UID: \"7303e352-744c-4c93-ac72-0c434384f090\") " pod="openstack/swift-proxy-68757c9bcf-d8gj4" Nov 28 13:39:01 crc kubenswrapper[4631]: I1128 13:39:01.133189 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7303e352-744c-4c93-ac72-0c434384f090-public-tls-certs\") pod \"swift-proxy-68757c9bcf-d8gj4\" (UID: \"7303e352-744c-4c93-ac72-0c434384f090\") " pod="openstack/swift-proxy-68757c9bcf-d8gj4" Nov 28 13:39:01 crc kubenswrapper[4631]: I1128 13:39:01.133265 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-klmwm\" (UniqueName: \"kubernetes.io/projected/7303e352-744c-4c93-ac72-0c434384f090-kube-api-access-klmwm\") pod \"swift-proxy-68757c9bcf-d8gj4\" (UID: \"7303e352-744c-4c93-ac72-0c434384f090\") " pod="openstack/swift-proxy-68757c9bcf-d8gj4" Nov 28 13:39:01 crc kubenswrapper[4631]: I1128 13:39:01.133309 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7303e352-744c-4c93-ac72-0c434384f090-combined-ca-bundle\") pod \"swift-proxy-68757c9bcf-d8gj4\" (UID: \"7303e352-744c-4c93-ac72-0c434384f090\") " pod="openstack/swift-proxy-68757c9bcf-d8gj4" Nov 28 13:39:01 crc kubenswrapper[4631]: I1128 13:39:01.133382 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7303e352-744c-4c93-ac72-0c434384f090-config-data\") pod \"swift-proxy-68757c9bcf-d8gj4\" (UID: \"7303e352-744c-4c93-ac72-0c434384f090\") " pod="openstack/swift-proxy-68757c9bcf-d8gj4" Nov 28 13:39:01 crc kubenswrapper[4631]: I1128 13:39:01.133400 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7303e352-744c-4c93-ac72-0c434384f090-internal-tls-certs\") pod \"swift-proxy-68757c9bcf-d8gj4\" (UID: \"7303e352-744c-4c93-ac72-0c434384f090\") " pod="openstack/swift-proxy-68757c9bcf-d8gj4" Nov 28 13:39:01 crc kubenswrapper[4631]: I1128 13:39:01.133422 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7303e352-744c-4c93-ac72-0c434384f090-run-httpd\") pod \"swift-proxy-68757c9bcf-d8gj4\" (UID: \"7303e352-744c-4c93-ac72-0c434384f090\") " pod="openstack/swift-proxy-68757c9bcf-d8gj4" Nov 28 13:39:01 crc kubenswrapper[4631]: I1128 13:39:01.133445 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/7303e352-744c-4c93-ac72-0c434384f090-etc-swift\") pod \"swift-proxy-68757c9bcf-d8gj4\" (UID: \"7303e352-744c-4c93-ac72-0c434384f090\") " pod="openstack/swift-proxy-68757c9bcf-d8gj4" Nov 28 13:39:01 crc kubenswrapper[4631]: I1128 13:39:01.235004 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7303e352-744c-4c93-ac72-0c434384f090-public-tls-certs\") pod \"swift-proxy-68757c9bcf-d8gj4\" (UID: \"7303e352-744c-4c93-ac72-0c434384f090\") " pod="openstack/swift-proxy-68757c9bcf-d8gj4" Nov 28 13:39:01 crc kubenswrapper[4631]: I1128 13:39:01.235107 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-klmwm\" (UniqueName: \"kubernetes.io/projected/7303e352-744c-4c93-ac72-0c434384f090-kube-api-access-klmwm\") pod \"swift-proxy-68757c9bcf-d8gj4\" (UID: \"7303e352-744c-4c93-ac72-0c434384f090\") " pod="openstack/swift-proxy-68757c9bcf-d8gj4" Nov 28 13:39:01 crc kubenswrapper[4631]: I1128 13:39:01.235139 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7303e352-744c-4c93-ac72-0c434384f090-combined-ca-bundle\") pod \"swift-proxy-68757c9bcf-d8gj4\" (UID: \"7303e352-744c-4c93-ac72-0c434384f090\") " pod="openstack/swift-proxy-68757c9bcf-d8gj4" Nov 28 13:39:01 crc kubenswrapper[4631]: I1128 13:39:01.235221 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7303e352-744c-4c93-ac72-0c434384f090-config-data\") pod \"swift-proxy-68757c9bcf-d8gj4\" (UID: \"7303e352-744c-4c93-ac72-0c434384f090\") " pod="openstack/swift-proxy-68757c9bcf-d8gj4" Nov 28 13:39:01 crc kubenswrapper[4631]: I1128 13:39:01.235248 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7303e352-744c-4c93-ac72-0c434384f090-internal-tls-certs\") pod \"swift-proxy-68757c9bcf-d8gj4\" (UID: \"7303e352-744c-4c93-ac72-0c434384f090\") " pod="openstack/swift-proxy-68757c9bcf-d8gj4" Nov 28 13:39:01 crc kubenswrapper[4631]: I1128 13:39:01.235271 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7303e352-744c-4c93-ac72-0c434384f090-run-httpd\") pod \"swift-proxy-68757c9bcf-d8gj4\" (UID: \"7303e352-744c-4c93-ac72-0c434384f090\") " pod="openstack/swift-proxy-68757c9bcf-d8gj4" Nov 28 13:39:01 crc kubenswrapper[4631]: I1128 13:39:01.235312 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/7303e352-744c-4c93-ac72-0c434384f090-etc-swift\") pod \"swift-proxy-68757c9bcf-d8gj4\" (UID: \"7303e352-744c-4c93-ac72-0c434384f090\") " pod="openstack/swift-proxy-68757c9bcf-d8gj4" Nov 28 13:39:01 crc kubenswrapper[4631]: I1128 13:39:01.235367 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7303e352-744c-4c93-ac72-0c434384f090-log-httpd\") pod \"swift-proxy-68757c9bcf-d8gj4\" (UID: \"7303e352-744c-4c93-ac72-0c434384f090\") " pod="openstack/swift-proxy-68757c9bcf-d8gj4" Nov 28 13:39:01 crc kubenswrapper[4631]: I1128 13:39:01.235884 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7303e352-744c-4c93-ac72-0c434384f090-log-httpd\") pod \"swift-proxy-68757c9bcf-d8gj4\" (UID: \"7303e352-744c-4c93-ac72-0c434384f090\") " pod="openstack/swift-proxy-68757c9bcf-d8gj4" Nov 28 13:39:01 crc kubenswrapper[4631]: I1128 13:39:01.237452 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7303e352-744c-4c93-ac72-0c434384f090-run-httpd\") pod \"swift-proxy-68757c9bcf-d8gj4\" (UID: \"7303e352-744c-4c93-ac72-0c434384f090\") " pod="openstack/swift-proxy-68757c9bcf-d8gj4" Nov 28 13:39:01 crc kubenswrapper[4631]: I1128 13:39:01.243557 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7303e352-744c-4c93-ac72-0c434384f090-internal-tls-certs\") pod \"swift-proxy-68757c9bcf-d8gj4\" (UID: \"7303e352-744c-4c93-ac72-0c434384f090\") " pod="openstack/swift-proxy-68757c9bcf-d8gj4" Nov 28 13:39:01 crc kubenswrapper[4631]: I1128 13:39:01.244605 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7303e352-744c-4c93-ac72-0c434384f090-config-data\") pod \"swift-proxy-68757c9bcf-d8gj4\" (UID: \"7303e352-744c-4c93-ac72-0c434384f090\") " pod="openstack/swift-proxy-68757c9bcf-d8gj4" Nov 28 13:39:01 crc kubenswrapper[4631]: I1128 13:39:01.244716 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/7303e352-744c-4c93-ac72-0c434384f090-etc-swift\") pod \"swift-proxy-68757c9bcf-d8gj4\" (UID: \"7303e352-744c-4c93-ac72-0c434384f090\") " pod="openstack/swift-proxy-68757c9bcf-d8gj4" Nov 28 13:39:01 crc kubenswrapper[4631]: I1128 13:39:01.245321 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7303e352-744c-4c93-ac72-0c434384f090-public-tls-certs\") pod \"swift-proxy-68757c9bcf-d8gj4\" (UID: \"7303e352-744c-4c93-ac72-0c434384f090\") " pod="openstack/swift-proxy-68757c9bcf-d8gj4" Nov 28 13:39:01 crc kubenswrapper[4631]: I1128 13:39:01.246626 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7303e352-744c-4c93-ac72-0c434384f090-combined-ca-bundle\") pod \"swift-proxy-68757c9bcf-d8gj4\" (UID: \"7303e352-744c-4c93-ac72-0c434384f090\") " pod="openstack/swift-proxy-68757c9bcf-d8gj4" Nov 28 13:39:01 crc kubenswrapper[4631]: I1128 13:39:01.261321 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-klmwm\" (UniqueName: \"kubernetes.io/projected/7303e352-744c-4c93-ac72-0c434384f090-kube-api-access-klmwm\") pod \"swift-proxy-68757c9bcf-d8gj4\" (UID: \"7303e352-744c-4c93-ac72-0c434384f090\") " pod="openstack/swift-proxy-68757c9bcf-d8gj4" Nov 28 13:39:01 crc kubenswrapper[4631]: I1128 13:39:01.392969 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-68757c9bcf-d8gj4" Nov 28 13:39:01 crc kubenswrapper[4631]: I1128 13:39:01.534279 4631 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d95d3536-1392-4576-8ee5-390c09fe7a49" path="/var/lib/kubelet/pods/d95d3536-1392-4576-8ee5-390c09fe7a49/volumes" Nov 28 13:39:01 crc kubenswrapper[4631]: I1128 13:39:01.933750 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ac113fd1-daaf-4958-b779-aaee117db1ea","Type":"ContainerStarted","Data":"00ecb9ef637423c12a9a4d2165068b41e5935cba5464a65193c60c523a342172"} Nov 28 13:39:03 crc kubenswrapper[4631]: I1128 13:39:03.959411 4631 generic.go:334] "Generic (PLEG): container finished" podID="cb966e50-d246-48dc-b4a6-a9de4dbb8569" containerID="a869370244f2d82eb690074a62782ef88fa1793d2d6fd2b724555af086cd397f" exitCode=0 Nov 28 13:39:03 crc kubenswrapper[4631]: I1128 13:39:03.959631 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"cb966e50-d246-48dc-b4a6-a9de4dbb8569","Type":"ContainerDied","Data":"a869370244f2d82eb690074a62782ef88fa1793d2d6fd2b724555af086cd397f"} Nov 28 13:39:05 crc kubenswrapper[4631]: I1128 13:39:05.207428 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 28 13:39:05 crc kubenswrapper[4631]: I1128 13:39:05.208198 4631 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="e1d15959-31a9-464e-9e70-57547ddad347" containerName="glance-log" containerID="cri-o://91af2216f4119efeb231926969248c0ed81dd41d8674ea566c2a0befef60820f" gracePeriod=30 Nov 28 13:39:05 crc kubenswrapper[4631]: I1128 13:39:05.208770 4631 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="e1d15959-31a9-464e-9e70-57547ddad347" containerName="glance-httpd" containerID="cri-o://998cd494d09a844d61a201f6da645a13de219f74073a03e8f64f574ed113822b" gracePeriod=30 Nov 28 13:39:05 crc kubenswrapper[4631]: I1128 13:39:05.332457 4631 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-64f8fdf97b-jn5mg" Nov 28 13:39:05 crc kubenswrapper[4631]: I1128 13:39:05.332505 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-64f8fdf97b-jn5mg" Nov 28 13:39:05 crc kubenswrapper[4631]: I1128 13:39:05.333501 4631 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-64f8fdf97b-jn5mg" podUID="111c5349-528d-4856-a3e4-1277e3b10889" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.146:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.146:8443: connect: connection refused" Nov 28 13:39:05 crc kubenswrapper[4631]: I1128 13:39:05.554580 4631 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-544496c768-rzhsw" Nov 28 13:39:05 crc kubenswrapper[4631]: I1128 13:39:05.555423 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-544496c768-rzhsw" Nov 28 13:39:05 crc kubenswrapper[4631]: I1128 13:39:05.556107 4631 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-544496c768-rzhsw" podUID="cd607dae-568a-4b81-af81-3310c0e95854" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.147:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.147:8443: connect: connection refused" Nov 28 13:39:05 crc kubenswrapper[4631]: I1128 13:39:05.991581 4631 generic.go:334] "Generic (PLEG): container finished" podID="bef6bb99-5262-4f84-a2f7-0908a080f3a0" containerID="2fae4b9d182ba1a2f4c1ae5abd2caf92fbcae61387ec588c09af0ad1460aaf0a" exitCode=137 Nov 28 13:39:05 crc kubenswrapper[4631]: I1128 13:39:05.992006 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"bef6bb99-5262-4f84-a2f7-0908a080f3a0","Type":"ContainerDied","Data":"2fae4b9d182ba1a2f4c1ae5abd2caf92fbcae61387ec588c09af0ad1460aaf0a"} Nov 28 13:39:05 crc kubenswrapper[4631]: I1128 13:39:05.995652 4631 generic.go:334] "Generic (PLEG): container finished" podID="e1d15959-31a9-464e-9e70-57547ddad347" containerID="91af2216f4119efeb231926969248c0ed81dd41d8674ea566c2a0befef60820f" exitCode=143 Nov 28 13:39:05 crc kubenswrapper[4631]: I1128 13:39:05.996425 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"e1d15959-31a9-464e-9e70-57547ddad347","Type":"ContainerDied","Data":"91af2216f4119efeb231926969248c0ed81dd41d8674ea566c2a0befef60820f"} Nov 28 13:39:09 crc kubenswrapper[4631]: I1128 13:39:09.036392 4631 generic.go:334] "Generic (PLEG): container finished" podID="e1d15959-31a9-464e-9e70-57547ddad347" containerID="998cd494d09a844d61a201f6da645a13de219f74073a03e8f64f574ed113822b" exitCode=0 Nov 28 13:39:09 crc kubenswrapper[4631]: I1128 13:39:09.036641 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"e1d15959-31a9-464e-9e70-57547ddad347","Type":"ContainerDied","Data":"998cd494d09a844d61a201f6da645a13de219f74073a03e8f64f574ed113822b"} Nov 28 13:39:10 crc kubenswrapper[4631]: I1128 13:39:10.184019 4631 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/cinder-api-0" podUID="bef6bb99-5262-4f84-a2f7-0908a080f3a0" containerName="cinder-api" probeResult="failure" output="Get \"http://10.217.0.161:8776/healthcheck\": dial tcp 10.217.0.161:8776: connect: connection refused" Nov 28 13:39:11 crc kubenswrapper[4631]: E1128 13:39:11.138238 4631 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-openstackclient:current-podified" Nov 28 13:39:11 crc kubenswrapper[4631]: E1128 13:39:11.139041 4631 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:openstackclient,Image:quay.io/podified-antelope-centos9/openstack-openstackclient:current-podified,Command:[/bin/sleep],Args:[infinity],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n68fh5f6h64dh698h5f5hbdh689h64ch65ch54dh6dh644h5bbh68bh64fh5d5hbbh5cch544h658h664h685h56hddh57fh56bh67h54ch69h66fh57bh5bbq,ValueFrom:nil,},EnvVar{Name:OS_CLOUD,Value:default,ValueFrom:nil,},EnvVar{Name:PROMETHEUS_HOST,Value:metric-storage-prometheus.openstack.svc,ValueFrom:nil,},EnvVar{Name:PROMETHEUS_PORT,Value:9090,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:openstack-config,ReadOnly:false,MountPath:/home/cloud-admin/.config/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config-secret,ReadOnly:false,MountPath:/home/cloud-admin/.config/openstack/secure.yaml,SubPath:secure.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config-secret,ReadOnly:false,MountPath:/home/cloud-admin/cloudrc,SubPath:cloudrc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-qwv27,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42401,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*42401,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod openstackclient_openstack(ea063a3c-330a-4642-a815-c013ae6b80c7): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 28 13:39:11 crc kubenswrapper[4631]: E1128 13:39:11.143372 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"openstackclient\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/openstackclient" podUID="ea063a3c-330a-4642-a815-c013ae6b80c7" Nov 28 13:39:11 crc kubenswrapper[4631]: I1128 13:39:11.676007 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 28 13:39:11 crc kubenswrapper[4631]: I1128 13:39:11.786017 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/bef6bb99-5262-4f84-a2f7-0908a080f3a0-config-data-custom\") pod \"bef6bb99-5262-4f84-a2f7-0908a080f3a0\" (UID: \"bef6bb99-5262-4f84-a2f7-0908a080f3a0\") " Nov 28 13:39:11 crc kubenswrapper[4631]: I1128 13:39:11.786085 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bef6bb99-5262-4f84-a2f7-0908a080f3a0-config-data\") pod \"bef6bb99-5262-4f84-a2f7-0908a080f3a0\" (UID: \"bef6bb99-5262-4f84-a2f7-0908a080f3a0\") " Nov 28 13:39:11 crc kubenswrapper[4631]: I1128 13:39:11.786109 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x6plr\" (UniqueName: \"kubernetes.io/projected/bef6bb99-5262-4f84-a2f7-0908a080f3a0-kube-api-access-x6plr\") pod \"bef6bb99-5262-4f84-a2f7-0908a080f3a0\" (UID: \"bef6bb99-5262-4f84-a2f7-0908a080f3a0\") " Nov 28 13:39:11 crc kubenswrapper[4631]: I1128 13:39:11.786141 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/bef6bb99-5262-4f84-a2f7-0908a080f3a0-etc-machine-id\") pod \"bef6bb99-5262-4f84-a2f7-0908a080f3a0\" (UID: \"bef6bb99-5262-4f84-a2f7-0908a080f3a0\") " Nov 28 13:39:11 crc kubenswrapper[4631]: I1128 13:39:11.786290 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bef6bb99-5262-4f84-a2f7-0908a080f3a0-logs\") pod \"bef6bb99-5262-4f84-a2f7-0908a080f3a0\" (UID: \"bef6bb99-5262-4f84-a2f7-0908a080f3a0\") " Nov 28 13:39:11 crc kubenswrapper[4631]: I1128 13:39:11.786330 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bef6bb99-5262-4f84-a2f7-0908a080f3a0-scripts\") pod \"bef6bb99-5262-4f84-a2f7-0908a080f3a0\" (UID: \"bef6bb99-5262-4f84-a2f7-0908a080f3a0\") " Nov 28 13:39:11 crc kubenswrapper[4631]: I1128 13:39:11.786473 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bef6bb99-5262-4f84-a2f7-0908a080f3a0-combined-ca-bundle\") pod \"bef6bb99-5262-4f84-a2f7-0908a080f3a0\" (UID: \"bef6bb99-5262-4f84-a2f7-0908a080f3a0\") " Nov 28 13:39:11 crc kubenswrapper[4631]: I1128 13:39:11.788096 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bef6bb99-5262-4f84-a2f7-0908a080f3a0-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "bef6bb99-5262-4f84-a2f7-0908a080f3a0" (UID: "bef6bb99-5262-4f84-a2f7-0908a080f3a0"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 13:39:11 crc kubenswrapper[4631]: I1128 13:39:11.788733 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bef6bb99-5262-4f84-a2f7-0908a080f3a0-logs" (OuterVolumeSpecName: "logs") pod "bef6bb99-5262-4f84-a2f7-0908a080f3a0" (UID: "bef6bb99-5262-4f84-a2f7-0908a080f3a0"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 13:39:11 crc kubenswrapper[4631]: I1128 13:39:11.805778 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bef6bb99-5262-4f84-a2f7-0908a080f3a0-kube-api-access-x6plr" (OuterVolumeSpecName: "kube-api-access-x6plr") pod "bef6bb99-5262-4f84-a2f7-0908a080f3a0" (UID: "bef6bb99-5262-4f84-a2f7-0908a080f3a0"). InnerVolumeSpecName "kube-api-access-x6plr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:39:11 crc kubenswrapper[4631]: I1128 13:39:11.820509 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bef6bb99-5262-4f84-a2f7-0908a080f3a0-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "bef6bb99-5262-4f84-a2f7-0908a080f3a0" (UID: "bef6bb99-5262-4f84-a2f7-0908a080f3a0"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:39:11 crc kubenswrapper[4631]: I1128 13:39:11.841104 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bef6bb99-5262-4f84-a2f7-0908a080f3a0-scripts" (OuterVolumeSpecName: "scripts") pod "bef6bb99-5262-4f84-a2f7-0908a080f3a0" (UID: "bef6bb99-5262-4f84-a2f7-0908a080f3a0"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:39:11 crc kubenswrapper[4631]: I1128 13:39:11.885677 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bef6bb99-5262-4f84-a2f7-0908a080f3a0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bef6bb99-5262-4f84-a2f7-0908a080f3a0" (UID: "bef6bb99-5262-4f84-a2f7-0908a080f3a0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:39:11 crc kubenswrapper[4631]: I1128 13:39:11.892954 4631 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bef6bb99-5262-4f84-a2f7-0908a080f3a0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 13:39:11 crc kubenswrapper[4631]: I1128 13:39:11.892996 4631 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/bef6bb99-5262-4f84-a2f7-0908a080f3a0-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 28 13:39:11 crc kubenswrapper[4631]: I1128 13:39:11.893006 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x6plr\" (UniqueName: \"kubernetes.io/projected/bef6bb99-5262-4f84-a2f7-0908a080f3a0-kube-api-access-x6plr\") on node \"crc\" DevicePath \"\"" Nov 28 13:39:11 crc kubenswrapper[4631]: I1128 13:39:11.893016 4631 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/bef6bb99-5262-4f84-a2f7-0908a080f3a0-etc-machine-id\") on node \"crc\" DevicePath \"\"" Nov 28 13:39:11 crc kubenswrapper[4631]: I1128 13:39:11.893026 4631 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bef6bb99-5262-4f84-a2f7-0908a080f3a0-logs\") on node \"crc\" DevicePath \"\"" Nov 28 13:39:11 crc kubenswrapper[4631]: I1128 13:39:11.893038 4631 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bef6bb99-5262-4f84-a2f7-0908a080f3a0-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 13:39:11 crc kubenswrapper[4631]: I1128 13:39:11.971603 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 28 13:39:12 crc kubenswrapper[4631]: I1128 13:39:12.015933 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bef6bb99-5262-4f84-a2f7-0908a080f3a0-config-data" (OuterVolumeSpecName: "config-data") pod "bef6bb99-5262-4f84-a2f7-0908a080f3a0" (UID: "bef6bb99-5262-4f84-a2f7-0908a080f3a0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:39:12 crc kubenswrapper[4631]: I1128 13:39:12.019101 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e1d15959-31a9-464e-9e70-57547ddad347-combined-ca-bundle\") pod \"e1d15959-31a9-464e-9e70-57547ddad347\" (UID: \"e1d15959-31a9-464e-9e70-57547ddad347\") " Nov 28 13:39:12 crc kubenswrapper[4631]: I1128 13:39:12.024120 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e1d15959-31a9-464e-9e70-57547ddad347-logs\") pod \"e1d15959-31a9-464e-9e70-57547ddad347\" (UID: \"e1d15959-31a9-464e-9e70-57547ddad347\") " Nov 28 13:39:12 crc kubenswrapper[4631]: I1128 13:39:12.024934 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e1d15959-31a9-464e-9e70-57547ddad347-logs" (OuterVolumeSpecName: "logs") pod "e1d15959-31a9-464e-9e70-57547ddad347" (UID: "e1d15959-31a9-464e-9e70-57547ddad347"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 13:39:12 crc kubenswrapper[4631]: I1128 13:39:12.025234 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e1d15959-31a9-464e-9e70-57547ddad347-config-data\") pod \"e1d15959-31a9-464e-9e70-57547ddad347\" (UID: \"e1d15959-31a9-464e-9e70-57547ddad347\") " Nov 28 13:39:12 crc kubenswrapper[4631]: I1128 13:39:12.025717 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e1d15959-31a9-464e-9e70-57547ddad347-httpd-run\") pod \"e1d15959-31a9-464e-9e70-57547ddad347\" (UID: \"e1d15959-31a9-464e-9e70-57547ddad347\") " Nov 28 13:39:12 crc kubenswrapper[4631]: I1128 13:39:12.034039 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e1d15959-31a9-464e-9e70-57547ddad347-scripts\") pod \"e1d15959-31a9-464e-9e70-57547ddad347\" (UID: \"e1d15959-31a9-464e-9e70-57547ddad347\") " Nov 28 13:39:12 crc kubenswrapper[4631]: I1128 13:39:12.030897 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e1d15959-31a9-464e-9e70-57547ddad347-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "e1d15959-31a9-464e-9e70-57547ddad347" (UID: "e1d15959-31a9-464e-9e70-57547ddad347"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 13:39:12 crc kubenswrapper[4631]: I1128 13:39:12.044038 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e1d15959-31a9-464e-9e70-57547ddad347-scripts" (OuterVolumeSpecName: "scripts") pod "e1d15959-31a9-464e-9e70-57547ddad347" (UID: "e1d15959-31a9-464e-9e70-57547ddad347"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:39:12 crc kubenswrapper[4631]: I1128 13:39:12.047377 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kpkmg\" (UniqueName: \"kubernetes.io/projected/e1d15959-31a9-464e-9e70-57547ddad347-kube-api-access-kpkmg\") pod \"e1d15959-31a9-464e-9e70-57547ddad347\" (UID: \"e1d15959-31a9-464e-9e70-57547ddad347\") " Nov 28 13:39:12 crc kubenswrapper[4631]: I1128 13:39:12.047531 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"e1d15959-31a9-464e-9e70-57547ddad347\" (UID: \"e1d15959-31a9-464e-9e70-57547ddad347\") " Nov 28 13:39:12 crc kubenswrapper[4631]: I1128 13:39:12.048063 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bef6bb99-5262-4f84-a2f7-0908a080f3a0-config-data\") pod \"bef6bb99-5262-4f84-a2f7-0908a080f3a0\" (UID: \"bef6bb99-5262-4f84-a2f7-0908a080f3a0\") " Nov 28 13:39:12 crc kubenswrapper[4631]: I1128 13:39:12.048207 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e1d15959-31a9-464e-9e70-57547ddad347-public-tls-certs\") pod \"e1d15959-31a9-464e-9e70-57547ddad347\" (UID: \"e1d15959-31a9-464e-9e70-57547ddad347\") " Nov 28 13:39:12 crc kubenswrapper[4631]: W1128 13:39:12.048459 4631 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/bef6bb99-5262-4f84-a2f7-0908a080f3a0/volumes/kubernetes.io~secret/config-data Nov 28 13:39:12 crc kubenswrapper[4631]: I1128 13:39:12.048495 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bef6bb99-5262-4f84-a2f7-0908a080f3a0-config-data" (OuterVolumeSpecName: "config-data") pod "bef6bb99-5262-4f84-a2f7-0908a080f3a0" (UID: "bef6bb99-5262-4f84-a2f7-0908a080f3a0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:39:12 crc kubenswrapper[4631]: I1128 13:39:12.051379 4631 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e1d15959-31a9-464e-9e70-57547ddad347-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 13:39:12 crc kubenswrapper[4631]: I1128 13:39:12.054057 4631 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bef6bb99-5262-4f84-a2f7-0908a080f3a0-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 13:39:12 crc kubenswrapper[4631]: I1128 13:39:12.054145 4631 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e1d15959-31a9-464e-9e70-57547ddad347-logs\") on node \"crc\" DevicePath \"\"" Nov 28 13:39:12 crc kubenswrapper[4631]: I1128 13:39:12.054217 4631 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e1d15959-31a9-464e-9e70-57547ddad347-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 28 13:39:12 crc kubenswrapper[4631]: I1128 13:39:12.051472 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e1d15959-31a9-464e-9e70-57547ddad347-kube-api-access-kpkmg" (OuterVolumeSpecName: "kube-api-access-kpkmg") pod "e1d15959-31a9-464e-9e70-57547ddad347" (UID: "e1d15959-31a9-464e-9e70-57547ddad347"). InnerVolumeSpecName "kube-api-access-kpkmg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:39:12 crc kubenswrapper[4631]: I1128 13:39:12.065498 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage04-crc" (OuterVolumeSpecName: "glance") pod "e1d15959-31a9-464e-9e70-57547ddad347" (UID: "e1d15959-31a9-464e-9e70-57547ddad347"). InnerVolumeSpecName "local-storage04-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 28 13:39:12 crc kubenswrapper[4631]: I1128 13:39:12.140344 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"e1d15959-31a9-464e-9e70-57547ddad347","Type":"ContainerDied","Data":"889aaf4a8a635294c7495304ce6462870ca87110eefe5667edfc7352a064dc36"} Nov 28 13:39:12 crc kubenswrapper[4631]: I1128 13:39:12.140401 4631 scope.go:117] "RemoveContainer" containerID="998cd494d09a844d61a201f6da645a13de219f74073a03e8f64f574ed113822b" Nov 28 13:39:12 crc kubenswrapper[4631]: I1128 13:39:12.140577 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 28 13:39:12 crc kubenswrapper[4631]: I1128 13:39:12.147154 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ac113fd1-daaf-4958-b779-aaee117db1ea","Type":"ContainerStarted","Data":"0053c434df76403c94f1221253f347b5af0b8d4b85b3e4d8847ca04e89fef6a6"} Nov 28 13:39:12 crc kubenswrapper[4631]: I1128 13:39:12.174157 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kpkmg\" (UniqueName: \"kubernetes.io/projected/e1d15959-31a9-464e-9e70-57547ddad347-kube-api-access-kpkmg\") on node \"crc\" DevicePath \"\"" Nov 28 13:39:12 crc kubenswrapper[4631]: I1128 13:39:12.174211 4631 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" " Nov 28 13:39:12 crc kubenswrapper[4631]: I1128 13:39:12.175225 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"bef6bb99-5262-4f84-a2f7-0908a080f3a0","Type":"ContainerDied","Data":"a63af8a28bdb74f133cd9aba1c8b756b01dccc91b42f05e57625c4b57feba93b"} Nov 28 13:39:12 crc kubenswrapper[4631]: I1128 13:39:12.176603 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 28 13:39:12 crc kubenswrapper[4631]: E1128 13:39:12.189405 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"openstackclient\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-openstackclient:current-podified\\\"\"" pod="openstack/openstackclient" podUID="ea063a3c-330a-4642-a815-c013ae6b80c7" Nov 28 13:39:12 crc kubenswrapper[4631]: I1128 13:39:12.239212 4631 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage04-crc" (UniqueName: "kubernetes.io/local-volume/local-storage04-crc") on node "crc" Nov 28 13:39:12 crc kubenswrapper[4631]: I1128 13:39:12.252831 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e1d15959-31a9-464e-9e70-57547ddad347-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e1d15959-31a9-464e-9e70-57547ddad347" (UID: "e1d15959-31a9-464e-9e70-57547ddad347"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:39:12 crc kubenswrapper[4631]: I1128 13:39:12.273796 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e1d15959-31a9-464e-9e70-57547ddad347-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "e1d15959-31a9-464e-9e70-57547ddad347" (UID: "e1d15959-31a9-464e-9e70-57547ddad347"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:39:12 crc kubenswrapper[4631]: I1128 13:39:12.276609 4631 reconciler_common.go:293] "Volume detached for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" DevicePath \"\"" Nov 28 13:39:12 crc kubenswrapper[4631]: I1128 13:39:12.276654 4631 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e1d15959-31a9-464e-9e70-57547ddad347-public-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 28 13:39:12 crc kubenswrapper[4631]: I1128 13:39:12.276665 4631 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e1d15959-31a9-464e-9e70-57547ddad347-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 13:39:12 crc kubenswrapper[4631]: I1128 13:39:12.305478 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e1d15959-31a9-464e-9e70-57547ddad347-config-data" (OuterVolumeSpecName: "config-data") pod "e1d15959-31a9-464e-9e70-57547ddad347" (UID: "e1d15959-31a9-464e-9e70-57547ddad347"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:39:12 crc kubenswrapper[4631]: I1128 13:39:12.379915 4631 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e1d15959-31a9-464e-9e70-57547ddad347-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 13:39:12 crc kubenswrapper[4631]: I1128 13:39:12.435521 4631 scope.go:117] "RemoveContainer" containerID="91af2216f4119efeb231926969248c0ed81dd41d8674ea566c2a0befef60820f" Nov 28 13:39:12 crc kubenswrapper[4631]: I1128 13:39:12.436712 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 28 13:39:12 crc kubenswrapper[4631]: I1128 13:39:12.551664 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Nov 28 13:39:12 crc kubenswrapper[4631]: I1128 13:39:12.580507 4631 scope.go:117] "RemoveContainer" containerID="2fae4b9d182ba1a2f4c1ae5abd2caf92fbcae61387ec588c09af0ad1460aaf0a" Nov 28 13:39:12 crc kubenswrapper[4631]: I1128 13:39:12.601153 4631 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Nov 28 13:39:12 crc kubenswrapper[4631]: I1128 13:39:12.632312 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cb966e50-d246-48dc-b4a6-a9de4dbb8569-config-data\") pod \"cb966e50-d246-48dc-b4a6-a9de4dbb8569\" (UID: \"cb966e50-d246-48dc-b4a6-a9de4dbb8569\") " Nov 28 13:39:12 crc kubenswrapper[4631]: I1128 13:39:12.632395 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/cb966e50-d246-48dc-b4a6-a9de4dbb8569-internal-tls-certs\") pod \"cb966e50-d246-48dc-b4a6-a9de4dbb8569\" (UID: \"cb966e50-d246-48dc-b4a6-a9de4dbb8569\") " Nov 28 13:39:12 crc kubenswrapper[4631]: I1128 13:39:12.632449 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cb966e50-d246-48dc-b4a6-a9de4dbb8569-logs\") pod \"cb966e50-d246-48dc-b4a6-a9de4dbb8569\" (UID: \"cb966e50-d246-48dc-b4a6-a9de4dbb8569\") " Nov 28 13:39:12 crc kubenswrapper[4631]: I1128 13:39:12.632502 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jksgc\" (UniqueName: \"kubernetes.io/projected/cb966e50-d246-48dc-b4a6-a9de4dbb8569-kube-api-access-jksgc\") pod \"cb966e50-d246-48dc-b4a6-a9de4dbb8569\" (UID: \"cb966e50-d246-48dc-b4a6-a9de4dbb8569\") " Nov 28 13:39:12 crc kubenswrapper[4631]: I1128 13:39:12.632656 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"cb966e50-d246-48dc-b4a6-a9de4dbb8569\" (UID: \"cb966e50-d246-48dc-b4a6-a9de4dbb8569\") " Nov 28 13:39:12 crc kubenswrapper[4631]: I1128 13:39:12.632792 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cb966e50-d246-48dc-b4a6-a9de4dbb8569-scripts\") pod \"cb966e50-d246-48dc-b4a6-a9de4dbb8569\" (UID: \"cb966e50-d246-48dc-b4a6-a9de4dbb8569\") " Nov 28 13:39:12 crc kubenswrapper[4631]: I1128 13:39:12.632820 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/cb966e50-d246-48dc-b4a6-a9de4dbb8569-httpd-run\") pod \"cb966e50-d246-48dc-b4a6-a9de4dbb8569\" (UID: \"cb966e50-d246-48dc-b4a6-a9de4dbb8569\") " Nov 28 13:39:12 crc kubenswrapper[4631]: I1128 13:39:12.637895 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cb966e50-d246-48dc-b4a6-a9de4dbb8569-combined-ca-bundle\") pod \"cb966e50-d246-48dc-b4a6-a9de4dbb8569\" (UID: \"cb966e50-d246-48dc-b4a6-a9de4dbb8569\") " Nov 28 13:39:12 crc kubenswrapper[4631]: I1128 13:39:12.670693 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cb966e50-d246-48dc-b4a6-a9de4dbb8569-logs" (OuterVolumeSpecName: "logs") pod "cb966e50-d246-48dc-b4a6-a9de4dbb8569" (UID: "cb966e50-d246-48dc-b4a6-a9de4dbb8569"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 13:39:12 crc kubenswrapper[4631]: I1128 13:39:12.671070 4631 scope.go:117] "RemoveContainer" containerID="d9367482c870497f738de7f33c554511341371eb34004d4968e56132c4122076" Nov 28 13:39:12 crc kubenswrapper[4631]: I1128 13:39:12.671928 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cb966e50-d246-48dc-b4a6-a9de4dbb8569-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "cb966e50-d246-48dc-b4a6-a9de4dbb8569" (UID: "cb966e50-d246-48dc-b4a6-a9de4dbb8569"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 13:39:12 crc kubenswrapper[4631]: I1128 13:39:12.671979 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Nov 28 13:39:12 crc kubenswrapper[4631]: E1128 13:39:12.676352 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bef6bb99-5262-4f84-a2f7-0908a080f3a0" containerName="cinder-api-log" Nov 28 13:39:12 crc kubenswrapper[4631]: I1128 13:39:12.676379 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="bef6bb99-5262-4f84-a2f7-0908a080f3a0" containerName="cinder-api-log" Nov 28 13:39:12 crc kubenswrapper[4631]: E1128 13:39:12.676417 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cb966e50-d246-48dc-b4a6-a9de4dbb8569" containerName="glance-httpd" Nov 28 13:39:12 crc kubenswrapper[4631]: I1128 13:39:12.676423 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="cb966e50-d246-48dc-b4a6-a9de4dbb8569" containerName="glance-httpd" Nov 28 13:39:12 crc kubenswrapper[4631]: E1128 13:39:12.676442 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bef6bb99-5262-4f84-a2f7-0908a080f3a0" containerName="cinder-api" Nov 28 13:39:12 crc kubenswrapper[4631]: I1128 13:39:12.676449 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="bef6bb99-5262-4f84-a2f7-0908a080f3a0" containerName="cinder-api" Nov 28 13:39:12 crc kubenswrapper[4631]: E1128 13:39:12.676471 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e1d15959-31a9-464e-9e70-57547ddad347" containerName="glance-httpd" Nov 28 13:39:12 crc kubenswrapper[4631]: I1128 13:39:12.676478 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="e1d15959-31a9-464e-9e70-57547ddad347" containerName="glance-httpd" Nov 28 13:39:12 crc kubenswrapper[4631]: E1128 13:39:12.676496 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e1d15959-31a9-464e-9e70-57547ddad347" containerName="glance-log" Nov 28 13:39:12 crc kubenswrapper[4631]: I1128 13:39:12.676502 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="e1d15959-31a9-464e-9e70-57547ddad347" containerName="glance-log" Nov 28 13:39:12 crc kubenswrapper[4631]: E1128 13:39:12.676532 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cb966e50-d246-48dc-b4a6-a9de4dbb8569" containerName="glance-log" Nov 28 13:39:12 crc kubenswrapper[4631]: I1128 13:39:12.676539 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="cb966e50-d246-48dc-b4a6-a9de4dbb8569" containerName="glance-log" Nov 28 13:39:12 crc kubenswrapper[4631]: I1128 13:39:12.676867 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="bef6bb99-5262-4f84-a2f7-0908a080f3a0" containerName="cinder-api-log" Nov 28 13:39:12 crc kubenswrapper[4631]: I1128 13:39:12.676908 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="e1d15959-31a9-464e-9e70-57547ddad347" containerName="glance-log" Nov 28 13:39:12 crc kubenswrapper[4631]: I1128 13:39:12.696239 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="e1d15959-31a9-464e-9e70-57547ddad347" containerName="glance-httpd" Nov 28 13:39:12 crc kubenswrapper[4631]: I1128 13:39:12.696317 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="cb966e50-d246-48dc-b4a6-a9de4dbb8569" containerName="glance-log" Nov 28 13:39:12 crc kubenswrapper[4631]: I1128 13:39:12.686246 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage08-crc" (OuterVolumeSpecName: "glance") pod "cb966e50-d246-48dc-b4a6-a9de4dbb8569" (UID: "cb966e50-d246-48dc-b4a6-a9de4dbb8569"). InnerVolumeSpecName "local-storage08-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 28 13:39:12 crc kubenswrapper[4631]: I1128 13:39:12.696084 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cb966e50-d246-48dc-b4a6-a9de4dbb8569-scripts" (OuterVolumeSpecName: "scripts") pod "cb966e50-d246-48dc-b4a6-a9de4dbb8569" (UID: "cb966e50-d246-48dc-b4a6-a9de4dbb8569"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:39:12 crc kubenswrapper[4631]: I1128 13:39:12.701928 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cb966e50-d246-48dc-b4a6-a9de4dbb8569-kube-api-access-jksgc" (OuterVolumeSpecName: "kube-api-access-jksgc") pod "cb966e50-d246-48dc-b4a6-a9de4dbb8569" (UID: "cb966e50-d246-48dc-b4a6-a9de4dbb8569"). InnerVolumeSpecName "kube-api-access-jksgc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:39:12 crc kubenswrapper[4631]: I1128 13:39:12.707191 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="bef6bb99-5262-4f84-a2f7-0908a080f3a0" containerName="cinder-api" Nov 28 13:39:12 crc kubenswrapper[4631]: I1128 13:39:12.707267 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="cb966e50-d246-48dc-b4a6-a9de4dbb8569" containerName="glance-httpd" Nov 28 13:39:12 crc kubenswrapper[4631]: I1128 13:39:12.723277 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 28 13:39:12 crc kubenswrapper[4631]: I1128 13:39:12.748907 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Nov 28 13:39:12 crc kubenswrapper[4631]: I1128 13:39:12.749184 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Nov 28 13:39:12 crc kubenswrapper[4631]: I1128 13:39:12.749375 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Nov 28 13:39:12 crc kubenswrapper[4631]: I1128 13:39:12.760531 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Nov 28 13:39:12 crc kubenswrapper[4631]: I1128 13:39:12.762692 4631 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cb966e50-d246-48dc-b4a6-a9de4dbb8569-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 13:39:12 crc kubenswrapper[4631]: I1128 13:39:12.762823 4631 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/cb966e50-d246-48dc-b4a6-a9de4dbb8569-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 28 13:39:12 crc kubenswrapper[4631]: I1128 13:39:12.762881 4631 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cb966e50-d246-48dc-b4a6-a9de4dbb8569-logs\") on node \"crc\" DevicePath \"\"" Nov 28 13:39:12 crc kubenswrapper[4631]: I1128 13:39:12.762942 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jksgc\" (UniqueName: \"kubernetes.io/projected/cb966e50-d246-48dc-b4a6-a9de4dbb8569-kube-api-access-jksgc\") on node \"crc\" DevicePath \"\"" Nov 28 13:39:12 crc kubenswrapper[4631]: I1128 13:39:12.763016 4631 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" " Nov 28 13:39:12 crc kubenswrapper[4631]: I1128 13:39:12.865023 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/86f5dab6-6b28-494c-b9d6-52528c726875-public-tls-certs\") pod \"cinder-api-0\" (UID: \"86f5dab6-6b28-494c-b9d6-52528c726875\") " pod="openstack/cinder-api-0" Nov 28 13:39:12 crc kubenswrapper[4631]: I1128 13:39:12.865089 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/86f5dab6-6b28-494c-b9d6-52528c726875-etc-machine-id\") pod \"cinder-api-0\" (UID: \"86f5dab6-6b28-494c-b9d6-52528c726875\") " pod="openstack/cinder-api-0" Nov 28 13:39:12 crc kubenswrapper[4631]: I1128 13:39:12.865118 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r9sps\" (UniqueName: \"kubernetes.io/projected/86f5dab6-6b28-494c-b9d6-52528c726875-kube-api-access-r9sps\") pod \"cinder-api-0\" (UID: \"86f5dab6-6b28-494c-b9d6-52528c726875\") " pod="openstack/cinder-api-0" Nov 28 13:39:12 crc kubenswrapper[4631]: I1128 13:39:12.865161 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/86f5dab6-6b28-494c-b9d6-52528c726875-scripts\") pod \"cinder-api-0\" (UID: \"86f5dab6-6b28-494c-b9d6-52528c726875\") " pod="openstack/cinder-api-0" Nov 28 13:39:12 crc kubenswrapper[4631]: I1128 13:39:12.865251 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/86f5dab6-6b28-494c-b9d6-52528c726875-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"86f5dab6-6b28-494c-b9d6-52528c726875\") " pod="openstack/cinder-api-0" Nov 28 13:39:12 crc kubenswrapper[4631]: I1128 13:39:12.865350 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/86f5dab6-6b28-494c-b9d6-52528c726875-config-data-custom\") pod \"cinder-api-0\" (UID: \"86f5dab6-6b28-494c-b9d6-52528c726875\") " pod="openstack/cinder-api-0" Nov 28 13:39:12 crc kubenswrapper[4631]: I1128 13:39:12.865423 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/86f5dab6-6b28-494c-b9d6-52528c726875-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"86f5dab6-6b28-494c-b9d6-52528c726875\") " pod="openstack/cinder-api-0" Nov 28 13:39:12 crc kubenswrapper[4631]: I1128 13:39:12.865458 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/86f5dab6-6b28-494c-b9d6-52528c726875-config-data\") pod \"cinder-api-0\" (UID: \"86f5dab6-6b28-494c-b9d6-52528c726875\") " pod="openstack/cinder-api-0" Nov 28 13:39:12 crc kubenswrapper[4631]: I1128 13:39:12.865502 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/86f5dab6-6b28-494c-b9d6-52528c726875-logs\") pod \"cinder-api-0\" (UID: \"86f5dab6-6b28-494c-b9d6-52528c726875\") " pod="openstack/cinder-api-0" Nov 28 13:39:12 crc kubenswrapper[4631]: I1128 13:39:12.867354 4631 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage08-crc" (UniqueName: "kubernetes.io/local-volume/local-storage08-crc") on node "crc" Nov 28 13:39:12 crc kubenswrapper[4631]: I1128 13:39:12.881489 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cb966e50-d246-48dc-b4a6-a9de4dbb8569-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "cb966e50-d246-48dc-b4a6-a9de4dbb8569" (UID: "cb966e50-d246-48dc-b4a6-a9de4dbb8569"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:39:12 crc kubenswrapper[4631]: I1128 13:39:12.886367 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-68757c9bcf-d8gj4"] Nov 28 13:39:12 crc kubenswrapper[4631]: I1128 13:39:12.898649 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cb966e50-d246-48dc-b4a6-a9de4dbb8569-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "cb966e50-d246-48dc-b4a6-a9de4dbb8569" (UID: "cb966e50-d246-48dc-b4a6-a9de4dbb8569"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:39:12 crc kubenswrapper[4631]: I1128 13:39:12.917413 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 28 13:39:12 crc kubenswrapper[4631]: I1128 13:39:12.917731 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cb966e50-d246-48dc-b4a6-a9de4dbb8569-config-data" (OuterVolumeSpecName: "config-data") pod "cb966e50-d246-48dc-b4a6-a9de4dbb8569" (UID: "cb966e50-d246-48dc-b4a6-a9de4dbb8569"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:39:12 crc kubenswrapper[4631]: I1128 13:39:12.947861 4631 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 28 13:39:12 crc kubenswrapper[4631]: I1128 13:39:12.970344 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/86f5dab6-6b28-494c-b9d6-52528c726875-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"86f5dab6-6b28-494c-b9d6-52528c726875\") " pod="openstack/cinder-api-0" Nov 28 13:39:12 crc kubenswrapper[4631]: I1128 13:39:12.970743 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/86f5dab6-6b28-494c-b9d6-52528c726875-config-data-custom\") pod \"cinder-api-0\" (UID: \"86f5dab6-6b28-494c-b9d6-52528c726875\") " pod="openstack/cinder-api-0" Nov 28 13:39:12 crc kubenswrapper[4631]: I1128 13:39:12.970784 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/86f5dab6-6b28-494c-b9d6-52528c726875-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"86f5dab6-6b28-494c-b9d6-52528c726875\") " pod="openstack/cinder-api-0" Nov 28 13:39:12 crc kubenswrapper[4631]: I1128 13:39:12.970805 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/86f5dab6-6b28-494c-b9d6-52528c726875-config-data\") pod \"cinder-api-0\" (UID: \"86f5dab6-6b28-494c-b9d6-52528c726875\") " pod="openstack/cinder-api-0" Nov 28 13:39:12 crc kubenswrapper[4631]: I1128 13:39:12.970842 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/86f5dab6-6b28-494c-b9d6-52528c726875-logs\") pod \"cinder-api-0\" (UID: \"86f5dab6-6b28-494c-b9d6-52528c726875\") " pod="openstack/cinder-api-0" Nov 28 13:39:12 crc kubenswrapper[4631]: I1128 13:39:12.970884 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/86f5dab6-6b28-494c-b9d6-52528c726875-public-tls-certs\") pod \"cinder-api-0\" (UID: \"86f5dab6-6b28-494c-b9d6-52528c726875\") " pod="openstack/cinder-api-0" Nov 28 13:39:12 crc kubenswrapper[4631]: I1128 13:39:12.970902 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/86f5dab6-6b28-494c-b9d6-52528c726875-etc-machine-id\") pod \"cinder-api-0\" (UID: \"86f5dab6-6b28-494c-b9d6-52528c726875\") " pod="openstack/cinder-api-0" Nov 28 13:39:12 crc kubenswrapper[4631]: I1128 13:39:12.970924 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r9sps\" (UniqueName: \"kubernetes.io/projected/86f5dab6-6b28-494c-b9d6-52528c726875-kube-api-access-r9sps\") pod \"cinder-api-0\" (UID: \"86f5dab6-6b28-494c-b9d6-52528c726875\") " pod="openstack/cinder-api-0" Nov 28 13:39:12 crc kubenswrapper[4631]: I1128 13:39:12.970955 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/86f5dab6-6b28-494c-b9d6-52528c726875-scripts\") pod \"cinder-api-0\" (UID: \"86f5dab6-6b28-494c-b9d6-52528c726875\") " pod="openstack/cinder-api-0" Nov 28 13:39:12 crc kubenswrapper[4631]: I1128 13:39:12.971065 4631 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cb966e50-d246-48dc-b4a6-a9de4dbb8569-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 13:39:12 crc kubenswrapper[4631]: I1128 13:39:12.971079 4631 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cb966e50-d246-48dc-b4a6-a9de4dbb8569-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 13:39:12 crc kubenswrapper[4631]: I1128 13:39:12.971089 4631 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/cb966e50-d246-48dc-b4a6-a9de4dbb8569-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 28 13:39:12 crc kubenswrapper[4631]: I1128 13:39:12.971100 4631 reconciler_common.go:293] "Volume detached for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" DevicePath \"\"" Nov 28 13:39:12 crc kubenswrapper[4631]: I1128 13:39:12.979096 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/86f5dab6-6b28-494c-b9d6-52528c726875-config-data\") pod \"cinder-api-0\" (UID: \"86f5dab6-6b28-494c-b9d6-52528c726875\") " pod="openstack/cinder-api-0" Nov 28 13:39:12 crc kubenswrapper[4631]: I1128 13:39:12.979200 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/86f5dab6-6b28-494c-b9d6-52528c726875-etc-machine-id\") pod \"cinder-api-0\" (UID: \"86f5dab6-6b28-494c-b9d6-52528c726875\") " pod="openstack/cinder-api-0" Nov 28 13:39:12 crc kubenswrapper[4631]: I1128 13:39:12.979992 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/86f5dab6-6b28-494c-b9d6-52528c726875-config-data-custom\") pod \"cinder-api-0\" (UID: \"86f5dab6-6b28-494c-b9d6-52528c726875\") " pod="openstack/cinder-api-0" Nov 28 13:39:12 crc kubenswrapper[4631]: I1128 13:39:12.981566 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/86f5dab6-6b28-494c-b9d6-52528c726875-logs\") pod \"cinder-api-0\" (UID: \"86f5dab6-6b28-494c-b9d6-52528c726875\") " pod="openstack/cinder-api-0" Nov 28 13:39:12 crc kubenswrapper[4631]: I1128 13:39:12.984842 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/86f5dab6-6b28-494c-b9d6-52528c726875-scripts\") pod \"cinder-api-0\" (UID: \"86f5dab6-6b28-494c-b9d6-52528c726875\") " pod="openstack/cinder-api-0" Nov 28 13:39:12 crc kubenswrapper[4631]: I1128 13:39:12.985878 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/86f5dab6-6b28-494c-b9d6-52528c726875-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"86f5dab6-6b28-494c-b9d6-52528c726875\") " pod="openstack/cinder-api-0" Nov 28 13:39:12 crc kubenswrapper[4631]: I1128 13:39:12.987445 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/86f5dab6-6b28-494c-b9d6-52528c726875-public-tls-certs\") pod \"cinder-api-0\" (UID: \"86f5dab6-6b28-494c-b9d6-52528c726875\") " pod="openstack/cinder-api-0" Nov 28 13:39:12 crc kubenswrapper[4631]: I1128 13:39:12.995073 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/86f5dab6-6b28-494c-b9d6-52528c726875-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"86f5dab6-6b28-494c-b9d6-52528c726875\") " pod="openstack/cinder-api-0" Nov 28 13:39:13 crc kubenswrapper[4631]: I1128 13:39:12.999962 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r9sps\" (UniqueName: \"kubernetes.io/projected/86f5dab6-6b28-494c-b9d6-52528c726875-kube-api-access-r9sps\") pod \"cinder-api-0\" (UID: \"86f5dab6-6b28-494c-b9d6-52528c726875\") " pod="openstack/cinder-api-0" Nov 28 13:39:13 crc kubenswrapper[4631]: I1128 13:39:13.002150 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Nov 28 13:39:13 crc kubenswrapper[4631]: I1128 13:39:13.004503 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 28 13:39:13 crc kubenswrapper[4631]: I1128 13:39:13.008816 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Nov 28 13:39:13 crc kubenswrapper[4631]: I1128 13:39:13.008998 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Nov 28 13:39:13 crc kubenswrapper[4631]: I1128 13:39:13.017717 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 28 13:39:13 crc kubenswrapper[4631]: I1128 13:39:13.072756 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rhc6t\" (UniqueName: \"kubernetes.io/projected/aeea572b-1f4e-486b-a916-1b34f0b84d10-kube-api-access-rhc6t\") pod \"glance-default-external-api-0\" (UID: \"aeea572b-1f4e-486b-a916-1b34f0b84d10\") " pod="openstack/glance-default-external-api-0" Nov 28 13:39:13 crc kubenswrapper[4631]: I1128 13:39:13.072881 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-external-api-0\" (UID: \"aeea572b-1f4e-486b-a916-1b34f0b84d10\") " pod="openstack/glance-default-external-api-0" Nov 28 13:39:13 crc kubenswrapper[4631]: I1128 13:39:13.072925 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/aeea572b-1f4e-486b-a916-1b34f0b84d10-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"aeea572b-1f4e-486b-a916-1b34f0b84d10\") " pod="openstack/glance-default-external-api-0" Nov 28 13:39:13 crc kubenswrapper[4631]: I1128 13:39:13.072984 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/aeea572b-1f4e-486b-a916-1b34f0b84d10-logs\") pod \"glance-default-external-api-0\" (UID: \"aeea572b-1f4e-486b-a916-1b34f0b84d10\") " pod="openstack/glance-default-external-api-0" Nov 28 13:39:13 crc kubenswrapper[4631]: I1128 13:39:13.073012 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/aeea572b-1f4e-486b-a916-1b34f0b84d10-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"aeea572b-1f4e-486b-a916-1b34f0b84d10\") " pod="openstack/glance-default-external-api-0" Nov 28 13:39:13 crc kubenswrapper[4631]: I1128 13:39:13.073039 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aeea572b-1f4e-486b-a916-1b34f0b84d10-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"aeea572b-1f4e-486b-a916-1b34f0b84d10\") " pod="openstack/glance-default-external-api-0" Nov 28 13:39:13 crc kubenswrapper[4631]: I1128 13:39:13.073087 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aeea572b-1f4e-486b-a916-1b34f0b84d10-config-data\") pod \"glance-default-external-api-0\" (UID: \"aeea572b-1f4e-486b-a916-1b34f0b84d10\") " pod="openstack/glance-default-external-api-0" Nov 28 13:39:13 crc kubenswrapper[4631]: I1128 13:39:13.073134 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aeea572b-1f4e-486b-a916-1b34f0b84d10-scripts\") pod \"glance-default-external-api-0\" (UID: \"aeea572b-1f4e-486b-a916-1b34f0b84d10\") " pod="openstack/glance-default-external-api-0" Nov 28 13:39:13 crc kubenswrapper[4631]: I1128 13:39:13.120625 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 28 13:39:13 crc kubenswrapper[4631]: I1128 13:39:13.174632 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-external-api-0\" (UID: \"aeea572b-1f4e-486b-a916-1b34f0b84d10\") " pod="openstack/glance-default-external-api-0" Nov 28 13:39:13 crc kubenswrapper[4631]: I1128 13:39:13.174682 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/aeea572b-1f4e-486b-a916-1b34f0b84d10-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"aeea572b-1f4e-486b-a916-1b34f0b84d10\") " pod="openstack/glance-default-external-api-0" Nov 28 13:39:13 crc kubenswrapper[4631]: I1128 13:39:13.174734 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/aeea572b-1f4e-486b-a916-1b34f0b84d10-logs\") pod \"glance-default-external-api-0\" (UID: \"aeea572b-1f4e-486b-a916-1b34f0b84d10\") " pod="openstack/glance-default-external-api-0" Nov 28 13:39:13 crc kubenswrapper[4631]: I1128 13:39:13.174755 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/aeea572b-1f4e-486b-a916-1b34f0b84d10-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"aeea572b-1f4e-486b-a916-1b34f0b84d10\") " pod="openstack/glance-default-external-api-0" Nov 28 13:39:13 crc kubenswrapper[4631]: I1128 13:39:13.174772 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aeea572b-1f4e-486b-a916-1b34f0b84d10-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"aeea572b-1f4e-486b-a916-1b34f0b84d10\") " pod="openstack/glance-default-external-api-0" Nov 28 13:39:13 crc kubenswrapper[4631]: I1128 13:39:13.174807 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aeea572b-1f4e-486b-a916-1b34f0b84d10-config-data\") pod \"glance-default-external-api-0\" (UID: \"aeea572b-1f4e-486b-a916-1b34f0b84d10\") " pod="openstack/glance-default-external-api-0" Nov 28 13:39:13 crc kubenswrapper[4631]: I1128 13:39:13.174843 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aeea572b-1f4e-486b-a916-1b34f0b84d10-scripts\") pod \"glance-default-external-api-0\" (UID: \"aeea572b-1f4e-486b-a916-1b34f0b84d10\") " pod="openstack/glance-default-external-api-0" Nov 28 13:39:13 crc kubenswrapper[4631]: I1128 13:39:13.174887 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rhc6t\" (UniqueName: \"kubernetes.io/projected/aeea572b-1f4e-486b-a916-1b34f0b84d10-kube-api-access-rhc6t\") pod \"glance-default-external-api-0\" (UID: \"aeea572b-1f4e-486b-a916-1b34f0b84d10\") " pod="openstack/glance-default-external-api-0" Nov 28 13:39:13 crc kubenswrapper[4631]: I1128 13:39:13.175998 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/aeea572b-1f4e-486b-a916-1b34f0b84d10-logs\") pod \"glance-default-external-api-0\" (UID: \"aeea572b-1f4e-486b-a916-1b34f0b84d10\") " pod="openstack/glance-default-external-api-0" Nov 28 13:39:13 crc kubenswrapper[4631]: I1128 13:39:13.176405 4631 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-external-api-0\" (UID: \"aeea572b-1f4e-486b-a916-1b34f0b84d10\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/glance-default-external-api-0" Nov 28 13:39:13 crc kubenswrapper[4631]: I1128 13:39:13.176624 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/aeea572b-1f4e-486b-a916-1b34f0b84d10-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"aeea572b-1f4e-486b-a916-1b34f0b84d10\") " pod="openstack/glance-default-external-api-0" Nov 28 13:39:13 crc kubenswrapper[4631]: I1128 13:39:13.182258 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/aeea572b-1f4e-486b-a916-1b34f0b84d10-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"aeea572b-1f4e-486b-a916-1b34f0b84d10\") " pod="openstack/glance-default-external-api-0" Nov 28 13:39:13 crc kubenswrapper[4631]: I1128 13:39:13.182762 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aeea572b-1f4e-486b-a916-1b34f0b84d10-scripts\") pod \"glance-default-external-api-0\" (UID: \"aeea572b-1f4e-486b-a916-1b34f0b84d10\") " pod="openstack/glance-default-external-api-0" Nov 28 13:39:13 crc kubenswrapper[4631]: I1128 13:39:13.198650 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aeea572b-1f4e-486b-a916-1b34f0b84d10-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"aeea572b-1f4e-486b-a916-1b34f0b84d10\") " pod="openstack/glance-default-external-api-0" Nov 28 13:39:13 crc kubenswrapper[4631]: I1128 13:39:13.201321 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"cb966e50-d246-48dc-b4a6-a9de4dbb8569","Type":"ContainerDied","Data":"598515cabf25ca4a2a23bcef4ffe5fa19863409e4a4da39027626c511adc8c82"} Nov 28 13:39:13 crc kubenswrapper[4631]: I1128 13:39:13.201374 4631 scope.go:117] "RemoveContainer" containerID="a869370244f2d82eb690074a62782ef88fa1793d2d6fd2b724555af086cd397f" Nov 28 13:39:13 crc kubenswrapper[4631]: I1128 13:39:13.201504 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 28 13:39:13 crc kubenswrapper[4631]: I1128 13:39:13.216231 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aeea572b-1f4e-486b-a916-1b34f0b84d10-config-data\") pod \"glance-default-external-api-0\" (UID: \"aeea572b-1f4e-486b-a916-1b34f0b84d10\") " pod="openstack/glance-default-external-api-0" Nov 28 13:39:13 crc kubenswrapper[4631]: I1128 13:39:13.231736 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rhc6t\" (UniqueName: \"kubernetes.io/projected/aeea572b-1f4e-486b-a916-1b34f0b84d10-kube-api-access-rhc6t\") pod \"glance-default-external-api-0\" (UID: \"aeea572b-1f4e-486b-a916-1b34f0b84d10\") " pod="openstack/glance-default-external-api-0" Nov 28 13:39:13 crc kubenswrapper[4631]: I1128 13:39:13.253972 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ac113fd1-daaf-4958-b779-aaee117db1ea","Type":"ContainerStarted","Data":"063cb00da8922ce657ae9b4bc1a12c502381a36f979aa22a8900b5d169e58dad"} Nov 28 13:39:13 crc kubenswrapper[4631]: I1128 13:39:13.254009 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-external-api-0\" (UID: \"aeea572b-1f4e-486b-a916-1b34f0b84d10\") " pod="openstack/glance-default-external-api-0" Nov 28 13:39:13 crc kubenswrapper[4631]: I1128 13:39:13.255836 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-68757c9bcf-d8gj4" event={"ID":"7303e352-744c-4c93-ac72-0c434384f090","Type":"ContainerStarted","Data":"bbe6a92c4a28d389bc4fcea410553a7cc348a369cf24f39e36f03d94e5b1bc3c"} Nov 28 13:39:13 crc kubenswrapper[4631]: I1128 13:39:13.340543 4631 scope.go:117] "RemoveContainer" containerID="6a153b478ff34b35b8edd10fed13a31e0f77eee95649346d1c104a57495f9a84" Nov 28 13:39:13 crc kubenswrapper[4631]: I1128 13:39:13.343539 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 28 13:39:13 crc kubenswrapper[4631]: I1128 13:39:13.402386 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 28 13:39:13 crc kubenswrapper[4631]: I1128 13:39:13.410914 4631 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 28 13:39:13 crc kubenswrapper[4631]: I1128 13:39:13.423591 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 28 13:39:13 crc kubenswrapper[4631]: I1128 13:39:13.425509 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 28 13:39:13 crc kubenswrapper[4631]: I1128 13:39:13.433743 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Nov 28 13:39:13 crc kubenswrapper[4631]: I1128 13:39:13.433945 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Nov 28 13:39:13 crc kubenswrapper[4631]: I1128 13:39:13.436745 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 28 13:39:13 crc kubenswrapper[4631]: I1128 13:39:13.490535 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b3727bc6-2743-41bf-9a94-45e4921eeb00-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"b3727bc6-2743-41bf-9a94-45e4921eeb00\") " pod="openstack/glance-default-internal-api-0" Nov 28 13:39:13 crc kubenswrapper[4631]: I1128 13:39:13.490582 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b3727bc6-2743-41bf-9a94-45e4921eeb00-logs\") pod \"glance-default-internal-api-0\" (UID: \"b3727bc6-2743-41bf-9a94-45e4921eeb00\") " pod="openstack/glance-default-internal-api-0" Nov 28 13:39:13 crc kubenswrapper[4631]: I1128 13:39:13.490760 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b3727bc6-2743-41bf-9a94-45e4921eeb00-scripts\") pod \"glance-default-internal-api-0\" (UID: \"b3727bc6-2743-41bf-9a94-45e4921eeb00\") " pod="openstack/glance-default-internal-api-0" Nov 28 13:39:13 crc kubenswrapper[4631]: I1128 13:39:13.490854 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b3727bc6-2743-41bf-9a94-45e4921eeb00-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"b3727bc6-2743-41bf-9a94-45e4921eeb00\") " pod="openstack/glance-default-internal-api-0" Nov 28 13:39:13 crc kubenswrapper[4631]: I1128 13:39:13.490934 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b3727bc6-2743-41bf-9a94-45e4921eeb00-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"b3727bc6-2743-41bf-9a94-45e4921eeb00\") " pod="openstack/glance-default-internal-api-0" Nov 28 13:39:13 crc kubenswrapper[4631]: I1128 13:39:13.491082 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-internal-api-0\" (UID: \"b3727bc6-2743-41bf-9a94-45e4921eeb00\") " pod="openstack/glance-default-internal-api-0" Nov 28 13:39:13 crc kubenswrapper[4631]: I1128 13:39:13.491111 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b3727bc6-2743-41bf-9a94-45e4921eeb00-config-data\") pod \"glance-default-internal-api-0\" (UID: \"b3727bc6-2743-41bf-9a94-45e4921eeb00\") " pod="openstack/glance-default-internal-api-0" Nov 28 13:39:13 crc kubenswrapper[4631]: I1128 13:39:13.491172 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gpjnn\" (UniqueName: \"kubernetes.io/projected/b3727bc6-2743-41bf-9a94-45e4921eeb00-kube-api-access-gpjnn\") pod \"glance-default-internal-api-0\" (UID: \"b3727bc6-2743-41bf-9a94-45e4921eeb00\") " pod="openstack/glance-default-internal-api-0" Nov 28 13:39:13 crc kubenswrapper[4631]: I1128 13:39:13.575738 4631 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bef6bb99-5262-4f84-a2f7-0908a080f3a0" path="/var/lib/kubelet/pods/bef6bb99-5262-4f84-a2f7-0908a080f3a0/volumes" Nov 28 13:39:13 crc kubenswrapper[4631]: I1128 13:39:13.577026 4631 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cb966e50-d246-48dc-b4a6-a9de4dbb8569" path="/var/lib/kubelet/pods/cb966e50-d246-48dc-b4a6-a9de4dbb8569/volumes" Nov 28 13:39:13 crc kubenswrapper[4631]: I1128 13:39:13.581966 4631 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e1d15959-31a9-464e-9e70-57547ddad347" path="/var/lib/kubelet/pods/e1d15959-31a9-464e-9e70-57547ddad347/volumes" Nov 28 13:39:13 crc kubenswrapper[4631]: I1128 13:39:13.592273 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b3727bc6-2743-41bf-9a94-45e4921eeb00-logs\") pod \"glance-default-internal-api-0\" (UID: \"b3727bc6-2743-41bf-9a94-45e4921eeb00\") " pod="openstack/glance-default-internal-api-0" Nov 28 13:39:13 crc kubenswrapper[4631]: I1128 13:39:13.592338 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b3727bc6-2743-41bf-9a94-45e4921eeb00-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"b3727bc6-2743-41bf-9a94-45e4921eeb00\") " pod="openstack/glance-default-internal-api-0" Nov 28 13:39:13 crc kubenswrapper[4631]: I1128 13:39:13.592362 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b3727bc6-2743-41bf-9a94-45e4921eeb00-scripts\") pod \"glance-default-internal-api-0\" (UID: \"b3727bc6-2743-41bf-9a94-45e4921eeb00\") " pod="openstack/glance-default-internal-api-0" Nov 28 13:39:13 crc kubenswrapper[4631]: I1128 13:39:13.592401 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b3727bc6-2743-41bf-9a94-45e4921eeb00-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"b3727bc6-2743-41bf-9a94-45e4921eeb00\") " pod="openstack/glance-default-internal-api-0" Nov 28 13:39:13 crc kubenswrapper[4631]: I1128 13:39:13.592442 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b3727bc6-2743-41bf-9a94-45e4921eeb00-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"b3727bc6-2743-41bf-9a94-45e4921eeb00\") " pod="openstack/glance-default-internal-api-0" Nov 28 13:39:13 crc kubenswrapper[4631]: I1128 13:39:13.592541 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-internal-api-0\" (UID: \"b3727bc6-2743-41bf-9a94-45e4921eeb00\") " pod="openstack/glance-default-internal-api-0" Nov 28 13:39:13 crc kubenswrapper[4631]: I1128 13:39:13.592561 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b3727bc6-2743-41bf-9a94-45e4921eeb00-config-data\") pod \"glance-default-internal-api-0\" (UID: \"b3727bc6-2743-41bf-9a94-45e4921eeb00\") " pod="openstack/glance-default-internal-api-0" Nov 28 13:39:13 crc kubenswrapper[4631]: I1128 13:39:13.592580 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gpjnn\" (UniqueName: \"kubernetes.io/projected/b3727bc6-2743-41bf-9a94-45e4921eeb00-kube-api-access-gpjnn\") pod \"glance-default-internal-api-0\" (UID: \"b3727bc6-2743-41bf-9a94-45e4921eeb00\") " pod="openstack/glance-default-internal-api-0" Nov 28 13:39:13 crc kubenswrapper[4631]: I1128 13:39:13.595913 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b3727bc6-2743-41bf-9a94-45e4921eeb00-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"b3727bc6-2743-41bf-9a94-45e4921eeb00\") " pod="openstack/glance-default-internal-api-0" Nov 28 13:39:13 crc kubenswrapper[4631]: I1128 13:39:13.600388 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b3727bc6-2743-41bf-9a94-45e4921eeb00-logs\") pod \"glance-default-internal-api-0\" (UID: \"b3727bc6-2743-41bf-9a94-45e4921eeb00\") " pod="openstack/glance-default-internal-api-0" Nov 28 13:39:13 crc kubenswrapper[4631]: I1128 13:39:13.600776 4631 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-internal-api-0\" (UID: \"b3727bc6-2743-41bf-9a94-45e4921eeb00\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/glance-default-internal-api-0" Nov 28 13:39:13 crc kubenswrapper[4631]: I1128 13:39:13.609153 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b3727bc6-2743-41bf-9a94-45e4921eeb00-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"b3727bc6-2743-41bf-9a94-45e4921eeb00\") " pod="openstack/glance-default-internal-api-0" Nov 28 13:39:13 crc kubenswrapper[4631]: I1128 13:39:13.634315 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b3727bc6-2743-41bf-9a94-45e4921eeb00-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"b3727bc6-2743-41bf-9a94-45e4921eeb00\") " pod="openstack/glance-default-internal-api-0" Nov 28 13:39:13 crc kubenswrapper[4631]: I1128 13:39:13.647947 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b3727bc6-2743-41bf-9a94-45e4921eeb00-config-data\") pod \"glance-default-internal-api-0\" (UID: \"b3727bc6-2743-41bf-9a94-45e4921eeb00\") " pod="openstack/glance-default-internal-api-0" Nov 28 13:39:13 crc kubenswrapper[4631]: I1128 13:39:13.651747 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b3727bc6-2743-41bf-9a94-45e4921eeb00-scripts\") pod \"glance-default-internal-api-0\" (UID: \"b3727bc6-2743-41bf-9a94-45e4921eeb00\") " pod="openstack/glance-default-internal-api-0" Nov 28 13:39:13 crc kubenswrapper[4631]: I1128 13:39:13.652453 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gpjnn\" (UniqueName: \"kubernetes.io/projected/b3727bc6-2743-41bf-9a94-45e4921eeb00-kube-api-access-gpjnn\") pod \"glance-default-internal-api-0\" (UID: \"b3727bc6-2743-41bf-9a94-45e4921eeb00\") " pod="openstack/glance-default-internal-api-0" Nov 28 13:39:13 crc kubenswrapper[4631]: I1128 13:39:13.711331 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-internal-api-0\" (UID: \"b3727bc6-2743-41bf-9a94-45e4921eeb00\") " pod="openstack/glance-default-internal-api-0" Nov 28 13:39:13 crc kubenswrapper[4631]: I1128 13:39:13.789645 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Nov 28 13:39:13 crc kubenswrapper[4631]: W1128 13:39:13.795875 4631 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod86f5dab6_6b28_494c_b9d6_52528c726875.slice/crio-ba4b46b1db8319f9f4efe1bdbd9ef960b89c9418ffae75f0124f0a6d16d778ef WatchSource:0}: Error finding container ba4b46b1db8319f9f4efe1bdbd9ef960b89c9418ffae75f0124f0a6d16d778ef: Status 404 returned error can't find the container with id ba4b46b1db8319f9f4efe1bdbd9ef960b89c9418ffae75f0124f0a6d16d778ef Nov 28 13:39:13 crc kubenswrapper[4631]: I1128 13:39:13.809543 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 28 13:39:14 crc kubenswrapper[4631]: I1128 13:39:14.402884 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 28 13:39:14 crc kubenswrapper[4631]: I1128 13:39:14.449420 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"86f5dab6-6b28-494c-b9d6-52528c726875","Type":"ContainerStarted","Data":"ba4b46b1db8319f9f4efe1bdbd9ef960b89c9418ffae75f0124f0a6d16d778ef"} Nov 28 13:39:14 crc kubenswrapper[4631]: I1128 13:39:14.515417 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-68757c9bcf-d8gj4" event={"ID":"7303e352-744c-4c93-ac72-0c434384f090","Type":"ContainerStarted","Data":"716f1efa174f58941d8e9e756217a5badd96d7e69340f8157f4bd397ffd0cccb"} Nov 28 13:39:14 crc kubenswrapper[4631]: I1128 13:39:14.515838 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-68757c9bcf-d8gj4" event={"ID":"7303e352-744c-4c93-ac72-0c434384f090","Type":"ContainerStarted","Data":"60a4be71dcfa25ea0f678611c62e0bbbd241cd81500fb62e2f28415e2c29a405"} Nov 28 13:39:14 crc kubenswrapper[4631]: I1128 13:39:14.516039 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-68757c9bcf-d8gj4" Nov 28 13:39:14 crc kubenswrapper[4631]: I1128 13:39:14.516155 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-68757c9bcf-d8gj4" Nov 28 13:39:14 crc kubenswrapper[4631]: I1128 13:39:14.552346 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-proxy-68757c9bcf-d8gj4" podStartSLOduration=14.552278822 podStartE2EDuration="14.552278822s" podCreationTimestamp="2025-11-28 13:39:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 13:39:14.545083349 +0000 UTC m=+1111.352386693" watchObservedRunningTime="2025-11-28 13:39:14.552278822 +0000 UTC m=+1111.359582166" Nov 28 13:39:14 crc kubenswrapper[4631]: I1128 13:39:14.738210 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 28 13:39:15 crc kubenswrapper[4631]: I1128 13:39:15.333457 4631 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-64f8fdf97b-jn5mg" podUID="111c5349-528d-4856-a3e4-1277e3b10889" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.146:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.146:8443: connect: connection refused" Nov 28 13:39:15 crc kubenswrapper[4631]: I1128 13:39:15.556986 4631 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-544496c768-rzhsw" podUID="cd607dae-568a-4b81-af81-3310c0e95854" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.147:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.147:8443: connect: connection refused" Nov 28 13:39:15 crc kubenswrapper[4631]: I1128 13:39:15.598944 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ac113fd1-daaf-4958-b779-aaee117db1ea","Type":"ContainerStarted","Data":"8fd41c24f5ebd9ec969cefedbbcb8b3f53e249e2e5a6b3d0a6e3a786e2884054"} Nov 28 13:39:15 crc kubenswrapper[4631]: I1128 13:39:15.601784 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"aeea572b-1f4e-486b-a916-1b34f0b84d10","Type":"ContainerStarted","Data":"ce90d35afc8577842fa8409429afc353a7f4d32fcc744862ad6830b46ec34dd9"} Nov 28 13:39:15 crc kubenswrapper[4631]: I1128 13:39:15.608529 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"b3727bc6-2743-41bf-9a94-45e4921eeb00","Type":"ContainerStarted","Data":"dc109835de67e9f9af39f39e3b0438ab99dbbd3865f0cd73946bc62754cf0df1"} Nov 28 13:39:15 crc kubenswrapper[4631]: I1128 13:39:15.641055 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"86f5dab6-6b28-494c-b9d6-52528c726875","Type":"ContainerStarted","Data":"d47edc098904e30e469bb75c29b690b13a6ef2ce5456a2ce6d6849f213269fd5"} Nov 28 13:39:16 crc kubenswrapper[4631]: I1128 13:39:16.679407 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"aeea572b-1f4e-486b-a916-1b34f0b84d10","Type":"ContainerStarted","Data":"19d79c71e6786344dfc50ab1c1d9bee695a485dfad32f7461a15062c391c662b"} Nov 28 13:39:16 crc kubenswrapper[4631]: I1128 13:39:16.695255 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"b3727bc6-2743-41bf-9a94-45e4921eeb00","Type":"ContainerStarted","Data":"5938095b61ae78e2339db5b05908bef66b8f0af9b0d03e4920bbe5eb6ffdbf29"} Nov 28 13:39:17 crc kubenswrapper[4631]: I1128 13:39:17.722334 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"b3727bc6-2743-41bf-9a94-45e4921eeb00","Type":"ContainerStarted","Data":"e601ea879f3af6986d7b09d6a6a1b9e19e1a71ec0d354af45993d496a7548818"} Nov 28 13:39:17 crc kubenswrapper[4631]: I1128 13:39:17.738087 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"86f5dab6-6b28-494c-b9d6-52528c726875","Type":"ContainerStarted","Data":"02de342bf873b69d04d1722727982c666eee21897f5482020c294618f390e9ef"} Nov 28 13:39:17 crc kubenswrapper[4631]: I1128 13:39:17.739205 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Nov 28 13:39:17 crc kubenswrapper[4631]: I1128 13:39:17.753860 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ac113fd1-daaf-4958-b779-aaee117db1ea","Type":"ContainerStarted","Data":"9e9842d6b641710ee7c5b462dcd1e8296f794fc8014a81cd8012c601bcf0a599"} Nov 28 13:39:17 crc kubenswrapper[4631]: I1128 13:39:17.754119 4631 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ac113fd1-daaf-4958-b779-aaee117db1ea" containerName="ceilometer-central-agent" containerID="cri-o://0053c434df76403c94f1221253f347b5af0b8d4b85b3e4d8847ca04e89fef6a6" gracePeriod=30 Nov 28 13:39:17 crc kubenswrapper[4631]: I1128 13:39:17.754224 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 28 13:39:17 crc kubenswrapper[4631]: I1128 13:39:17.754268 4631 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ac113fd1-daaf-4958-b779-aaee117db1ea" containerName="proxy-httpd" containerID="cri-o://9e9842d6b641710ee7c5b462dcd1e8296f794fc8014a81cd8012c601bcf0a599" gracePeriod=30 Nov 28 13:39:17 crc kubenswrapper[4631]: I1128 13:39:17.754340 4631 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ac113fd1-daaf-4958-b779-aaee117db1ea" containerName="sg-core" containerID="cri-o://8fd41c24f5ebd9ec969cefedbbcb8b3f53e249e2e5a6b3d0a6e3a786e2884054" gracePeriod=30 Nov 28 13:39:17 crc kubenswrapper[4631]: I1128 13:39:17.754393 4631 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ac113fd1-daaf-4958-b779-aaee117db1ea" containerName="ceilometer-notification-agent" containerID="cri-o://063cb00da8922ce657ae9b4bc1a12c502381a36f979aa22a8900b5d169e58dad" gracePeriod=30 Nov 28 13:39:17 crc kubenswrapper[4631]: I1128 13:39:17.756463 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=4.756441123 podStartE2EDuration="4.756441123s" podCreationTimestamp="2025-11-28 13:39:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 13:39:17.750388468 +0000 UTC m=+1114.557691822" watchObservedRunningTime="2025-11-28 13:39:17.756441123 +0000 UTC m=+1114.563744467" Nov 28 13:39:17 crc kubenswrapper[4631]: I1128 13:39:17.775828 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"aeea572b-1f4e-486b-a916-1b34f0b84d10","Type":"ContainerStarted","Data":"b8b9014a08843e29af2b7c46163babbf11d1efbbcb6f495f5e627a2ca5e8dfbe"} Nov 28 13:39:17 crc kubenswrapper[4631]: I1128 13:39:17.795878 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=5.795853656 podStartE2EDuration="5.795853656s" podCreationTimestamp="2025-11-28 13:39:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 13:39:17.786756699 +0000 UTC m=+1114.594060043" watchObservedRunningTime="2025-11-28 13:39:17.795853656 +0000 UTC m=+1114.603156990" Nov 28 13:39:17 crc kubenswrapper[4631]: I1128 13:39:17.830680 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=3.820351135 podStartE2EDuration="18.83064793s" podCreationTimestamp="2025-11-28 13:38:59 +0000 UTC" firstStartedPulling="2025-11-28 13:39:00.977369875 +0000 UTC m=+1097.784673219" lastFinishedPulling="2025-11-28 13:39:15.98766667 +0000 UTC m=+1112.794970014" observedRunningTime="2025-11-28 13:39:17.827799912 +0000 UTC m=+1114.635103266" watchObservedRunningTime="2025-11-28 13:39:17.83064793 +0000 UTC m=+1114.637951274" Nov 28 13:39:17 crc kubenswrapper[4631]: I1128 13:39:17.863458 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=5.863425445 podStartE2EDuration="5.863425445s" podCreationTimestamp="2025-11-28 13:39:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 13:39:17.852122004 +0000 UTC m=+1114.659425348" watchObservedRunningTime="2025-11-28 13:39:17.863425445 +0000 UTC m=+1114.670728789" Nov 28 13:39:18 crc kubenswrapper[4631]: I1128 13:39:18.812743 4631 generic.go:334] "Generic (PLEG): container finished" podID="ac113fd1-daaf-4958-b779-aaee117db1ea" containerID="9e9842d6b641710ee7c5b462dcd1e8296f794fc8014a81cd8012c601bcf0a599" exitCode=0 Nov 28 13:39:18 crc kubenswrapper[4631]: I1128 13:39:18.813136 4631 generic.go:334] "Generic (PLEG): container finished" podID="ac113fd1-daaf-4958-b779-aaee117db1ea" containerID="8fd41c24f5ebd9ec969cefedbbcb8b3f53e249e2e5a6b3d0a6e3a786e2884054" exitCode=2 Nov 28 13:39:18 crc kubenswrapper[4631]: I1128 13:39:18.813146 4631 generic.go:334] "Generic (PLEG): container finished" podID="ac113fd1-daaf-4958-b779-aaee117db1ea" containerID="063cb00da8922ce657ae9b4bc1a12c502381a36f979aa22a8900b5d169e58dad" exitCode=0 Nov 28 13:39:18 crc kubenswrapper[4631]: I1128 13:39:18.813154 4631 generic.go:334] "Generic (PLEG): container finished" podID="ac113fd1-daaf-4958-b779-aaee117db1ea" containerID="0053c434df76403c94f1221253f347b5af0b8d4b85b3e4d8847ca04e89fef6a6" exitCode=0 Nov 28 13:39:18 crc kubenswrapper[4631]: I1128 13:39:18.813366 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ac113fd1-daaf-4958-b779-aaee117db1ea","Type":"ContainerDied","Data":"9e9842d6b641710ee7c5b462dcd1e8296f794fc8014a81cd8012c601bcf0a599"} Nov 28 13:39:18 crc kubenswrapper[4631]: I1128 13:39:18.813445 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ac113fd1-daaf-4958-b779-aaee117db1ea","Type":"ContainerDied","Data":"8fd41c24f5ebd9ec969cefedbbcb8b3f53e249e2e5a6b3d0a6e3a786e2884054"} Nov 28 13:39:18 crc kubenswrapper[4631]: I1128 13:39:18.813457 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ac113fd1-daaf-4958-b779-aaee117db1ea","Type":"ContainerDied","Data":"063cb00da8922ce657ae9b4bc1a12c502381a36f979aa22a8900b5d169e58dad"} Nov 28 13:39:18 crc kubenswrapper[4631]: I1128 13:39:18.813470 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ac113fd1-daaf-4958-b779-aaee117db1ea","Type":"ContainerDied","Data":"0053c434df76403c94f1221253f347b5af0b8d4b85b3e4d8847ca04e89fef6a6"} Nov 28 13:39:18 crc kubenswrapper[4631]: I1128 13:39:18.993244 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 28 13:39:19 crc kubenswrapper[4631]: I1128 13:39:19.033944 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ac113fd1-daaf-4958-b779-aaee117db1ea-run-httpd\") pod \"ac113fd1-daaf-4958-b779-aaee117db1ea\" (UID: \"ac113fd1-daaf-4958-b779-aaee117db1ea\") " Nov 28 13:39:19 crc kubenswrapper[4631]: I1128 13:39:19.034086 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ac113fd1-daaf-4958-b779-aaee117db1ea-log-httpd\") pod \"ac113fd1-daaf-4958-b779-aaee117db1ea\" (UID: \"ac113fd1-daaf-4958-b779-aaee117db1ea\") " Nov 28 13:39:19 crc kubenswrapper[4631]: I1128 13:39:19.034190 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ac113fd1-daaf-4958-b779-aaee117db1ea-sg-core-conf-yaml\") pod \"ac113fd1-daaf-4958-b779-aaee117db1ea\" (UID: \"ac113fd1-daaf-4958-b779-aaee117db1ea\") " Nov 28 13:39:19 crc kubenswrapper[4631]: I1128 13:39:19.034310 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ac113fd1-daaf-4958-b779-aaee117db1ea-config-data\") pod \"ac113fd1-daaf-4958-b779-aaee117db1ea\" (UID: \"ac113fd1-daaf-4958-b779-aaee117db1ea\") " Nov 28 13:39:19 crc kubenswrapper[4631]: I1128 13:39:19.034426 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ac113fd1-daaf-4958-b779-aaee117db1ea-scripts\") pod \"ac113fd1-daaf-4958-b779-aaee117db1ea\" (UID: \"ac113fd1-daaf-4958-b779-aaee117db1ea\") " Nov 28 13:39:19 crc kubenswrapper[4631]: I1128 13:39:19.034466 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ac113fd1-daaf-4958-b779-aaee117db1ea-combined-ca-bundle\") pod \"ac113fd1-daaf-4958-b779-aaee117db1ea\" (UID: \"ac113fd1-daaf-4958-b779-aaee117db1ea\") " Nov 28 13:39:19 crc kubenswrapper[4631]: I1128 13:39:19.034520 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-npcz5\" (UniqueName: \"kubernetes.io/projected/ac113fd1-daaf-4958-b779-aaee117db1ea-kube-api-access-npcz5\") pod \"ac113fd1-daaf-4958-b779-aaee117db1ea\" (UID: \"ac113fd1-daaf-4958-b779-aaee117db1ea\") " Nov 28 13:39:19 crc kubenswrapper[4631]: I1128 13:39:19.035668 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ac113fd1-daaf-4958-b779-aaee117db1ea-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "ac113fd1-daaf-4958-b779-aaee117db1ea" (UID: "ac113fd1-daaf-4958-b779-aaee117db1ea"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 13:39:19 crc kubenswrapper[4631]: I1128 13:39:19.039678 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ac113fd1-daaf-4958-b779-aaee117db1ea-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "ac113fd1-daaf-4958-b779-aaee117db1ea" (UID: "ac113fd1-daaf-4958-b779-aaee117db1ea"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 13:39:19 crc kubenswrapper[4631]: I1128 13:39:19.055501 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ac113fd1-daaf-4958-b779-aaee117db1ea-scripts" (OuterVolumeSpecName: "scripts") pod "ac113fd1-daaf-4958-b779-aaee117db1ea" (UID: "ac113fd1-daaf-4958-b779-aaee117db1ea"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:39:19 crc kubenswrapper[4631]: I1128 13:39:19.079849 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ac113fd1-daaf-4958-b779-aaee117db1ea-kube-api-access-npcz5" (OuterVolumeSpecName: "kube-api-access-npcz5") pod "ac113fd1-daaf-4958-b779-aaee117db1ea" (UID: "ac113fd1-daaf-4958-b779-aaee117db1ea"). InnerVolumeSpecName "kube-api-access-npcz5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:39:19 crc kubenswrapper[4631]: I1128 13:39:19.095915 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ac113fd1-daaf-4958-b779-aaee117db1ea-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "ac113fd1-daaf-4958-b779-aaee117db1ea" (UID: "ac113fd1-daaf-4958-b779-aaee117db1ea"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:39:19 crc kubenswrapper[4631]: I1128 13:39:19.137763 4631 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ac113fd1-daaf-4958-b779-aaee117db1ea-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 13:39:19 crc kubenswrapper[4631]: I1128 13:39:19.137803 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-npcz5\" (UniqueName: \"kubernetes.io/projected/ac113fd1-daaf-4958-b779-aaee117db1ea-kube-api-access-npcz5\") on node \"crc\" DevicePath \"\"" Nov 28 13:39:19 crc kubenswrapper[4631]: I1128 13:39:19.137817 4631 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ac113fd1-daaf-4958-b779-aaee117db1ea-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 28 13:39:19 crc kubenswrapper[4631]: I1128 13:39:19.137831 4631 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ac113fd1-daaf-4958-b779-aaee117db1ea-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 28 13:39:19 crc kubenswrapper[4631]: I1128 13:39:19.137847 4631 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ac113fd1-daaf-4958-b779-aaee117db1ea-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 28 13:39:19 crc kubenswrapper[4631]: I1128 13:39:19.253204 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ac113fd1-daaf-4958-b779-aaee117db1ea-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ac113fd1-daaf-4958-b779-aaee117db1ea" (UID: "ac113fd1-daaf-4958-b779-aaee117db1ea"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:39:19 crc kubenswrapper[4631]: I1128 13:39:19.254344 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ac113fd1-daaf-4958-b779-aaee117db1ea-config-data" (OuterVolumeSpecName: "config-data") pod "ac113fd1-daaf-4958-b779-aaee117db1ea" (UID: "ac113fd1-daaf-4958-b779-aaee117db1ea"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:39:19 crc kubenswrapper[4631]: I1128 13:39:19.341884 4631 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ac113fd1-daaf-4958-b779-aaee117db1ea-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 13:39:19 crc kubenswrapper[4631]: I1128 13:39:19.341932 4631 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ac113fd1-daaf-4958-b779-aaee117db1ea-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 13:39:19 crc kubenswrapper[4631]: I1128 13:39:19.827535 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 28 13:39:19 crc kubenswrapper[4631]: I1128 13:39:19.827510 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ac113fd1-daaf-4958-b779-aaee117db1ea","Type":"ContainerDied","Data":"00ecb9ef637423c12a9a4d2165068b41e5935cba5464a65193c60c523a342172"} Nov 28 13:39:19 crc kubenswrapper[4631]: I1128 13:39:19.829105 4631 scope.go:117] "RemoveContainer" containerID="9e9842d6b641710ee7c5b462dcd1e8296f794fc8014a81cd8012c601bcf0a599" Nov 28 13:39:19 crc kubenswrapper[4631]: I1128 13:39:19.865019 4631 scope.go:117] "RemoveContainer" containerID="8fd41c24f5ebd9ec969cefedbbcb8b3f53e249e2e5a6b3d0a6e3a786e2884054" Nov 28 13:39:19 crc kubenswrapper[4631]: I1128 13:39:19.871517 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 28 13:39:19 crc kubenswrapper[4631]: I1128 13:39:19.916841 4631 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 28 13:39:19 crc kubenswrapper[4631]: I1128 13:39:19.918886 4631 scope.go:117] "RemoveContainer" containerID="063cb00da8922ce657ae9b4bc1a12c502381a36f979aa22a8900b5d169e58dad" Nov 28 13:39:19 crc kubenswrapper[4631]: I1128 13:39:19.928720 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 28 13:39:19 crc kubenswrapper[4631]: E1128 13:39:19.929406 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ac113fd1-daaf-4958-b779-aaee117db1ea" containerName="ceilometer-central-agent" Nov 28 13:39:19 crc kubenswrapper[4631]: I1128 13:39:19.929429 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="ac113fd1-daaf-4958-b779-aaee117db1ea" containerName="ceilometer-central-agent" Nov 28 13:39:19 crc kubenswrapper[4631]: E1128 13:39:19.929450 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ac113fd1-daaf-4958-b779-aaee117db1ea" containerName="proxy-httpd" Nov 28 13:39:19 crc kubenswrapper[4631]: I1128 13:39:19.929458 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="ac113fd1-daaf-4958-b779-aaee117db1ea" containerName="proxy-httpd" Nov 28 13:39:19 crc kubenswrapper[4631]: E1128 13:39:19.929477 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ac113fd1-daaf-4958-b779-aaee117db1ea" containerName="ceilometer-notification-agent" Nov 28 13:39:19 crc kubenswrapper[4631]: I1128 13:39:19.929485 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="ac113fd1-daaf-4958-b779-aaee117db1ea" containerName="ceilometer-notification-agent" Nov 28 13:39:19 crc kubenswrapper[4631]: E1128 13:39:19.929505 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ac113fd1-daaf-4958-b779-aaee117db1ea" containerName="sg-core" Nov 28 13:39:19 crc kubenswrapper[4631]: I1128 13:39:19.929513 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="ac113fd1-daaf-4958-b779-aaee117db1ea" containerName="sg-core" Nov 28 13:39:19 crc kubenswrapper[4631]: I1128 13:39:19.929953 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="ac113fd1-daaf-4958-b779-aaee117db1ea" containerName="ceilometer-central-agent" Nov 28 13:39:19 crc kubenswrapper[4631]: I1128 13:39:19.929986 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="ac113fd1-daaf-4958-b779-aaee117db1ea" containerName="sg-core" Nov 28 13:39:19 crc kubenswrapper[4631]: I1128 13:39:19.930004 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="ac113fd1-daaf-4958-b779-aaee117db1ea" containerName="proxy-httpd" Nov 28 13:39:19 crc kubenswrapper[4631]: I1128 13:39:19.930012 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="ac113fd1-daaf-4958-b779-aaee117db1ea" containerName="ceilometer-notification-agent" Nov 28 13:39:19 crc kubenswrapper[4631]: I1128 13:39:19.933792 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 28 13:39:19 crc kubenswrapper[4631]: I1128 13:39:19.937367 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 28 13:39:19 crc kubenswrapper[4631]: I1128 13:39:19.937705 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 28 13:39:19 crc kubenswrapper[4631]: I1128 13:39:19.949446 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 28 13:39:19 crc kubenswrapper[4631]: I1128 13:39:19.956163 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/da6d90be-70f1-42d8-9b95-a338d635a460-log-httpd\") pod \"ceilometer-0\" (UID: \"da6d90be-70f1-42d8-9b95-a338d635a460\") " pod="openstack/ceilometer-0" Nov 28 13:39:19 crc kubenswrapper[4631]: I1128 13:39:19.956671 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/da6d90be-70f1-42d8-9b95-a338d635a460-run-httpd\") pod \"ceilometer-0\" (UID: \"da6d90be-70f1-42d8-9b95-a338d635a460\") " pod="openstack/ceilometer-0" Nov 28 13:39:19 crc kubenswrapper[4631]: I1128 13:39:19.956817 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/da6d90be-70f1-42d8-9b95-a338d635a460-config-data\") pod \"ceilometer-0\" (UID: \"da6d90be-70f1-42d8-9b95-a338d635a460\") " pod="openstack/ceilometer-0" Nov 28 13:39:19 crc kubenswrapper[4631]: I1128 13:39:19.956863 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/da6d90be-70f1-42d8-9b95-a338d635a460-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"da6d90be-70f1-42d8-9b95-a338d635a460\") " pod="openstack/ceilometer-0" Nov 28 13:39:19 crc kubenswrapper[4631]: I1128 13:39:19.956900 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/da6d90be-70f1-42d8-9b95-a338d635a460-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"da6d90be-70f1-42d8-9b95-a338d635a460\") " pod="openstack/ceilometer-0" Nov 28 13:39:19 crc kubenswrapper[4631]: I1128 13:39:19.956958 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/da6d90be-70f1-42d8-9b95-a338d635a460-scripts\") pod \"ceilometer-0\" (UID: \"da6d90be-70f1-42d8-9b95-a338d635a460\") " pod="openstack/ceilometer-0" Nov 28 13:39:19 crc kubenswrapper[4631]: I1128 13:39:19.957049 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kqpjv\" (UniqueName: \"kubernetes.io/projected/da6d90be-70f1-42d8-9b95-a338d635a460-kube-api-access-kqpjv\") pod \"ceilometer-0\" (UID: \"da6d90be-70f1-42d8-9b95-a338d635a460\") " pod="openstack/ceilometer-0" Nov 28 13:39:20 crc kubenswrapper[4631]: I1128 13:39:20.030814 4631 scope.go:117] "RemoveContainer" containerID="0053c434df76403c94f1221253f347b5af0b8d4b85b3e4d8847ca04e89fef6a6" Nov 28 13:39:20 crc kubenswrapper[4631]: I1128 13:39:20.597819 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/da6d90be-70f1-42d8-9b95-a338d635a460-config-data\") pod \"ceilometer-0\" (UID: \"da6d90be-70f1-42d8-9b95-a338d635a460\") " pod="openstack/ceilometer-0" Nov 28 13:39:20 crc kubenswrapper[4631]: I1128 13:39:20.597887 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/da6d90be-70f1-42d8-9b95-a338d635a460-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"da6d90be-70f1-42d8-9b95-a338d635a460\") " pod="openstack/ceilometer-0" Nov 28 13:39:20 crc kubenswrapper[4631]: I1128 13:39:20.597919 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/da6d90be-70f1-42d8-9b95-a338d635a460-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"da6d90be-70f1-42d8-9b95-a338d635a460\") " pod="openstack/ceilometer-0" Nov 28 13:39:20 crc kubenswrapper[4631]: I1128 13:39:20.597960 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/da6d90be-70f1-42d8-9b95-a338d635a460-scripts\") pod \"ceilometer-0\" (UID: \"da6d90be-70f1-42d8-9b95-a338d635a460\") " pod="openstack/ceilometer-0" Nov 28 13:39:20 crc kubenswrapper[4631]: I1128 13:39:20.598025 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kqpjv\" (UniqueName: \"kubernetes.io/projected/da6d90be-70f1-42d8-9b95-a338d635a460-kube-api-access-kqpjv\") pod \"ceilometer-0\" (UID: \"da6d90be-70f1-42d8-9b95-a338d635a460\") " pod="openstack/ceilometer-0" Nov 28 13:39:20 crc kubenswrapper[4631]: I1128 13:39:20.598050 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/da6d90be-70f1-42d8-9b95-a338d635a460-log-httpd\") pod \"ceilometer-0\" (UID: \"da6d90be-70f1-42d8-9b95-a338d635a460\") " pod="openstack/ceilometer-0" Nov 28 13:39:20 crc kubenswrapper[4631]: I1128 13:39:20.598069 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/da6d90be-70f1-42d8-9b95-a338d635a460-run-httpd\") pod \"ceilometer-0\" (UID: \"da6d90be-70f1-42d8-9b95-a338d635a460\") " pod="openstack/ceilometer-0" Nov 28 13:39:20 crc kubenswrapper[4631]: I1128 13:39:20.598784 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/da6d90be-70f1-42d8-9b95-a338d635a460-run-httpd\") pod \"ceilometer-0\" (UID: \"da6d90be-70f1-42d8-9b95-a338d635a460\") " pod="openstack/ceilometer-0" Nov 28 13:39:20 crc kubenswrapper[4631]: I1128 13:39:20.599622 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/da6d90be-70f1-42d8-9b95-a338d635a460-log-httpd\") pod \"ceilometer-0\" (UID: \"da6d90be-70f1-42d8-9b95-a338d635a460\") " pod="openstack/ceilometer-0" Nov 28 13:39:20 crc kubenswrapper[4631]: I1128 13:39:20.611161 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/da6d90be-70f1-42d8-9b95-a338d635a460-scripts\") pod \"ceilometer-0\" (UID: \"da6d90be-70f1-42d8-9b95-a338d635a460\") " pod="openstack/ceilometer-0" Nov 28 13:39:20 crc kubenswrapper[4631]: I1128 13:39:20.613244 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/da6d90be-70f1-42d8-9b95-a338d635a460-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"da6d90be-70f1-42d8-9b95-a338d635a460\") " pod="openstack/ceilometer-0" Nov 28 13:39:20 crc kubenswrapper[4631]: I1128 13:39:20.620761 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/da6d90be-70f1-42d8-9b95-a338d635a460-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"da6d90be-70f1-42d8-9b95-a338d635a460\") " pod="openstack/ceilometer-0" Nov 28 13:39:20 crc kubenswrapper[4631]: I1128 13:39:20.624991 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/da6d90be-70f1-42d8-9b95-a338d635a460-config-data\") pod \"ceilometer-0\" (UID: \"da6d90be-70f1-42d8-9b95-a338d635a460\") " pod="openstack/ceilometer-0" Nov 28 13:39:20 crc kubenswrapper[4631]: I1128 13:39:20.641446 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kqpjv\" (UniqueName: \"kubernetes.io/projected/da6d90be-70f1-42d8-9b95-a338d635a460-kube-api-access-kqpjv\") pod \"ceilometer-0\" (UID: \"da6d90be-70f1-42d8-9b95-a338d635a460\") " pod="openstack/ceilometer-0" Nov 28 13:39:20 crc kubenswrapper[4631]: I1128 13:39:20.927898 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 28 13:39:21 crc kubenswrapper[4631]: I1128 13:39:21.400339 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-68757c9bcf-d8gj4" Nov 28 13:39:21 crc kubenswrapper[4631]: I1128 13:39:21.410372 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-68757c9bcf-d8gj4" Nov 28 13:39:21 crc kubenswrapper[4631]: I1128 13:39:21.511730 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 28 13:39:21 crc kubenswrapper[4631]: I1128 13:39:21.550730 4631 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ac113fd1-daaf-4958-b779-aaee117db1ea" path="/var/lib/kubelet/pods/ac113fd1-daaf-4958-b779-aaee117db1ea/volumes" Nov 28 13:39:21 crc kubenswrapper[4631]: I1128 13:39:21.856161 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"da6d90be-70f1-42d8-9b95-a338d635a460","Type":"ContainerStarted","Data":"ddfae1373058f71a2b60260b97703d1bac3a7d2ad5ad4f39b7827262b995d16b"} Nov 28 13:39:22 crc kubenswrapper[4631]: I1128 13:39:22.442739 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 28 13:39:22 crc kubenswrapper[4631]: I1128 13:39:22.874914 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"da6d90be-70f1-42d8-9b95-a338d635a460","Type":"ContainerStarted","Data":"fb69638e625d934534f92ebfbe0519166ed42fb448f2dda43bf3db42c019a610"} Nov 28 13:39:23 crc kubenswrapper[4631]: I1128 13:39:23.344032 4631 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Nov 28 13:39:23 crc kubenswrapper[4631]: I1128 13:39:23.344624 4631 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Nov 28 13:39:23 crc kubenswrapper[4631]: I1128 13:39:23.395799 4631 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Nov 28 13:39:23 crc kubenswrapper[4631]: I1128 13:39:23.432991 4631 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Nov 28 13:39:23 crc kubenswrapper[4631]: I1128 13:39:23.810516 4631 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Nov 28 13:39:23 crc kubenswrapper[4631]: I1128 13:39:23.810581 4631 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Nov 28 13:39:23 crc kubenswrapper[4631]: I1128 13:39:23.865961 4631 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Nov 28 13:39:23 crc kubenswrapper[4631]: I1128 13:39:23.890783 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"da6d90be-70f1-42d8-9b95-a338d635a460","Type":"ContainerStarted","Data":"11f9d2b9a8e6191a9bcfbb12937052b29700f5e513889b5e941883d14b1e060e"} Nov 28 13:39:23 crc kubenswrapper[4631]: I1128 13:39:23.891463 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Nov 28 13:39:23 crc kubenswrapper[4631]: I1128 13:39:23.891540 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Nov 28 13:39:23 crc kubenswrapper[4631]: I1128 13:39:23.891555 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Nov 28 13:39:23 crc kubenswrapper[4631]: I1128 13:39:23.912860 4631 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Nov 28 13:39:24 crc kubenswrapper[4631]: I1128 13:39:24.905394 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"da6d90be-70f1-42d8-9b95-a338d635a460","Type":"ContainerStarted","Data":"18f1eeaaf309fee86444a43afd8cd2087c028118589bb2e14b693df9349d4714"} Nov 28 13:39:24 crc kubenswrapper[4631]: I1128 13:39:24.907240 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Nov 28 13:39:25 crc kubenswrapper[4631]: I1128 13:39:25.332402 4631 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-64f8fdf97b-jn5mg" podUID="111c5349-528d-4856-a3e4-1277e3b10889" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.146:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.146:8443: connect: connection refused" Nov 28 13:39:25 crc kubenswrapper[4631]: I1128 13:39:25.332487 4631 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-64f8fdf97b-jn5mg" Nov 28 13:39:25 crc kubenswrapper[4631]: I1128 13:39:25.333305 4631 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="horizon" containerStatusID={"Type":"cri-o","ID":"ddca46c54a9ba33270a5c14213a03ffc4ee751ec03afaccc3b37cfa0eac87eb7"} pod="openstack/horizon-64f8fdf97b-jn5mg" containerMessage="Container horizon failed startup probe, will be restarted" Nov 28 13:39:25 crc kubenswrapper[4631]: I1128 13:39:25.333351 4631 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-64f8fdf97b-jn5mg" podUID="111c5349-528d-4856-a3e4-1277e3b10889" containerName="horizon" containerID="cri-o://ddca46c54a9ba33270a5c14213a03ffc4ee751ec03afaccc3b37cfa0eac87eb7" gracePeriod=30 Nov 28 13:39:25 crc kubenswrapper[4631]: I1128 13:39:25.554459 4631 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-544496c768-rzhsw" podUID="cd607dae-568a-4b81-af81-3310c0e95854" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.147:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.147:8443: connect: connection refused" Nov 28 13:39:25 crc kubenswrapper[4631]: I1128 13:39:25.554937 4631 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-544496c768-rzhsw" Nov 28 13:39:25 crc kubenswrapper[4631]: I1128 13:39:25.555913 4631 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="horizon" containerStatusID={"Type":"cri-o","ID":"26b4a5d749292bcec07819e5de34fe07da049e5be75efa94d19e13fdfcfb064d"} pod="openstack/horizon-544496c768-rzhsw" containerMessage="Container horizon failed startup probe, will be restarted" Nov 28 13:39:25 crc kubenswrapper[4631]: I1128 13:39:25.555957 4631 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-544496c768-rzhsw" podUID="cd607dae-568a-4b81-af81-3310c0e95854" containerName="horizon" containerID="cri-o://26b4a5d749292bcec07819e5de34fe07da049e5be75efa94d19e13fdfcfb064d" gracePeriod=30 Nov 28 13:39:25 crc kubenswrapper[4631]: I1128 13:39:25.923908 4631 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 28 13:39:25 crc kubenswrapper[4631]: I1128 13:39:25.924049 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"da6d90be-70f1-42d8-9b95-a338d635a460","Type":"ContainerStarted","Data":"82b1ce68db6d83ee312c79d1f5eb6eae79e2cd43dc3c03e78e2d29df0f446585"} Nov 28 13:39:25 crc kubenswrapper[4631]: I1128 13:39:25.924256 4631 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="da6d90be-70f1-42d8-9b95-a338d635a460" containerName="ceilometer-central-agent" containerID="cri-o://fb69638e625d934534f92ebfbe0519166ed42fb448f2dda43bf3db42c019a610" gracePeriod=30 Nov 28 13:39:25 crc kubenswrapper[4631]: I1128 13:39:25.925104 4631 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="da6d90be-70f1-42d8-9b95-a338d635a460" containerName="proxy-httpd" containerID="cri-o://82b1ce68db6d83ee312c79d1f5eb6eae79e2cd43dc3c03e78e2d29df0f446585" gracePeriod=30 Nov 28 13:39:25 crc kubenswrapper[4631]: I1128 13:39:25.925173 4631 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="da6d90be-70f1-42d8-9b95-a338d635a460" containerName="sg-core" containerID="cri-o://18f1eeaaf309fee86444a43afd8cd2087c028118589bb2e14b693df9349d4714" gracePeriod=30 Nov 28 13:39:25 crc kubenswrapper[4631]: I1128 13:39:25.925227 4631 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="da6d90be-70f1-42d8-9b95-a338d635a460" containerName="ceilometer-notification-agent" containerID="cri-o://11f9d2b9a8e6191a9bcfbb12937052b29700f5e513889b5e941883d14b1e060e" gracePeriod=30 Nov 28 13:39:25 crc kubenswrapper[4631]: I1128 13:39:25.974378 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=3.285534528 podStartE2EDuration="6.974351826s" podCreationTimestamp="2025-11-28 13:39:19 +0000 UTC" firstStartedPulling="2025-11-28 13:39:21.561820254 +0000 UTC m=+1118.369123598" lastFinishedPulling="2025-11-28 13:39:25.250637552 +0000 UTC m=+1122.057940896" observedRunningTime="2025-11-28 13:39:25.967889921 +0000 UTC m=+1122.775193265" watchObservedRunningTime="2025-11-28 13:39:25.974351826 +0000 UTC m=+1122.781655170" Nov 28 13:39:26 crc kubenswrapper[4631]: I1128 13:39:26.784265 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Nov 28 13:39:26 crc kubenswrapper[4631]: I1128 13:39:26.943067 4631 generic.go:334] "Generic (PLEG): container finished" podID="da6d90be-70f1-42d8-9b95-a338d635a460" containerID="82b1ce68db6d83ee312c79d1f5eb6eae79e2cd43dc3c03e78e2d29df0f446585" exitCode=0 Nov 28 13:39:26 crc kubenswrapper[4631]: I1128 13:39:26.943105 4631 generic.go:334] "Generic (PLEG): container finished" podID="da6d90be-70f1-42d8-9b95-a338d635a460" containerID="18f1eeaaf309fee86444a43afd8cd2087c028118589bb2e14b693df9349d4714" exitCode=2 Nov 28 13:39:26 crc kubenswrapper[4631]: I1128 13:39:26.943114 4631 generic.go:334] "Generic (PLEG): container finished" podID="da6d90be-70f1-42d8-9b95-a338d635a460" containerID="11f9d2b9a8e6191a9bcfbb12937052b29700f5e513889b5e941883d14b1e060e" exitCode=0 Nov 28 13:39:26 crc kubenswrapper[4631]: I1128 13:39:26.943138 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"da6d90be-70f1-42d8-9b95-a338d635a460","Type":"ContainerDied","Data":"82b1ce68db6d83ee312c79d1f5eb6eae79e2cd43dc3c03e78e2d29df0f446585"} Nov 28 13:39:26 crc kubenswrapper[4631]: I1128 13:39:26.943173 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"da6d90be-70f1-42d8-9b95-a338d635a460","Type":"ContainerDied","Data":"18f1eeaaf309fee86444a43afd8cd2087c028118589bb2e14b693df9349d4714"} Nov 28 13:39:26 crc kubenswrapper[4631]: I1128 13:39:26.943183 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"da6d90be-70f1-42d8-9b95-a338d635a460","Type":"ContainerDied","Data":"11f9d2b9a8e6191a9bcfbb12937052b29700f5e513889b5e941883d14b1e060e"} Nov 28 13:39:28 crc kubenswrapper[4631]: I1128 13:39:28.519780 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Nov 28 13:39:28 crc kubenswrapper[4631]: I1128 13:39:28.520345 4631 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 28 13:39:28 crc kubenswrapper[4631]: I1128 13:39:28.523619 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Nov 28 13:39:28 crc kubenswrapper[4631]: I1128 13:39:28.914516 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Nov 28 13:39:28 crc kubenswrapper[4631]: I1128 13:39:28.914696 4631 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 28 13:39:28 crc kubenswrapper[4631]: I1128 13:39:28.968785 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"ea063a3c-330a-4642-a815-c013ae6b80c7","Type":"ContainerStarted","Data":"26893e935624d9c36a8de5f9296d81d05043202047c2795efcaded7e0a10b1e0"} Nov 28 13:39:29 crc kubenswrapper[4631]: I1128 13:39:29.653442 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Nov 28 13:39:29 crc kubenswrapper[4631]: I1128 13:39:29.713303 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=2.637605434 podStartE2EDuration="36.713264315s" podCreationTimestamp="2025-11-28 13:38:53 +0000 UTC" firstStartedPulling="2025-11-28 13:38:54.034419517 +0000 UTC m=+1090.841722861" lastFinishedPulling="2025-11-28 13:39:28.110078398 +0000 UTC m=+1124.917381742" observedRunningTime="2025-11-28 13:39:29.026397214 +0000 UTC m=+1125.833700568" watchObservedRunningTime="2025-11-28 13:39:29.713264315 +0000 UTC m=+1126.520567659" Nov 28 13:39:31 crc kubenswrapper[4631]: I1128 13:39:31.011662 4631 generic.go:334] "Generic (PLEG): container finished" podID="da6d90be-70f1-42d8-9b95-a338d635a460" containerID="fb69638e625d934534f92ebfbe0519166ed42fb448f2dda43bf3db42c019a610" exitCode=0 Nov 28 13:39:31 crc kubenswrapper[4631]: I1128 13:39:31.011757 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"da6d90be-70f1-42d8-9b95-a338d635a460","Type":"ContainerDied","Data":"fb69638e625d934534f92ebfbe0519166ed42fb448f2dda43bf3db42c019a610"} Nov 28 13:39:33 crc kubenswrapper[4631]: I1128 13:39:33.521420 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 28 13:39:33 crc kubenswrapper[4631]: I1128 13:39:33.592101 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/da6d90be-70f1-42d8-9b95-a338d635a460-config-data\") pod \"da6d90be-70f1-42d8-9b95-a338d635a460\" (UID: \"da6d90be-70f1-42d8-9b95-a338d635a460\") " Nov 28 13:39:33 crc kubenswrapper[4631]: I1128 13:39:33.592255 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/da6d90be-70f1-42d8-9b95-a338d635a460-log-httpd\") pod \"da6d90be-70f1-42d8-9b95-a338d635a460\" (UID: \"da6d90be-70f1-42d8-9b95-a338d635a460\") " Nov 28 13:39:33 crc kubenswrapper[4631]: I1128 13:39:33.592314 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/da6d90be-70f1-42d8-9b95-a338d635a460-combined-ca-bundle\") pod \"da6d90be-70f1-42d8-9b95-a338d635a460\" (UID: \"da6d90be-70f1-42d8-9b95-a338d635a460\") " Nov 28 13:39:33 crc kubenswrapper[4631]: I1128 13:39:33.592335 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/da6d90be-70f1-42d8-9b95-a338d635a460-run-httpd\") pod \"da6d90be-70f1-42d8-9b95-a338d635a460\" (UID: \"da6d90be-70f1-42d8-9b95-a338d635a460\") " Nov 28 13:39:33 crc kubenswrapper[4631]: I1128 13:39:33.592436 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/da6d90be-70f1-42d8-9b95-a338d635a460-sg-core-conf-yaml\") pod \"da6d90be-70f1-42d8-9b95-a338d635a460\" (UID: \"da6d90be-70f1-42d8-9b95-a338d635a460\") " Nov 28 13:39:33 crc kubenswrapper[4631]: I1128 13:39:33.592456 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/da6d90be-70f1-42d8-9b95-a338d635a460-scripts\") pod \"da6d90be-70f1-42d8-9b95-a338d635a460\" (UID: \"da6d90be-70f1-42d8-9b95-a338d635a460\") " Nov 28 13:39:33 crc kubenswrapper[4631]: I1128 13:39:33.592481 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kqpjv\" (UniqueName: \"kubernetes.io/projected/da6d90be-70f1-42d8-9b95-a338d635a460-kube-api-access-kqpjv\") pod \"da6d90be-70f1-42d8-9b95-a338d635a460\" (UID: \"da6d90be-70f1-42d8-9b95-a338d635a460\") " Nov 28 13:39:33 crc kubenswrapper[4631]: I1128 13:39:33.593924 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/da6d90be-70f1-42d8-9b95-a338d635a460-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "da6d90be-70f1-42d8-9b95-a338d635a460" (UID: "da6d90be-70f1-42d8-9b95-a338d635a460"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 13:39:33 crc kubenswrapper[4631]: I1128 13:39:33.594566 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/da6d90be-70f1-42d8-9b95-a338d635a460-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "da6d90be-70f1-42d8-9b95-a338d635a460" (UID: "da6d90be-70f1-42d8-9b95-a338d635a460"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 13:39:33 crc kubenswrapper[4631]: I1128 13:39:33.605739 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/da6d90be-70f1-42d8-9b95-a338d635a460-kube-api-access-kqpjv" (OuterVolumeSpecName: "kube-api-access-kqpjv") pod "da6d90be-70f1-42d8-9b95-a338d635a460" (UID: "da6d90be-70f1-42d8-9b95-a338d635a460"). InnerVolumeSpecName "kube-api-access-kqpjv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:39:33 crc kubenswrapper[4631]: I1128 13:39:33.606252 4631 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/da6d90be-70f1-42d8-9b95-a338d635a460-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 28 13:39:33 crc kubenswrapper[4631]: I1128 13:39:33.616011 4631 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/da6d90be-70f1-42d8-9b95-a338d635a460-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 28 13:39:33 crc kubenswrapper[4631]: I1128 13:39:33.634572 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/da6d90be-70f1-42d8-9b95-a338d635a460-scripts" (OuterVolumeSpecName: "scripts") pod "da6d90be-70f1-42d8-9b95-a338d635a460" (UID: "da6d90be-70f1-42d8-9b95-a338d635a460"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:39:33 crc kubenswrapper[4631]: I1128 13:39:33.680687 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/da6d90be-70f1-42d8-9b95-a338d635a460-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "da6d90be-70f1-42d8-9b95-a338d635a460" (UID: "da6d90be-70f1-42d8-9b95-a338d635a460"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:39:33 crc kubenswrapper[4631]: I1128 13:39:33.718158 4631 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/da6d90be-70f1-42d8-9b95-a338d635a460-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 28 13:39:33 crc kubenswrapper[4631]: I1128 13:39:33.718203 4631 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/da6d90be-70f1-42d8-9b95-a338d635a460-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 13:39:33 crc kubenswrapper[4631]: I1128 13:39:33.718216 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kqpjv\" (UniqueName: \"kubernetes.io/projected/da6d90be-70f1-42d8-9b95-a338d635a460-kube-api-access-kqpjv\") on node \"crc\" DevicePath \"\"" Nov 28 13:39:33 crc kubenswrapper[4631]: I1128 13:39:33.734505 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/da6d90be-70f1-42d8-9b95-a338d635a460-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "da6d90be-70f1-42d8-9b95-a338d635a460" (UID: "da6d90be-70f1-42d8-9b95-a338d635a460"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:39:33 crc kubenswrapper[4631]: I1128 13:39:33.771259 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/da6d90be-70f1-42d8-9b95-a338d635a460-config-data" (OuterVolumeSpecName: "config-data") pod "da6d90be-70f1-42d8-9b95-a338d635a460" (UID: "da6d90be-70f1-42d8-9b95-a338d635a460"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:39:33 crc kubenswrapper[4631]: I1128 13:39:33.819974 4631 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/da6d90be-70f1-42d8-9b95-a338d635a460-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 13:39:33 crc kubenswrapper[4631]: I1128 13:39:33.820020 4631 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/da6d90be-70f1-42d8-9b95-a338d635a460-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 13:39:34 crc kubenswrapper[4631]: I1128 13:39:34.068384 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"da6d90be-70f1-42d8-9b95-a338d635a460","Type":"ContainerDied","Data":"ddfae1373058f71a2b60260b97703d1bac3a7d2ad5ad4f39b7827262b995d16b"} Nov 28 13:39:34 crc kubenswrapper[4631]: I1128 13:39:34.069072 4631 scope.go:117] "RemoveContainer" containerID="82b1ce68db6d83ee312c79d1f5eb6eae79e2cd43dc3c03e78e2d29df0f446585" Nov 28 13:39:34 crc kubenswrapper[4631]: I1128 13:39:34.068594 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 28 13:39:34 crc kubenswrapper[4631]: I1128 13:39:34.133901 4631 scope.go:117] "RemoveContainer" containerID="18f1eeaaf309fee86444a43afd8cd2087c028118589bb2e14b693df9349d4714" Nov 28 13:39:34 crc kubenswrapper[4631]: I1128 13:39:34.138535 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 28 13:39:34 crc kubenswrapper[4631]: I1128 13:39:34.172233 4631 scope.go:117] "RemoveContainer" containerID="11f9d2b9a8e6191a9bcfbb12937052b29700f5e513889b5e941883d14b1e060e" Nov 28 13:39:34 crc kubenswrapper[4631]: I1128 13:39:34.176366 4631 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 28 13:39:34 crc kubenswrapper[4631]: I1128 13:39:34.193046 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 28 13:39:34 crc kubenswrapper[4631]: E1128 13:39:34.193728 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="da6d90be-70f1-42d8-9b95-a338d635a460" containerName="ceilometer-central-agent" Nov 28 13:39:34 crc kubenswrapper[4631]: I1128 13:39:34.193750 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="da6d90be-70f1-42d8-9b95-a338d635a460" containerName="ceilometer-central-agent" Nov 28 13:39:34 crc kubenswrapper[4631]: E1128 13:39:34.193764 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="da6d90be-70f1-42d8-9b95-a338d635a460" containerName="ceilometer-notification-agent" Nov 28 13:39:34 crc kubenswrapper[4631]: I1128 13:39:34.193771 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="da6d90be-70f1-42d8-9b95-a338d635a460" containerName="ceilometer-notification-agent" Nov 28 13:39:34 crc kubenswrapper[4631]: E1128 13:39:34.193784 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="da6d90be-70f1-42d8-9b95-a338d635a460" containerName="sg-core" Nov 28 13:39:34 crc kubenswrapper[4631]: I1128 13:39:34.193790 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="da6d90be-70f1-42d8-9b95-a338d635a460" containerName="sg-core" Nov 28 13:39:34 crc kubenswrapper[4631]: E1128 13:39:34.193823 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="da6d90be-70f1-42d8-9b95-a338d635a460" containerName="proxy-httpd" Nov 28 13:39:34 crc kubenswrapper[4631]: I1128 13:39:34.193829 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="da6d90be-70f1-42d8-9b95-a338d635a460" containerName="proxy-httpd" Nov 28 13:39:34 crc kubenswrapper[4631]: I1128 13:39:34.194002 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="da6d90be-70f1-42d8-9b95-a338d635a460" containerName="ceilometer-central-agent" Nov 28 13:39:34 crc kubenswrapper[4631]: I1128 13:39:34.194016 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="da6d90be-70f1-42d8-9b95-a338d635a460" containerName="sg-core" Nov 28 13:39:34 crc kubenswrapper[4631]: I1128 13:39:34.194025 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="da6d90be-70f1-42d8-9b95-a338d635a460" containerName="ceilometer-notification-agent" Nov 28 13:39:34 crc kubenswrapper[4631]: I1128 13:39:34.194035 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="da6d90be-70f1-42d8-9b95-a338d635a460" containerName="proxy-httpd" Nov 28 13:39:34 crc kubenswrapper[4631]: I1128 13:39:34.196001 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 28 13:39:34 crc kubenswrapper[4631]: I1128 13:39:34.198726 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 28 13:39:34 crc kubenswrapper[4631]: I1128 13:39:34.199019 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 28 13:39:34 crc kubenswrapper[4631]: I1128 13:39:34.207079 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 28 13:39:34 crc kubenswrapper[4631]: I1128 13:39:34.269626 4631 scope.go:117] "RemoveContainer" containerID="fb69638e625d934534f92ebfbe0519166ed42fb448f2dda43bf3db42c019a610" Nov 28 13:39:34 crc kubenswrapper[4631]: I1128 13:39:34.331038 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d511d3b7-eff1-45f3-ab81-0f71b4324dc7-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d511d3b7-eff1-45f3-ab81-0f71b4324dc7\") " pod="openstack/ceilometer-0" Nov 28 13:39:34 crc kubenswrapper[4631]: I1128 13:39:34.331123 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d511d3b7-eff1-45f3-ab81-0f71b4324dc7-log-httpd\") pod \"ceilometer-0\" (UID: \"d511d3b7-eff1-45f3-ab81-0f71b4324dc7\") " pod="openstack/ceilometer-0" Nov 28 13:39:34 crc kubenswrapper[4631]: I1128 13:39:34.331212 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d511d3b7-eff1-45f3-ab81-0f71b4324dc7-scripts\") pod \"ceilometer-0\" (UID: \"d511d3b7-eff1-45f3-ab81-0f71b4324dc7\") " pod="openstack/ceilometer-0" Nov 28 13:39:34 crc kubenswrapper[4631]: I1128 13:39:34.331235 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d511d3b7-eff1-45f3-ab81-0f71b4324dc7-config-data\") pod \"ceilometer-0\" (UID: \"d511d3b7-eff1-45f3-ab81-0f71b4324dc7\") " pod="openstack/ceilometer-0" Nov 28 13:39:34 crc kubenswrapper[4631]: I1128 13:39:34.331268 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d9h2l\" (UniqueName: \"kubernetes.io/projected/d511d3b7-eff1-45f3-ab81-0f71b4324dc7-kube-api-access-d9h2l\") pod \"ceilometer-0\" (UID: \"d511d3b7-eff1-45f3-ab81-0f71b4324dc7\") " pod="openstack/ceilometer-0" Nov 28 13:39:34 crc kubenswrapper[4631]: I1128 13:39:34.331320 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d511d3b7-eff1-45f3-ab81-0f71b4324dc7-run-httpd\") pod \"ceilometer-0\" (UID: \"d511d3b7-eff1-45f3-ab81-0f71b4324dc7\") " pod="openstack/ceilometer-0" Nov 28 13:39:34 crc kubenswrapper[4631]: I1128 13:39:34.331361 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d511d3b7-eff1-45f3-ab81-0f71b4324dc7-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d511d3b7-eff1-45f3-ab81-0f71b4324dc7\") " pod="openstack/ceilometer-0" Nov 28 13:39:34 crc kubenswrapper[4631]: I1128 13:39:34.432965 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d511d3b7-eff1-45f3-ab81-0f71b4324dc7-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d511d3b7-eff1-45f3-ab81-0f71b4324dc7\") " pod="openstack/ceilometer-0" Nov 28 13:39:34 crc kubenswrapper[4631]: I1128 13:39:34.433189 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d511d3b7-eff1-45f3-ab81-0f71b4324dc7-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d511d3b7-eff1-45f3-ab81-0f71b4324dc7\") " pod="openstack/ceilometer-0" Nov 28 13:39:34 crc kubenswrapper[4631]: I1128 13:39:34.433249 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d511d3b7-eff1-45f3-ab81-0f71b4324dc7-log-httpd\") pod \"ceilometer-0\" (UID: \"d511d3b7-eff1-45f3-ab81-0f71b4324dc7\") " pod="openstack/ceilometer-0" Nov 28 13:39:34 crc kubenswrapper[4631]: I1128 13:39:34.433319 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d511d3b7-eff1-45f3-ab81-0f71b4324dc7-scripts\") pod \"ceilometer-0\" (UID: \"d511d3b7-eff1-45f3-ab81-0f71b4324dc7\") " pod="openstack/ceilometer-0" Nov 28 13:39:34 crc kubenswrapper[4631]: I1128 13:39:34.433367 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d511d3b7-eff1-45f3-ab81-0f71b4324dc7-config-data\") pod \"ceilometer-0\" (UID: \"d511d3b7-eff1-45f3-ab81-0f71b4324dc7\") " pod="openstack/ceilometer-0" Nov 28 13:39:34 crc kubenswrapper[4631]: I1128 13:39:34.433400 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d9h2l\" (UniqueName: \"kubernetes.io/projected/d511d3b7-eff1-45f3-ab81-0f71b4324dc7-kube-api-access-d9h2l\") pod \"ceilometer-0\" (UID: \"d511d3b7-eff1-45f3-ab81-0f71b4324dc7\") " pod="openstack/ceilometer-0" Nov 28 13:39:34 crc kubenswrapper[4631]: I1128 13:39:34.433435 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d511d3b7-eff1-45f3-ab81-0f71b4324dc7-run-httpd\") pod \"ceilometer-0\" (UID: \"d511d3b7-eff1-45f3-ab81-0f71b4324dc7\") " pod="openstack/ceilometer-0" Nov 28 13:39:34 crc kubenswrapper[4631]: I1128 13:39:34.435734 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d511d3b7-eff1-45f3-ab81-0f71b4324dc7-run-httpd\") pod \"ceilometer-0\" (UID: \"d511d3b7-eff1-45f3-ab81-0f71b4324dc7\") " pod="openstack/ceilometer-0" Nov 28 13:39:34 crc kubenswrapper[4631]: I1128 13:39:34.435969 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d511d3b7-eff1-45f3-ab81-0f71b4324dc7-log-httpd\") pod \"ceilometer-0\" (UID: \"d511d3b7-eff1-45f3-ab81-0f71b4324dc7\") " pod="openstack/ceilometer-0" Nov 28 13:39:34 crc kubenswrapper[4631]: I1128 13:39:34.439838 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d511d3b7-eff1-45f3-ab81-0f71b4324dc7-config-data\") pod \"ceilometer-0\" (UID: \"d511d3b7-eff1-45f3-ab81-0f71b4324dc7\") " pod="openstack/ceilometer-0" Nov 28 13:39:34 crc kubenswrapper[4631]: I1128 13:39:34.440639 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d511d3b7-eff1-45f3-ab81-0f71b4324dc7-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d511d3b7-eff1-45f3-ab81-0f71b4324dc7\") " pod="openstack/ceilometer-0" Nov 28 13:39:34 crc kubenswrapper[4631]: I1128 13:39:34.440665 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d511d3b7-eff1-45f3-ab81-0f71b4324dc7-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d511d3b7-eff1-45f3-ab81-0f71b4324dc7\") " pod="openstack/ceilometer-0" Nov 28 13:39:34 crc kubenswrapper[4631]: I1128 13:39:34.444958 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d511d3b7-eff1-45f3-ab81-0f71b4324dc7-scripts\") pod \"ceilometer-0\" (UID: \"d511d3b7-eff1-45f3-ab81-0f71b4324dc7\") " pod="openstack/ceilometer-0" Nov 28 13:39:34 crc kubenswrapper[4631]: I1128 13:39:34.456195 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d9h2l\" (UniqueName: \"kubernetes.io/projected/d511d3b7-eff1-45f3-ab81-0f71b4324dc7-kube-api-access-d9h2l\") pod \"ceilometer-0\" (UID: \"d511d3b7-eff1-45f3-ab81-0f71b4324dc7\") " pod="openstack/ceilometer-0" Nov 28 13:39:34 crc kubenswrapper[4631]: I1128 13:39:34.536832 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 28 13:39:35 crc kubenswrapper[4631]: I1128 13:39:35.364211 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 28 13:39:35 crc kubenswrapper[4631]: W1128 13:39:35.366175 4631 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd511d3b7_eff1_45f3_ab81_0f71b4324dc7.slice/crio-18317c37a5634a2985b1a09f1f859abc40b0a1ad1a849fed91b80714aa2742c7 WatchSource:0}: Error finding container 18317c37a5634a2985b1a09f1f859abc40b0a1ad1a849fed91b80714aa2742c7: Status 404 returned error can't find the container with id 18317c37a5634a2985b1a09f1f859abc40b0a1ad1a849fed91b80714aa2742c7 Nov 28 13:39:35 crc kubenswrapper[4631]: I1128 13:39:35.534418 4631 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="da6d90be-70f1-42d8-9b95-a338d635a460" path="/var/lib/kubelet/pods/da6d90be-70f1-42d8-9b95-a338d635a460/volumes" Nov 28 13:39:36 crc kubenswrapper[4631]: I1128 13:39:36.108840 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d511d3b7-eff1-45f3-ab81-0f71b4324dc7","Type":"ContainerStarted","Data":"18317c37a5634a2985b1a09f1f859abc40b0a1ad1a849fed91b80714aa2742c7"} Nov 28 13:39:36 crc kubenswrapper[4631]: I1128 13:39:36.716355 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-sqgzw"] Nov 28 13:39:36 crc kubenswrapper[4631]: I1128 13:39:36.720944 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-sqgzw" Nov 28 13:39:36 crc kubenswrapper[4631]: I1128 13:39:36.788142 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-sqgzw"] Nov 28 13:39:36 crc kubenswrapper[4631]: I1128 13:39:36.798593 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bcf77\" (UniqueName: \"kubernetes.io/projected/325fdb2e-59f8-4d94-860f-3ae7bd2e231e-kube-api-access-bcf77\") pod \"nova-api-db-create-sqgzw\" (UID: \"325fdb2e-59f8-4d94-860f-3ae7bd2e231e\") " pod="openstack/nova-api-db-create-sqgzw" Nov 28 13:39:36 crc kubenswrapper[4631]: I1128 13:39:36.798690 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/325fdb2e-59f8-4d94-860f-3ae7bd2e231e-operator-scripts\") pod \"nova-api-db-create-sqgzw\" (UID: \"325fdb2e-59f8-4d94-860f-3ae7bd2e231e\") " pod="openstack/nova-api-db-create-sqgzw" Nov 28 13:39:36 crc kubenswrapper[4631]: I1128 13:39:36.903614 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bcf77\" (UniqueName: \"kubernetes.io/projected/325fdb2e-59f8-4d94-860f-3ae7bd2e231e-kube-api-access-bcf77\") pod \"nova-api-db-create-sqgzw\" (UID: \"325fdb2e-59f8-4d94-860f-3ae7bd2e231e\") " pod="openstack/nova-api-db-create-sqgzw" Nov 28 13:39:36 crc kubenswrapper[4631]: I1128 13:39:36.904105 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/325fdb2e-59f8-4d94-860f-3ae7bd2e231e-operator-scripts\") pod \"nova-api-db-create-sqgzw\" (UID: \"325fdb2e-59f8-4d94-860f-3ae7bd2e231e\") " pod="openstack/nova-api-db-create-sqgzw" Nov 28 13:39:36 crc kubenswrapper[4631]: I1128 13:39:36.905190 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/325fdb2e-59f8-4d94-860f-3ae7bd2e231e-operator-scripts\") pod \"nova-api-db-create-sqgzw\" (UID: \"325fdb2e-59f8-4d94-860f-3ae7bd2e231e\") " pod="openstack/nova-api-db-create-sqgzw" Nov 28 13:39:36 crc kubenswrapper[4631]: I1128 13:39:36.923527 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-fc8wj"] Nov 28 13:39:36 crc kubenswrapper[4631]: I1128 13:39:36.925181 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-fc8wj" Nov 28 13:39:36 crc kubenswrapper[4631]: I1128 13:39:36.947948 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-5cd3-account-create-update-hrtjp"] Nov 28 13:39:36 crc kubenswrapper[4631]: I1128 13:39:36.949453 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-5cd3-account-create-update-hrtjp" Nov 28 13:39:36 crc kubenswrapper[4631]: I1128 13:39:36.952705 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Nov 28 13:39:37 crc kubenswrapper[4631]: I1128 13:39:37.001165 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bcf77\" (UniqueName: \"kubernetes.io/projected/325fdb2e-59f8-4d94-860f-3ae7bd2e231e-kube-api-access-bcf77\") pod \"nova-api-db-create-sqgzw\" (UID: \"325fdb2e-59f8-4d94-860f-3ae7bd2e231e\") " pod="openstack/nova-api-db-create-sqgzw" Nov 28 13:39:37 crc kubenswrapper[4631]: I1128 13:39:37.001280 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-5cd3-account-create-update-hrtjp"] Nov 28 13:39:37 crc kubenswrapper[4631]: I1128 13:39:37.008838 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0f61d189-17d8-4154-ac95-df78d4042055-operator-scripts\") pod \"nova-api-5cd3-account-create-update-hrtjp\" (UID: \"0f61d189-17d8-4154-ac95-df78d4042055\") " pod="openstack/nova-api-5cd3-account-create-update-hrtjp" Nov 28 13:39:37 crc kubenswrapper[4631]: I1128 13:39:37.008907 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/641b1d24-f22a-4d63-8e9e-d76c3dde2135-operator-scripts\") pod \"nova-cell0-db-create-fc8wj\" (UID: \"641b1d24-f22a-4d63-8e9e-d76c3dde2135\") " pod="openstack/nova-cell0-db-create-fc8wj" Nov 28 13:39:37 crc kubenswrapper[4631]: I1128 13:39:37.008993 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5rkjw\" (UniqueName: \"kubernetes.io/projected/641b1d24-f22a-4d63-8e9e-d76c3dde2135-kube-api-access-5rkjw\") pod \"nova-cell0-db-create-fc8wj\" (UID: \"641b1d24-f22a-4d63-8e9e-d76c3dde2135\") " pod="openstack/nova-cell0-db-create-fc8wj" Nov 28 13:39:37 crc kubenswrapper[4631]: I1128 13:39:37.009060 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z2lbg\" (UniqueName: \"kubernetes.io/projected/0f61d189-17d8-4154-ac95-df78d4042055-kube-api-access-z2lbg\") pod \"nova-api-5cd3-account-create-update-hrtjp\" (UID: \"0f61d189-17d8-4154-ac95-df78d4042055\") " pod="openstack/nova-api-5cd3-account-create-update-hrtjp" Nov 28 13:39:37 crc kubenswrapper[4631]: I1128 13:39:37.029304 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-fc8wj"] Nov 28 13:39:37 crc kubenswrapper[4631]: I1128 13:39:37.104832 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-x4t6b"] Nov 28 13:39:37 crc kubenswrapper[4631]: I1128 13:39:37.106427 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-x4t6b" Nov 28 13:39:37 crc kubenswrapper[4631]: I1128 13:39:37.110735 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z2lbg\" (UniqueName: \"kubernetes.io/projected/0f61d189-17d8-4154-ac95-df78d4042055-kube-api-access-z2lbg\") pod \"nova-api-5cd3-account-create-update-hrtjp\" (UID: \"0f61d189-17d8-4154-ac95-df78d4042055\") " pod="openstack/nova-api-5cd3-account-create-update-hrtjp" Nov 28 13:39:37 crc kubenswrapper[4631]: I1128 13:39:37.110805 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0f61d189-17d8-4154-ac95-df78d4042055-operator-scripts\") pod \"nova-api-5cd3-account-create-update-hrtjp\" (UID: \"0f61d189-17d8-4154-ac95-df78d4042055\") " pod="openstack/nova-api-5cd3-account-create-update-hrtjp" Nov 28 13:39:37 crc kubenswrapper[4631]: I1128 13:39:37.110840 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/641b1d24-f22a-4d63-8e9e-d76c3dde2135-operator-scripts\") pod \"nova-cell0-db-create-fc8wj\" (UID: \"641b1d24-f22a-4d63-8e9e-d76c3dde2135\") " pod="openstack/nova-cell0-db-create-fc8wj" Nov 28 13:39:37 crc kubenswrapper[4631]: I1128 13:39:37.110896 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5rkjw\" (UniqueName: \"kubernetes.io/projected/641b1d24-f22a-4d63-8e9e-d76c3dde2135-kube-api-access-5rkjw\") pod \"nova-cell0-db-create-fc8wj\" (UID: \"641b1d24-f22a-4d63-8e9e-d76c3dde2135\") " pod="openstack/nova-cell0-db-create-fc8wj" Nov 28 13:39:37 crc kubenswrapper[4631]: I1128 13:39:37.112135 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0f61d189-17d8-4154-ac95-df78d4042055-operator-scripts\") pod \"nova-api-5cd3-account-create-update-hrtjp\" (UID: \"0f61d189-17d8-4154-ac95-df78d4042055\") " pod="openstack/nova-api-5cd3-account-create-update-hrtjp" Nov 28 13:39:37 crc kubenswrapper[4631]: I1128 13:39:37.112736 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/641b1d24-f22a-4d63-8e9e-d76c3dde2135-operator-scripts\") pod \"nova-cell0-db-create-fc8wj\" (UID: \"641b1d24-f22a-4d63-8e9e-d76c3dde2135\") " pod="openstack/nova-cell0-db-create-fc8wj" Nov 28 13:39:37 crc kubenswrapper[4631]: I1128 13:39:37.125842 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-sqgzw" Nov 28 13:39:37 crc kubenswrapper[4631]: I1128 13:39:37.128829 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-x4t6b"] Nov 28 13:39:37 crc kubenswrapper[4631]: I1128 13:39:37.146607 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d511d3b7-eff1-45f3-ab81-0f71b4324dc7","Type":"ContainerStarted","Data":"c452f408f264bcc78b2e6824e66bfabf1f3b5b5fb29219ddb376b76dc921bb99"} Nov 28 13:39:37 crc kubenswrapper[4631]: I1128 13:39:37.146663 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d511d3b7-eff1-45f3-ab81-0f71b4324dc7","Type":"ContainerStarted","Data":"978325b49b1129c5ccb877420d3040e957d3e5e744cc5b5115cc7d1c5668e094"} Nov 28 13:39:37 crc kubenswrapper[4631]: I1128 13:39:37.147506 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z2lbg\" (UniqueName: \"kubernetes.io/projected/0f61d189-17d8-4154-ac95-df78d4042055-kube-api-access-z2lbg\") pod \"nova-api-5cd3-account-create-update-hrtjp\" (UID: \"0f61d189-17d8-4154-ac95-df78d4042055\") " pod="openstack/nova-api-5cd3-account-create-update-hrtjp" Nov 28 13:39:37 crc kubenswrapper[4631]: I1128 13:39:37.157707 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5rkjw\" (UniqueName: \"kubernetes.io/projected/641b1d24-f22a-4d63-8e9e-d76c3dde2135-kube-api-access-5rkjw\") pod \"nova-cell0-db-create-fc8wj\" (UID: \"641b1d24-f22a-4d63-8e9e-d76c3dde2135\") " pod="openstack/nova-cell0-db-create-fc8wj" Nov 28 13:39:37 crc kubenswrapper[4631]: I1128 13:39:37.161356 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-099f-account-create-update-f2xhr"] Nov 28 13:39:37 crc kubenswrapper[4631]: I1128 13:39:37.163024 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-099f-account-create-update-f2xhr" Nov 28 13:39:37 crc kubenswrapper[4631]: I1128 13:39:37.168866 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Nov 28 13:39:37 crc kubenswrapper[4631]: I1128 13:39:37.169717 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-099f-account-create-update-f2xhr"] Nov 28 13:39:37 crc kubenswrapper[4631]: I1128 13:39:37.213656 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/98edb37a-2698-4c30-a397-70d83441063d-operator-scripts\") pod \"nova-cell0-099f-account-create-update-f2xhr\" (UID: \"98edb37a-2698-4c30-a397-70d83441063d\") " pod="openstack/nova-cell0-099f-account-create-update-f2xhr" Nov 28 13:39:37 crc kubenswrapper[4631]: I1128 13:39:37.213730 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mkt2j\" (UniqueName: \"kubernetes.io/projected/98edb37a-2698-4c30-a397-70d83441063d-kube-api-access-mkt2j\") pod \"nova-cell0-099f-account-create-update-f2xhr\" (UID: \"98edb37a-2698-4c30-a397-70d83441063d\") " pod="openstack/nova-cell0-099f-account-create-update-f2xhr" Nov 28 13:39:37 crc kubenswrapper[4631]: I1128 13:39:37.213851 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c5a28e08-ff16-4ecf-97ae-543c9de270f8-operator-scripts\") pod \"nova-cell1-db-create-x4t6b\" (UID: \"c5a28e08-ff16-4ecf-97ae-543c9de270f8\") " pod="openstack/nova-cell1-db-create-x4t6b" Nov 28 13:39:37 crc kubenswrapper[4631]: I1128 13:39:37.213904 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vkxp8\" (UniqueName: \"kubernetes.io/projected/c5a28e08-ff16-4ecf-97ae-543c9de270f8-kube-api-access-vkxp8\") pod \"nova-cell1-db-create-x4t6b\" (UID: \"c5a28e08-ff16-4ecf-97ae-543c9de270f8\") " pod="openstack/nova-cell1-db-create-x4t6b" Nov 28 13:39:37 crc kubenswrapper[4631]: I1128 13:39:37.316899 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c5a28e08-ff16-4ecf-97ae-543c9de270f8-operator-scripts\") pod \"nova-cell1-db-create-x4t6b\" (UID: \"c5a28e08-ff16-4ecf-97ae-543c9de270f8\") " pod="openstack/nova-cell1-db-create-x4t6b" Nov 28 13:39:37 crc kubenswrapper[4631]: I1128 13:39:37.317302 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vkxp8\" (UniqueName: \"kubernetes.io/projected/c5a28e08-ff16-4ecf-97ae-543c9de270f8-kube-api-access-vkxp8\") pod \"nova-cell1-db-create-x4t6b\" (UID: \"c5a28e08-ff16-4ecf-97ae-543c9de270f8\") " pod="openstack/nova-cell1-db-create-x4t6b" Nov 28 13:39:37 crc kubenswrapper[4631]: I1128 13:39:37.317342 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/98edb37a-2698-4c30-a397-70d83441063d-operator-scripts\") pod \"nova-cell0-099f-account-create-update-f2xhr\" (UID: \"98edb37a-2698-4c30-a397-70d83441063d\") " pod="openstack/nova-cell0-099f-account-create-update-f2xhr" Nov 28 13:39:37 crc kubenswrapper[4631]: I1128 13:39:37.317382 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mkt2j\" (UniqueName: \"kubernetes.io/projected/98edb37a-2698-4c30-a397-70d83441063d-kube-api-access-mkt2j\") pod \"nova-cell0-099f-account-create-update-f2xhr\" (UID: \"98edb37a-2698-4c30-a397-70d83441063d\") " pod="openstack/nova-cell0-099f-account-create-update-f2xhr" Nov 28 13:39:37 crc kubenswrapper[4631]: I1128 13:39:37.318604 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c5a28e08-ff16-4ecf-97ae-543c9de270f8-operator-scripts\") pod \"nova-cell1-db-create-x4t6b\" (UID: \"c5a28e08-ff16-4ecf-97ae-543c9de270f8\") " pod="openstack/nova-cell1-db-create-x4t6b" Nov 28 13:39:37 crc kubenswrapper[4631]: I1128 13:39:37.319218 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/98edb37a-2698-4c30-a397-70d83441063d-operator-scripts\") pod \"nova-cell0-099f-account-create-update-f2xhr\" (UID: \"98edb37a-2698-4c30-a397-70d83441063d\") " pod="openstack/nova-cell0-099f-account-create-update-f2xhr" Nov 28 13:39:37 crc kubenswrapper[4631]: I1128 13:39:37.324555 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-0720-account-create-update-fhl44"] Nov 28 13:39:37 crc kubenswrapper[4631]: I1128 13:39:37.329412 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-fc8wj" Nov 28 13:39:37 crc kubenswrapper[4631]: I1128 13:39:37.335611 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-0720-account-create-update-fhl44" Nov 28 13:39:37 crc kubenswrapper[4631]: I1128 13:39:37.347231 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-5cd3-account-create-update-hrtjp" Nov 28 13:39:37 crc kubenswrapper[4631]: I1128 13:39:37.351219 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mkt2j\" (UniqueName: \"kubernetes.io/projected/98edb37a-2698-4c30-a397-70d83441063d-kube-api-access-mkt2j\") pod \"nova-cell0-099f-account-create-update-f2xhr\" (UID: \"98edb37a-2698-4c30-a397-70d83441063d\") " pod="openstack/nova-cell0-099f-account-create-update-f2xhr" Nov 28 13:39:37 crc kubenswrapper[4631]: I1128 13:39:37.352479 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Nov 28 13:39:37 crc kubenswrapper[4631]: I1128 13:39:37.363761 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vkxp8\" (UniqueName: \"kubernetes.io/projected/c5a28e08-ff16-4ecf-97ae-543c9de270f8-kube-api-access-vkxp8\") pod \"nova-cell1-db-create-x4t6b\" (UID: \"c5a28e08-ff16-4ecf-97ae-543c9de270f8\") " pod="openstack/nova-cell1-db-create-x4t6b" Nov 28 13:39:37 crc kubenswrapper[4631]: I1128 13:39:37.385601 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-0720-account-create-update-fhl44"] Nov 28 13:39:37 crc kubenswrapper[4631]: I1128 13:39:37.422373 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5a1eb7ab-04ab-476e-aa72-2a3a3f81eb21-operator-scripts\") pod \"nova-cell1-0720-account-create-update-fhl44\" (UID: \"5a1eb7ab-04ab-476e-aa72-2a3a3f81eb21\") " pod="openstack/nova-cell1-0720-account-create-update-fhl44" Nov 28 13:39:37 crc kubenswrapper[4631]: I1128 13:39:37.422430 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hks8v\" (UniqueName: \"kubernetes.io/projected/5a1eb7ab-04ab-476e-aa72-2a3a3f81eb21-kube-api-access-hks8v\") pod \"nova-cell1-0720-account-create-update-fhl44\" (UID: \"5a1eb7ab-04ab-476e-aa72-2a3a3f81eb21\") " pod="openstack/nova-cell1-0720-account-create-update-fhl44" Nov 28 13:39:37 crc kubenswrapper[4631]: I1128 13:39:37.431309 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-x4t6b" Nov 28 13:39:37 crc kubenswrapper[4631]: I1128 13:39:37.524307 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5a1eb7ab-04ab-476e-aa72-2a3a3f81eb21-operator-scripts\") pod \"nova-cell1-0720-account-create-update-fhl44\" (UID: \"5a1eb7ab-04ab-476e-aa72-2a3a3f81eb21\") " pod="openstack/nova-cell1-0720-account-create-update-fhl44" Nov 28 13:39:37 crc kubenswrapper[4631]: I1128 13:39:37.524503 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hks8v\" (UniqueName: \"kubernetes.io/projected/5a1eb7ab-04ab-476e-aa72-2a3a3f81eb21-kube-api-access-hks8v\") pod \"nova-cell1-0720-account-create-update-fhl44\" (UID: \"5a1eb7ab-04ab-476e-aa72-2a3a3f81eb21\") " pod="openstack/nova-cell1-0720-account-create-update-fhl44" Nov 28 13:39:37 crc kubenswrapper[4631]: I1128 13:39:37.527177 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5a1eb7ab-04ab-476e-aa72-2a3a3f81eb21-operator-scripts\") pod \"nova-cell1-0720-account-create-update-fhl44\" (UID: \"5a1eb7ab-04ab-476e-aa72-2a3a3f81eb21\") " pod="openstack/nova-cell1-0720-account-create-update-fhl44" Nov 28 13:39:37 crc kubenswrapper[4631]: I1128 13:39:37.546072 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hks8v\" (UniqueName: \"kubernetes.io/projected/5a1eb7ab-04ab-476e-aa72-2a3a3f81eb21-kube-api-access-hks8v\") pod \"nova-cell1-0720-account-create-update-fhl44\" (UID: \"5a1eb7ab-04ab-476e-aa72-2a3a3f81eb21\") " pod="openstack/nova-cell1-0720-account-create-update-fhl44" Nov 28 13:39:37 crc kubenswrapper[4631]: I1128 13:39:37.558638 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-099f-account-create-update-f2xhr" Nov 28 13:39:37 crc kubenswrapper[4631]: I1128 13:39:37.680233 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-0720-account-create-update-fhl44" Nov 28 13:39:37 crc kubenswrapper[4631]: I1128 13:39:37.804654 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-sqgzw"] Nov 28 13:39:38 crc kubenswrapper[4631]: I1128 13:39:38.183245 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-sqgzw" event={"ID":"325fdb2e-59f8-4d94-860f-3ae7bd2e231e","Type":"ContainerStarted","Data":"4d5022c831b0fef5c476ac9ef9ab45cacf078c669990e0396599fb7965acc187"} Nov 28 13:39:38 crc kubenswrapper[4631]: I1128 13:39:38.246643 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-fc8wj"] Nov 28 13:39:38 crc kubenswrapper[4631]: I1128 13:39:38.447011 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-x4t6b"] Nov 28 13:39:39 crc kubenswrapper[4631]: I1128 13:39:39.204461 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-fc8wj" event={"ID":"641b1d24-f22a-4d63-8e9e-d76c3dde2135","Type":"ContainerStarted","Data":"3ddf2c9e1b460818b8085657a4e3cb74f75e0172d40feb83b851eb50ccd9d229"} Nov 28 13:39:39 crc kubenswrapper[4631]: I1128 13:39:39.208379 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-x4t6b" event={"ID":"c5a28e08-ff16-4ecf-97ae-543c9de270f8","Type":"ContainerStarted","Data":"534eff893a6ef7cafd7fb5b6bacbc255c1e4c9a1102bc06d9125061c3451f7c3"} Nov 28 13:39:39 crc kubenswrapper[4631]: I1128 13:39:39.362684 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 28 13:39:40 crc kubenswrapper[4631]: I1128 13:39:40.023471 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-5cd3-account-create-update-hrtjp"] Nov 28 13:39:40 crc kubenswrapper[4631]: I1128 13:39:40.036173 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-099f-account-create-update-f2xhr"] Nov 28 13:39:40 crc kubenswrapper[4631]: W1128 13:39:40.052210 4631 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod98edb37a_2698_4c30_a397_70d83441063d.slice/crio-c1f50b03d90e478dde95101498bf1138c22ac879e00e032679c2e386793c11a2 WatchSource:0}: Error finding container c1f50b03d90e478dde95101498bf1138c22ac879e00e032679c2e386793c11a2: Status 404 returned error can't find the container with id c1f50b03d90e478dde95101498bf1138c22ac879e00e032679c2e386793c11a2 Nov 28 13:39:40 crc kubenswrapper[4631]: I1128 13:39:40.059377 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-0720-account-create-update-fhl44"] Nov 28 13:39:40 crc kubenswrapper[4631]: W1128 13:39:40.074993 4631 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5a1eb7ab_04ab_476e_aa72_2a3a3f81eb21.slice/crio-de7231651223aa6d09ea4b703fd5feb347d9652c7f13fcf1ef8e1e737c642618 WatchSource:0}: Error finding container de7231651223aa6d09ea4b703fd5feb347d9652c7f13fcf1ef8e1e737c642618: Status 404 returned error can't find the container with id de7231651223aa6d09ea4b703fd5feb347d9652c7f13fcf1ef8e1e737c642618 Nov 28 13:39:40 crc kubenswrapper[4631]: I1128 13:39:40.223923 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-099f-account-create-update-f2xhr" event={"ID":"98edb37a-2698-4c30-a397-70d83441063d","Type":"ContainerStarted","Data":"c1f50b03d90e478dde95101498bf1138c22ac879e00e032679c2e386793c11a2"} Nov 28 13:39:40 crc kubenswrapper[4631]: I1128 13:39:40.225521 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-0720-account-create-update-fhl44" event={"ID":"5a1eb7ab-04ab-476e-aa72-2a3a3f81eb21","Type":"ContainerStarted","Data":"de7231651223aa6d09ea4b703fd5feb347d9652c7f13fcf1ef8e1e737c642618"} Nov 28 13:39:40 crc kubenswrapper[4631]: I1128 13:39:40.227211 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-5cd3-account-create-update-hrtjp" event={"ID":"0f61d189-17d8-4154-ac95-df78d4042055","Type":"ContainerStarted","Data":"0a32d7ae2c3a534184dba524af21eb67d2088a5219dc164fd35b2643f3e54ddd"} Nov 28 13:39:41 crc kubenswrapper[4631]: I1128 13:39:41.242397 4631 generic.go:334] "Generic (PLEG): container finished" podID="5a1eb7ab-04ab-476e-aa72-2a3a3f81eb21" containerID="a08b1a5b404934f0700e2e2d8cf33b72868086a48b28b7e15d65392a9e1d62a1" exitCode=0 Nov 28 13:39:41 crc kubenswrapper[4631]: I1128 13:39:41.242511 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-0720-account-create-update-fhl44" event={"ID":"5a1eb7ab-04ab-476e-aa72-2a3a3f81eb21","Type":"ContainerDied","Data":"a08b1a5b404934f0700e2e2d8cf33b72868086a48b28b7e15d65392a9e1d62a1"} Nov 28 13:39:41 crc kubenswrapper[4631]: I1128 13:39:41.246975 4631 generic.go:334] "Generic (PLEG): container finished" podID="0f61d189-17d8-4154-ac95-df78d4042055" containerID="fa715c259033c082f538da5fdbeffa2585161112c7dcee2b390f49966d22329d" exitCode=0 Nov 28 13:39:41 crc kubenswrapper[4631]: I1128 13:39:41.247200 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-5cd3-account-create-update-hrtjp" event={"ID":"0f61d189-17d8-4154-ac95-df78d4042055","Type":"ContainerDied","Data":"fa715c259033c082f538da5fdbeffa2585161112c7dcee2b390f49966d22329d"} Nov 28 13:39:41 crc kubenswrapper[4631]: I1128 13:39:41.252023 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d511d3b7-eff1-45f3-ab81-0f71b4324dc7","Type":"ContainerStarted","Data":"8445f38e73c31d3d4d1042303914d59653deadd33acca0865c3b7269356144b8"} Nov 28 13:39:41 crc kubenswrapper[4631]: I1128 13:39:41.254532 4631 generic.go:334] "Generic (PLEG): container finished" podID="325fdb2e-59f8-4d94-860f-3ae7bd2e231e" containerID="ef5349a016a2aa007616d1b0bd682c055321668c090392075d244fb0a424d9e2" exitCode=0 Nov 28 13:39:41 crc kubenswrapper[4631]: I1128 13:39:41.254613 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-sqgzw" event={"ID":"325fdb2e-59f8-4d94-860f-3ae7bd2e231e","Type":"ContainerDied","Data":"ef5349a016a2aa007616d1b0bd682c055321668c090392075d244fb0a424d9e2"} Nov 28 13:39:41 crc kubenswrapper[4631]: I1128 13:39:41.257532 4631 generic.go:334] "Generic (PLEG): container finished" podID="641b1d24-f22a-4d63-8e9e-d76c3dde2135" containerID="b742ee5d45fe80719ab66c8cdecf4b2c06b42061a669d4bff8463258e674cf40" exitCode=0 Nov 28 13:39:41 crc kubenswrapper[4631]: I1128 13:39:41.257617 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-fc8wj" event={"ID":"641b1d24-f22a-4d63-8e9e-d76c3dde2135","Type":"ContainerDied","Data":"b742ee5d45fe80719ab66c8cdecf4b2c06b42061a669d4bff8463258e674cf40"} Nov 28 13:39:41 crc kubenswrapper[4631]: I1128 13:39:41.261413 4631 generic.go:334] "Generic (PLEG): container finished" podID="c5a28e08-ff16-4ecf-97ae-543c9de270f8" containerID="92e24cc7e65e275d64d28d4779c3e78ddb7fa944fc5c9460736e81d594bb492f" exitCode=0 Nov 28 13:39:41 crc kubenswrapper[4631]: I1128 13:39:41.261579 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-x4t6b" event={"ID":"c5a28e08-ff16-4ecf-97ae-543c9de270f8","Type":"ContainerDied","Data":"92e24cc7e65e275d64d28d4779c3e78ddb7fa944fc5c9460736e81d594bb492f"} Nov 28 13:39:41 crc kubenswrapper[4631]: I1128 13:39:41.264212 4631 generic.go:334] "Generic (PLEG): container finished" podID="98edb37a-2698-4c30-a397-70d83441063d" containerID="aba4ee440f097b8a145c46ec9a1b706abef9c12de69dbf06a20556acea03afa4" exitCode=0 Nov 28 13:39:41 crc kubenswrapper[4631]: I1128 13:39:41.264265 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-099f-account-create-update-f2xhr" event={"ID":"98edb37a-2698-4c30-a397-70d83441063d","Type":"ContainerDied","Data":"aba4ee440f097b8a145c46ec9a1b706abef9c12de69dbf06a20556acea03afa4"} Nov 28 13:39:42 crc kubenswrapper[4631]: I1128 13:39:42.275829 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d511d3b7-eff1-45f3-ab81-0f71b4324dc7","Type":"ContainerStarted","Data":"75b00bb29802e3263f0954877c7012663f5f708e8959b1b05c30795cf3e7c879"} Nov 28 13:39:42 crc kubenswrapper[4631]: I1128 13:39:42.275923 4631 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d511d3b7-eff1-45f3-ab81-0f71b4324dc7" containerName="ceilometer-central-agent" containerID="cri-o://978325b49b1129c5ccb877420d3040e957d3e5e744cc5b5115cc7d1c5668e094" gracePeriod=30 Nov 28 13:39:42 crc kubenswrapper[4631]: I1128 13:39:42.276165 4631 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d511d3b7-eff1-45f3-ab81-0f71b4324dc7" containerName="sg-core" containerID="cri-o://8445f38e73c31d3d4d1042303914d59653deadd33acca0865c3b7269356144b8" gracePeriod=30 Nov 28 13:39:42 crc kubenswrapper[4631]: I1128 13:39:42.276185 4631 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d511d3b7-eff1-45f3-ab81-0f71b4324dc7" containerName="proxy-httpd" containerID="cri-o://75b00bb29802e3263f0954877c7012663f5f708e8959b1b05c30795cf3e7c879" gracePeriod=30 Nov 28 13:39:42 crc kubenswrapper[4631]: I1128 13:39:42.276195 4631 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d511d3b7-eff1-45f3-ab81-0f71b4324dc7" containerName="ceilometer-notification-agent" containerID="cri-o://c452f408f264bcc78b2e6824e66bfabf1f3b5b5fb29219ddb376b76dc921bb99" gracePeriod=30 Nov 28 13:39:42 crc kubenswrapper[4631]: I1128 13:39:42.285699 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 28 13:39:42 crc kubenswrapper[4631]: I1128 13:39:42.311189 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.110158373 podStartE2EDuration="8.3111434s" podCreationTimestamp="2025-11-28 13:39:34 +0000 UTC" firstStartedPulling="2025-11-28 13:39:35.369546705 +0000 UTC m=+1132.176850059" lastFinishedPulling="2025-11-28 13:39:41.570531742 +0000 UTC m=+1138.377835086" observedRunningTime="2025-11-28 13:39:42.301818708 +0000 UTC m=+1139.109122062" watchObservedRunningTime="2025-11-28 13:39:42.3111434 +0000 UTC m=+1139.118446754" Nov 28 13:39:43 crc kubenswrapper[4631]: I1128 13:39:43.034666 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-sqgzw" Nov 28 13:39:43 crc kubenswrapper[4631]: I1128 13:39:43.074878 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/325fdb2e-59f8-4d94-860f-3ae7bd2e231e-operator-scripts\") pod \"325fdb2e-59f8-4d94-860f-3ae7bd2e231e\" (UID: \"325fdb2e-59f8-4d94-860f-3ae7bd2e231e\") " Nov 28 13:39:43 crc kubenswrapper[4631]: I1128 13:39:43.074962 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bcf77\" (UniqueName: \"kubernetes.io/projected/325fdb2e-59f8-4d94-860f-3ae7bd2e231e-kube-api-access-bcf77\") pod \"325fdb2e-59f8-4d94-860f-3ae7bd2e231e\" (UID: \"325fdb2e-59f8-4d94-860f-3ae7bd2e231e\") " Nov 28 13:39:43 crc kubenswrapper[4631]: I1128 13:39:43.076269 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/325fdb2e-59f8-4d94-860f-3ae7bd2e231e-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "325fdb2e-59f8-4d94-860f-3ae7bd2e231e" (UID: "325fdb2e-59f8-4d94-860f-3ae7bd2e231e"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 13:39:43 crc kubenswrapper[4631]: I1128 13:39:43.083400 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/325fdb2e-59f8-4d94-860f-3ae7bd2e231e-kube-api-access-bcf77" (OuterVolumeSpecName: "kube-api-access-bcf77") pod "325fdb2e-59f8-4d94-860f-3ae7bd2e231e" (UID: "325fdb2e-59f8-4d94-860f-3ae7bd2e231e"). InnerVolumeSpecName "kube-api-access-bcf77". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:39:43 crc kubenswrapper[4631]: I1128 13:39:43.178942 4631 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/325fdb2e-59f8-4d94-860f-3ae7bd2e231e-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 13:39:43 crc kubenswrapper[4631]: I1128 13:39:43.178985 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bcf77\" (UniqueName: \"kubernetes.io/projected/325fdb2e-59f8-4d94-860f-3ae7bd2e231e-kube-api-access-bcf77\") on node \"crc\" DevicePath \"\"" Nov 28 13:39:43 crc kubenswrapper[4631]: I1128 13:39:43.189062 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-5cd3-account-create-update-hrtjp" Nov 28 13:39:43 crc kubenswrapper[4631]: I1128 13:39:43.204107 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-099f-account-create-update-f2xhr" Nov 28 13:39:43 crc kubenswrapper[4631]: I1128 13:39:43.219870 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-fc8wj" Nov 28 13:39:43 crc kubenswrapper[4631]: I1128 13:39:43.230847 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-0720-account-create-update-fhl44" Nov 28 13:39:43 crc kubenswrapper[4631]: I1128 13:39:43.244615 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-x4t6b" Nov 28 13:39:43 crc kubenswrapper[4631]: I1128 13:39:43.298624 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-099f-account-create-update-f2xhr" Nov 28 13:39:43 crc kubenswrapper[4631]: I1128 13:39:43.299604 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-099f-account-create-update-f2xhr" event={"ID":"98edb37a-2698-4c30-a397-70d83441063d","Type":"ContainerDied","Data":"c1f50b03d90e478dde95101498bf1138c22ac879e00e032679c2e386793c11a2"} Nov 28 13:39:43 crc kubenswrapper[4631]: I1128 13:39:43.299641 4631 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c1f50b03d90e478dde95101498bf1138c22ac879e00e032679c2e386793c11a2" Nov 28 13:39:43 crc kubenswrapper[4631]: I1128 13:39:43.304060 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-0720-account-create-update-fhl44" event={"ID":"5a1eb7ab-04ab-476e-aa72-2a3a3f81eb21","Type":"ContainerDied","Data":"de7231651223aa6d09ea4b703fd5feb347d9652c7f13fcf1ef8e1e737c642618"} Nov 28 13:39:43 crc kubenswrapper[4631]: I1128 13:39:43.304078 4631 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="de7231651223aa6d09ea4b703fd5feb347d9652c7f13fcf1ef8e1e737c642618" Nov 28 13:39:43 crc kubenswrapper[4631]: I1128 13:39:43.304121 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-0720-account-create-update-fhl44" Nov 28 13:39:43 crc kubenswrapper[4631]: I1128 13:39:43.310674 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-5cd3-account-create-update-hrtjp" event={"ID":"0f61d189-17d8-4154-ac95-df78d4042055","Type":"ContainerDied","Data":"0a32d7ae2c3a534184dba524af21eb67d2088a5219dc164fd35b2643f3e54ddd"} Nov 28 13:39:43 crc kubenswrapper[4631]: I1128 13:39:43.310713 4631 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0a32d7ae2c3a534184dba524af21eb67d2088a5219dc164fd35b2643f3e54ddd" Nov 28 13:39:43 crc kubenswrapper[4631]: I1128 13:39:43.310779 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-5cd3-account-create-update-hrtjp" Nov 28 13:39:43 crc kubenswrapper[4631]: I1128 13:39:43.328143 4631 generic.go:334] "Generic (PLEG): container finished" podID="d511d3b7-eff1-45f3-ab81-0f71b4324dc7" containerID="75b00bb29802e3263f0954877c7012663f5f708e8959b1b05c30795cf3e7c879" exitCode=0 Nov 28 13:39:43 crc kubenswrapper[4631]: I1128 13:39:43.328183 4631 generic.go:334] "Generic (PLEG): container finished" podID="d511d3b7-eff1-45f3-ab81-0f71b4324dc7" containerID="8445f38e73c31d3d4d1042303914d59653deadd33acca0865c3b7269356144b8" exitCode=2 Nov 28 13:39:43 crc kubenswrapper[4631]: I1128 13:39:43.328191 4631 generic.go:334] "Generic (PLEG): container finished" podID="d511d3b7-eff1-45f3-ab81-0f71b4324dc7" containerID="c452f408f264bcc78b2e6824e66bfabf1f3b5b5fb29219ddb376b76dc921bb99" exitCode=0 Nov 28 13:39:43 crc kubenswrapper[4631]: I1128 13:39:43.328199 4631 generic.go:334] "Generic (PLEG): container finished" podID="d511d3b7-eff1-45f3-ab81-0f71b4324dc7" containerID="978325b49b1129c5ccb877420d3040e957d3e5e744cc5b5115cc7d1c5668e094" exitCode=0 Nov 28 13:39:43 crc kubenswrapper[4631]: I1128 13:39:43.328266 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d511d3b7-eff1-45f3-ab81-0f71b4324dc7","Type":"ContainerDied","Data":"75b00bb29802e3263f0954877c7012663f5f708e8959b1b05c30795cf3e7c879"} Nov 28 13:39:43 crc kubenswrapper[4631]: I1128 13:39:43.328589 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d511d3b7-eff1-45f3-ab81-0f71b4324dc7","Type":"ContainerDied","Data":"8445f38e73c31d3d4d1042303914d59653deadd33acca0865c3b7269356144b8"} Nov 28 13:39:43 crc kubenswrapper[4631]: I1128 13:39:43.328604 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d511d3b7-eff1-45f3-ab81-0f71b4324dc7","Type":"ContainerDied","Data":"c452f408f264bcc78b2e6824e66bfabf1f3b5b5fb29219ddb376b76dc921bb99"} Nov 28 13:39:43 crc kubenswrapper[4631]: I1128 13:39:43.328614 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d511d3b7-eff1-45f3-ab81-0f71b4324dc7","Type":"ContainerDied","Data":"978325b49b1129c5ccb877420d3040e957d3e5e744cc5b5115cc7d1c5668e094"} Nov 28 13:39:43 crc kubenswrapper[4631]: I1128 13:39:43.332614 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-sqgzw" Nov 28 13:39:43 crc kubenswrapper[4631]: I1128 13:39:43.332970 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-sqgzw" event={"ID":"325fdb2e-59f8-4d94-860f-3ae7bd2e231e","Type":"ContainerDied","Data":"4d5022c831b0fef5c476ac9ef9ab45cacf078c669990e0396599fb7965acc187"} Nov 28 13:39:43 crc kubenswrapper[4631]: I1128 13:39:43.333002 4631 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4d5022c831b0fef5c476ac9ef9ab45cacf078c669990e0396599fb7965acc187" Nov 28 13:39:43 crc kubenswrapper[4631]: I1128 13:39:43.336263 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-fc8wj" event={"ID":"641b1d24-f22a-4d63-8e9e-d76c3dde2135","Type":"ContainerDied","Data":"3ddf2c9e1b460818b8085657a4e3cb74f75e0172d40feb83b851eb50ccd9d229"} Nov 28 13:39:43 crc kubenswrapper[4631]: I1128 13:39:43.336306 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-fc8wj" Nov 28 13:39:43 crc kubenswrapper[4631]: I1128 13:39:43.336312 4631 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3ddf2c9e1b460818b8085657a4e3cb74f75e0172d40feb83b851eb50ccd9d229" Nov 28 13:39:43 crc kubenswrapper[4631]: I1128 13:39:43.338464 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-x4t6b" event={"ID":"c5a28e08-ff16-4ecf-97ae-543c9de270f8","Type":"ContainerDied","Data":"534eff893a6ef7cafd7fb5b6bacbc255c1e4c9a1102bc06d9125061c3451f7c3"} Nov 28 13:39:43 crc kubenswrapper[4631]: I1128 13:39:43.338491 4631 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="534eff893a6ef7cafd7fb5b6bacbc255c1e4c9a1102bc06d9125061c3451f7c3" Nov 28 13:39:43 crc kubenswrapper[4631]: I1128 13:39:43.338530 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-x4t6b" Nov 28 13:39:43 crc kubenswrapper[4631]: I1128 13:39:43.341609 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 28 13:39:43 crc kubenswrapper[4631]: I1128 13:39:43.399554 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d511d3b7-eff1-45f3-ab81-0f71b4324dc7-config-data\") pod \"d511d3b7-eff1-45f3-ab81-0f71b4324dc7\" (UID: \"d511d3b7-eff1-45f3-ab81-0f71b4324dc7\") " Nov 28 13:39:43 crc kubenswrapper[4631]: I1128 13:39:43.399627 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vkxp8\" (UniqueName: \"kubernetes.io/projected/c5a28e08-ff16-4ecf-97ae-543c9de270f8-kube-api-access-vkxp8\") pod \"c5a28e08-ff16-4ecf-97ae-543c9de270f8\" (UID: \"c5a28e08-ff16-4ecf-97ae-543c9de270f8\") " Nov 28 13:39:43 crc kubenswrapper[4631]: I1128 13:39:43.399658 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d511d3b7-eff1-45f3-ab81-0f71b4324dc7-log-httpd\") pod \"d511d3b7-eff1-45f3-ab81-0f71b4324dc7\" (UID: \"d511d3b7-eff1-45f3-ab81-0f71b4324dc7\") " Nov 28 13:39:43 crc kubenswrapper[4631]: I1128 13:39:43.399695 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5rkjw\" (UniqueName: \"kubernetes.io/projected/641b1d24-f22a-4d63-8e9e-d76c3dde2135-kube-api-access-5rkjw\") pod \"641b1d24-f22a-4d63-8e9e-d76c3dde2135\" (UID: \"641b1d24-f22a-4d63-8e9e-d76c3dde2135\") " Nov 28 13:39:43 crc kubenswrapper[4631]: I1128 13:39:43.399737 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d511d3b7-eff1-45f3-ab81-0f71b4324dc7-combined-ca-bundle\") pod \"d511d3b7-eff1-45f3-ab81-0f71b4324dc7\" (UID: \"d511d3b7-eff1-45f3-ab81-0f71b4324dc7\") " Nov 28 13:39:43 crc kubenswrapper[4631]: I1128 13:39:43.399764 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d511d3b7-eff1-45f3-ab81-0f71b4324dc7-sg-core-conf-yaml\") pod \"d511d3b7-eff1-45f3-ab81-0f71b4324dc7\" (UID: \"d511d3b7-eff1-45f3-ab81-0f71b4324dc7\") " Nov 28 13:39:43 crc kubenswrapper[4631]: I1128 13:39:43.399812 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c5a28e08-ff16-4ecf-97ae-543c9de270f8-operator-scripts\") pod \"c5a28e08-ff16-4ecf-97ae-543c9de270f8\" (UID: \"c5a28e08-ff16-4ecf-97ae-543c9de270f8\") " Nov 28 13:39:43 crc kubenswrapper[4631]: I1128 13:39:43.399877 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/98edb37a-2698-4c30-a397-70d83441063d-operator-scripts\") pod \"98edb37a-2698-4c30-a397-70d83441063d\" (UID: \"98edb37a-2698-4c30-a397-70d83441063d\") " Nov 28 13:39:43 crc kubenswrapper[4631]: I1128 13:39:43.399958 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0f61d189-17d8-4154-ac95-df78d4042055-operator-scripts\") pod \"0f61d189-17d8-4154-ac95-df78d4042055\" (UID: \"0f61d189-17d8-4154-ac95-df78d4042055\") " Nov 28 13:39:43 crc kubenswrapper[4631]: I1128 13:39:43.399997 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mkt2j\" (UniqueName: \"kubernetes.io/projected/98edb37a-2698-4c30-a397-70d83441063d-kube-api-access-mkt2j\") pod \"98edb37a-2698-4c30-a397-70d83441063d\" (UID: \"98edb37a-2698-4c30-a397-70d83441063d\") " Nov 28 13:39:43 crc kubenswrapper[4631]: I1128 13:39:43.400013 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d9h2l\" (UniqueName: \"kubernetes.io/projected/d511d3b7-eff1-45f3-ab81-0f71b4324dc7-kube-api-access-d9h2l\") pod \"d511d3b7-eff1-45f3-ab81-0f71b4324dc7\" (UID: \"d511d3b7-eff1-45f3-ab81-0f71b4324dc7\") " Nov 28 13:39:43 crc kubenswrapper[4631]: I1128 13:39:43.400149 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5a1eb7ab-04ab-476e-aa72-2a3a3f81eb21-operator-scripts\") pod \"5a1eb7ab-04ab-476e-aa72-2a3a3f81eb21\" (UID: \"5a1eb7ab-04ab-476e-aa72-2a3a3f81eb21\") " Nov 28 13:39:43 crc kubenswrapper[4631]: I1128 13:39:43.400181 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d511d3b7-eff1-45f3-ab81-0f71b4324dc7-run-httpd\") pod \"d511d3b7-eff1-45f3-ab81-0f71b4324dc7\" (UID: \"d511d3b7-eff1-45f3-ab81-0f71b4324dc7\") " Nov 28 13:39:43 crc kubenswrapper[4631]: I1128 13:39:43.400201 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hks8v\" (UniqueName: \"kubernetes.io/projected/5a1eb7ab-04ab-476e-aa72-2a3a3f81eb21-kube-api-access-hks8v\") pod \"5a1eb7ab-04ab-476e-aa72-2a3a3f81eb21\" (UID: \"5a1eb7ab-04ab-476e-aa72-2a3a3f81eb21\") " Nov 28 13:39:43 crc kubenswrapper[4631]: I1128 13:39:43.400301 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d511d3b7-eff1-45f3-ab81-0f71b4324dc7-scripts\") pod \"d511d3b7-eff1-45f3-ab81-0f71b4324dc7\" (UID: \"d511d3b7-eff1-45f3-ab81-0f71b4324dc7\") " Nov 28 13:39:43 crc kubenswrapper[4631]: I1128 13:39:43.400334 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z2lbg\" (UniqueName: \"kubernetes.io/projected/0f61d189-17d8-4154-ac95-df78d4042055-kube-api-access-z2lbg\") pod \"0f61d189-17d8-4154-ac95-df78d4042055\" (UID: \"0f61d189-17d8-4154-ac95-df78d4042055\") " Nov 28 13:39:43 crc kubenswrapper[4631]: I1128 13:39:43.400390 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/641b1d24-f22a-4d63-8e9e-d76c3dde2135-operator-scripts\") pod \"641b1d24-f22a-4d63-8e9e-d76c3dde2135\" (UID: \"641b1d24-f22a-4d63-8e9e-d76c3dde2135\") " Nov 28 13:39:43 crc kubenswrapper[4631]: I1128 13:39:43.400424 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d511d3b7-eff1-45f3-ab81-0f71b4324dc7-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "d511d3b7-eff1-45f3-ab81-0f71b4324dc7" (UID: "d511d3b7-eff1-45f3-ab81-0f71b4324dc7"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 13:39:43 crc kubenswrapper[4631]: I1128 13:39:43.400859 4631 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d511d3b7-eff1-45f3-ab81-0f71b4324dc7-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 28 13:39:43 crc kubenswrapper[4631]: I1128 13:39:43.411502 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c5a28e08-ff16-4ecf-97ae-543c9de270f8-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "c5a28e08-ff16-4ecf-97ae-543c9de270f8" (UID: "c5a28e08-ff16-4ecf-97ae-543c9de270f8"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 13:39:43 crc kubenswrapper[4631]: I1128 13:39:43.412844 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c5a28e08-ff16-4ecf-97ae-543c9de270f8-kube-api-access-vkxp8" (OuterVolumeSpecName: "kube-api-access-vkxp8") pod "c5a28e08-ff16-4ecf-97ae-543c9de270f8" (UID: "c5a28e08-ff16-4ecf-97ae-543c9de270f8"). InnerVolumeSpecName "kube-api-access-vkxp8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:39:43 crc kubenswrapper[4631]: I1128 13:39:43.413599 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0f61d189-17d8-4154-ac95-df78d4042055-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "0f61d189-17d8-4154-ac95-df78d4042055" (UID: "0f61d189-17d8-4154-ac95-df78d4042055"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 13:39:43 crc kubenswrapper[4631]: I1128 13:39:43.414209 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5a1eb7ab-04ab-476e-aa72-2a3a3f81eb21-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "5a1eb7ab-04ab-476e-aa72-2a3a3f81eb21" (UID: "5a1eb7ab-04ab-476e-aa72-2a3a3f81eb21"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 13:39:43 crc kubenswrapper[4631]: I1128 13:39:43.414565 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d511d3b7-eff1-45f3-ab81-0f71b4324dc7-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "d511d3b7-eff1-45f3-ab81-0f71b4324dc7" (UID: "d511d3b7-eff1-45f3-ab81-0f71b4324dc7"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 13:39:43 crc kubenswrapper[4631]: I1128 13:39:43.414746 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/98edb37a-2698-4c30-a397-70d83441063d-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "98edb37a-2698-4c30-a397-70d83441063d" (UID: "98edb37a-2698-4c30-a397-70d83441063d"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 13:39:43 crc kubenswrapper[4631]: I1128 13:39:43.414782 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/641b1d24-f22a-4d63-8e9e-d76c3dde2135-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "641b1d24-f22a-4d63-8e9e-d76c3dde2135" (UID: "641b1d24-f22a-4d63-8e9e-d76c3dde2135"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 13:39:43 crc kubenswrapper[4631]: I1128 13:39:43.420584 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d511d3b7-eff1-45f3-ab81-0f71b4324dc7-kube-api-access-d9h2l" (OuterVolumeSpecName: "kube-api-access-d9h2l") pod "d511d3b7-eff1-45f3-ab81-0f71b4324dc7" (UID: "d511d3b7-eff1-45f3-ab81-0f71b4324dc7"). InnerVolumeSpecName "kube-api-access-d9h2l". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:39:43 crc kubenswrapper[4631]: I1128 13:39:43.422067 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/641b1d24-f22a-4d63-8e9e-d76c3dde2135-kube-api-access-5rkjw" (OuterVolumeSpecName: "kube-api-access-5rkjw") pod "641b1d24-f22a-4d63-8e9e-d76c3dde2135" (UID: "641b1d24-f22a-4d63-8e9e-d76c3dde2135"). InnerVolumeSpecName "kube-api-access-5rkjw". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:39:43 crc kubenswrapper[4631]: I1128 13:39:43.422172 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/98edb37a-2698-4c30-a397-70d83441063d-kube-api-access-mkt2j" (OuterVolumeSpecName: "kube-api-access-mkt2j") pod "98edb37a-2698-4c30-a397-70d83441063d" (UID: "98edb37a-2698-4c30-a397-70d83441063d"). InnerVolumeSpecName "kube-api-access-mkt2j". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:39:43 crc kubenswrapper[4631]: I1128 13:39:43.422342 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d511d3b7-eff1-45f3-ab81-0f71b4324dc7-scripts" (OuterVolumeSpecName: "scripts") pod "d511d3b7-eff1-45f3-ab81-0f71b4324dc7" (UID: "d511d3b7-eff1-45f3-ab81-0f71b4324dc7"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:39:43 crc kubenswrapper[4631]: I1128 13:39:43.423605 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5a1eb7ab-04ab-476e-aa72-2a3a3f81eb21-kube-api-access-hks8v" (OuterVolumeSpecName: "kube-api-access-hks8v") pod "5a1eb7ab-04ab-476e-aa72-2a3a3f81eb21" (UID: "5a1eb7ab-04ab-476e-aa72-2a3a3f81eb21"). InnerVolumeSpecName "kube-api-access-hks8v". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:39:43 crc kubenswrapper[4631]: I1128 13:39:43.428033 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0f61d189-17d8-4154-ac95-df78d4042055-kube-api-access-z2lbg" (OuterVolumeSpecName: "kube-api-access-z2lbg") pod "0f61d189-17d8-4154-ac95-df78d4042055" (UID: "0f61d189-17d8-4154-ac95-df78d4042055"). InnerVolumeSpecName "kube-api-access-z2lbg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:39:43 crc kubenswrapper[4631]: I1128 13:39:43.473262 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d511d3b7-eff1-45f3-ab81-0f71b4324dc7-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "d511d3b7-eff1-45f3-ab81-0f71b4324dc7" (UID: "d511d3b7-eff1-45f3-ab81-0f71b4324dc7"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:39:43 crc kubenswrapper[4631]: I1128 13:39:43.503028 4631 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d511d3b7-eff1-45f3-ab81-0f71b4324dc7-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 28 13:39:43 crc kubenswrapper[4631]: I1128 13:39:43.503079 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hks8v\" (UniqueName: \"kubernetes.io/projected/5a1eb7ab-04ab-476e-aa72-2a3a3f81eb21-kube-api-access-hks8v\") on node \"crc\" DevicePath \"\"" Nov 28 13:39:43 crc kubenswrapper[4631]: I1128 13:39:43.503094 4631 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d511d3b7-eff1-45f3-ab81-0f71b4324dc7-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 13:39:43 crc kubenswrapper[4631]: I1128 13:39:43.503106 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z2lbg\" (UniqueName: \"kubernetes.io/projected/0f61d189-17d8-4154-ac95-df78d4042055-kube-api-access-z2lbg\") on node \"crc\" DevicePath \"\"" Nov 28 13:39:43 crc kubenswrapper[4631]: I1128 13:39:43.503118 4631 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/641b1d24-f22a-4d63-8e9e-d76c3dde2135-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 13:39:43 crc kubenswrapper[4631]: I1128 13:39:43.503133 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vkxp8\" (UniqueName: \"kubernetes.io/projected/c5a28e08-ff16-4ecf-97ae-543c9de270f8-kube-api-access-vkxp8\") on node \"crc\" DevicePath \"\"" Nov 28 13:39:43 crc kubenswrapper[4631]: I1128 13:39:43.503145 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5rkjw\" (UniqueName: \"kubernetes.io/projected/641b1d24-f22a-4d63-8e9e-d76c3dde2135-kube-api-access-5rkjw\") on node \"crc\" DevicePath \"\"" Nov 28 13:39:43 crc kubenswrapper[4631]: I1128 13:39:43.503156 4631 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d511d3b7-eff1-45f3-ab81-0f71b4324dc7-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 28 13:39:43 crc kubenswrapper[4631]: I1128 13:39:43.503167 4631 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c5a28e08-ff16-4ecf-97ae-543c9de270f8-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 13:39:43 crc kubenswrapper[4631]: I1128 13:39:43.503180 4631 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/98edb37a-2698-4c30-a397-70d83441063d-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 13:39:43 crc kubenswrapper[4631]: I1128 13:39:43.503191 4631 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0f61d189-17d8-4154-ac95-df78d4042055-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 13:39:43 crc kubenswrapper[4631]: I1128 13:39:43.503204 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mkt2j\" (UniqueName: \"kubernetes.io/projected/98edb37a-2698-4c30-a397-70d83441063d-kube-api-access-mkt2j\") on node \"crc\" DevicePath \"\"" Nov 28 13:39:43 crc kubenswrapper[4631]: I1128 13:39:43.503215 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d9h2l\" (UniqueName: \"kubernetes.io/projected/d511d3b7-eff1-45f3-ab81-0f71b4324dc7-kube-api-access-d9h2l\") on node \"crc\" DevicePath \"\"" Nov 28 13:39:43 crc kubenswrapper[4631]: I1128 13:39:43.503229 4631 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5a1eb7ab-04ab-476e-aa72-2a3a3f81eb21-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 13:39:43 crc kubenswrapper[4631]: I1128 13:39:43.533951 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d511d3b7-eff1-45f3-ab81-0f71b4324dc7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d511d3b7-eff1-45f3-ab81-0f71b4324dc7" (UID: "d511d3b7-eff1-45f3-ab81-0f71b4324dc7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:39:43 crc kubenswrapper[4631]: I1128 13:39:43.536264 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d511d3b7-eff1-45f3-ab81-0f71b4324dc7-config-data" (OuterVolumeSpecName: "config-data") pod "d511d3b7-eff1-45f3-ab81-0f71b4324dc7" (UID: "d511d3b7-eff1-45f3-ab81-0f71b4324dc7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:39:43 crc kubenswrapper[4631]: I1128 13:39:43.605816 4631 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d511d3b7-eff1-45f3-ab81-0f71b4324dc7-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 13:39:43 crc kubenswrapper[4631]: I1128 13:39:43.606282 4631 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d511d3b7-eff1-45f3-ab81-0f71b4324dc7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 13:39:44 crc kubenswrapper[4631]: I1128 13:39:44.354693 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d511d3b7-eff1-45f3-ab81-0f71b4324dc7","Type":"ContainerDied","Data":"18317c37a5634a2985b1a09f1f859abc40b0a1ad1a849fed91b80714aa2742c7"} Nov 28 13:39:44 crc kubenswrapper[4631]: I1128 13:39:44.354764 4631 scope.go:117] "RemoveContainer" containerID="75b00bb29802e3263f0954877c7012663f5f708e8959b1b05c30795cf3e7c879" Nov 28 13:39:44 crc kubenswrapper[4631]: I1128 13:39:44.354937 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 28 13:39:44 crc kubenswrapper[4631]: I1128 13:39:44.400452 4631 scope.go:117] "RemoveContainer" containerID="8445f38e73c31d3d4d1042303914d59653deadd33acca0865c3b7269356144b8" Nov 28 13:39:44 crc kubenswrapper[4631]: I1128 13:39:44.435744 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 28 13:39:44 crc kubenswrapper[4631]: I1128 13:39:44.445008 4631 scope.go:117] "RemoveContainer" containerID="c452f408f264bcc78b2e6824e66bfabf1f3b5b5fb29219ddb376b76dc921bb99" Nov 28 13:39:44 crc kubenswrapper[4631]: I1128 13:39:44.478161 4631 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 28 13:39:44 crc kubenswrapper[4631]: I1128 13:39:44.493014 4631 scope.go:117] "RemoveContainer" containerID="978325b49b1129c5ccb877420d3040e957d3e5e744cc5b5115cc7d1c5668e094" Nov 28 13:39:44 crc kubenswrapper[4631]: I1128 13:39:44.506403 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 28 13:39:44 crc kubenswrapper[4631]: E1128 13:39:44.506982 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="98edb37a-2698-4c30-a397-70d83441063d" containerName="mariadb-account-create-update" Nov 28 13:39:44 crc kubenswrapper[4631]: I1128 13:39:44.507005 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="98edb37a-2698-4c30-a397-70d83441063d" containerName="mariadb-account-create-update" Nov 28 13:39:44 crc kubenswrapper[4631]: E1128 13:39:44.507023 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d511d3b7-eff1-45f3-ab81-0f71b4324dc7" containerName="ceilometer-notification-agent" Nov 28 13:39:44 crc kubenswrapper[4631]: I1128 13:39:44.507031 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="d511d3b7-eff1-45f3-ab81-0f71b4324dc7" containerName="ceilometer-notification-agent" Nov 28 13:39:44 crc kubenswrapper[4631]: E1128 13:39:44.507042 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="641b1d24-f22a-4d63-8e9e-d76c3dde2135" containerName="mariadb-database-create" Nov 28 13:39:44 crc kubenswrapper[4631]: I1128 13:39:44.507048 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="641b1d24-f22a-4d63-8e9e-d76c3dde2135" containerName="mariadb-database-create" Nov 28 13:39:44 crc kubenswrapper[4631]: E1128 13:39:44.507069 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f61d189-17d8-4154-ac95-df78d4042055" containerName="mariadb-account-create-update" Nov 28 13:39:44 crc kubenswrapper[4631]: I1128 13:39:44.507078 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f61d189-17d8-4154-ac95-df78d4042055" containerName="mariadb-account-create-update" Nov 28 13:39:44 crc kubenswrapper[4631]: E1128 13:39:44.507086 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5a1eb7ab-04ab-476e-aa72-2a3a3f81eb21" containerName="mariadb-account-create-update" Nov 28 13:39:44 crc kubenswrapper[4631]: I1128 13:39:44.507092 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="5a1eb7ab-04ab-476e-aa72-2a3a3f81eb21" containerName="mariadb-account-create-update" Nov 28 13:39:44 crc kubenswrapper[4631]: E1128 13:39:44.507103 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d511d3b7-eff1-45f3-ab81-0f71b4324dc7" containerName="ceilometer-central-agent" Nov 28 13:39:44 crc kubenswrapper[4631]: I1128 13:39:44.507109 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="d511d3b7-eff1-45f3-ab81-0f71b4324dc7" containerName="ceilometer-central-agent" Nov 28 13:39:44 crc kubenswrapper[4631]: E1128 13:39:44.507121 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c5a28e08-ff16-4ecf-97ae-543c9de270f8" containerName="mariadb-database-create" Nov 28 13:39:44 crc kubenswrapper[4631]: I1128 13:39:44.507129 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="c5a28e08-ff16-4ecf-97ae-543c9de270f8" containerName="mariadb-database-create" Nov 28 13:39:44 crc kubenswrapper[4631]: E1128 13:39:44.507149 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="325fdb2e-59f8-4d94-860f-3ae7bd2e231e" containerName="mariadb-database-create" Nov 28 13:39:44 crc kubenswrapper[4631]: I1128 13:39:44.507157 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="325fdb2e-59f8-4d94-860f-3ae7bd2e231e" containerName="mariadb-database-create" Nov 28 13:39:44 crc kubenswrapper[4631]: E1128 13:39:44.507170 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d511d3b7-eff1-45f3-ab81-0f71b4324dc7" containerName="sg-core" Nov 28 13:39:44 crc kubenswrapper[4631]: I1128 13:39:44.507181 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="d511d3b7-eff1-45f3-ab81-0f71b4324dc7" containerName="sg-core" Nov 28 13:39:44 crc kubenswrapper[4631]: E1128 13:39:44.507192 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d511d3b7-eff1-45f3-ab81-0f71b4324dc7" containerName="proxy-httpd" Nov 28 13:39:44 crc kubenswrapper[4631]: I1128 13:39:44.507199 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="d511d3b7-eff1-45f3-ab81-0f71b4324dc7" containerName="proxy-httpd" Nov 28 13:39:44 crc kubenswrapper[4631]: I1128 13:39:44.507426 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="98edb37a-2698-4c30-a397-70d83441063d" containerName="mariadb-account-create-update" Nov 28 13:39:44 crc kubenswrapper[4631]: I1128 13:39:44.507441 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="c5a28e08-ff16-4ecf-97ae-543c9de270f8" containerName="mariadb-database-create" Nov 28 13:39:44 crc kubenswrapper[4631]: I1128 13:39:44.507453 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="d511d3b7-eff1-45f3-ab81-0f71b4324dc7" containerName="proxy-httpd" Nov 28 13:39:44 crc kubenswrapper[4631]: I1128 13:39:44.507464 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="5a1eb7ab-04ab-476e-aa72-2a3a3f81eb21" containerName="mariadb-account-create-update" Nov 28 13:39:44 crc kubenswrapper[4631]: I1128 13:39:44.507479 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="641b1d24-f22a-4d63-8e9e-d76c3dde2135" containerName="mariadb-database-create" Nov 28 13:39:44 crc kubenswrapper[4631]: I1128 13:39:44.507494 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="d511d3b7-eff1-45f3-ab81-0f71b4324dc7" containerName="sg-core" Nov 28 13:39:44 crc kubenswrapper[4631]: I1128 13:39:44.507507 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="d511d3b7-eff1-45f3-ab81-0f71b4324dc7" containerName="ceilometer-notification-agent" Nov 28 13:39:44 crc kubenswrapper[4631]: I1128 13:39:44.507514 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="d511d3b7-eff1-45f3-ab81-0f71b4324dc7" containerName="ceilometer-central-agent" Nov 28 13:39:44 crc kubenswrapper[4631]: I1128 13:39:44.507523 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="325fdb2e-59f8-4d94-860f-3ae7bd2e231e" containerName="mariadb-database-create" Nov 28 13:39:44 crc kubenswrapper[4631]: I1128 13:39:44.507531 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="0f61d189-17d8-4154-ac95-df78d4042055" containerName="mariadb-account-create-update" Nov 28 13:39:44 crc kubenswrapper[4631]: I1128 13:39:44.509521 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 28 13:39:44 crc kubenswrapper[4631]: I1128 13:39:44.537055 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 28 13:39:44 crc kubenswrapper[4631]: I1128 13:39:44.537823 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 28 13:39:44 crc kubenswrapper[4631]: I1128 13:39:44.538022 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 28 13:39:44 crc kubenswrapper[4631]: I1128 13:39:44.542763 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/736b1888-f3ac-46f6-9532-cdd4d4e310c8-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"736b1888-f3ac-46f6-9532-cdd4d4e310c8\") " pod="openstack/ceilometer-0" Nov 28 13:39:44 crc kubenswrapper[4631]: I1128 13:39:44.542875 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hrdmr\" (UniqueName: \"kubernetes.io/projected/736b1888-f3ac-46f6-9532-cdd4d4e310c8-kube-api-access-hrdmr\") pod \"ceilometer-0\" (UID: \"736b1888-f3ac-46f6-9532-cdd4d4e310c8\") " pod="openstack/ceilometer-0" Nov 28 13:39:44 crc kubenswrapper[4631]: I1128 13:39:44.542906 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/736b1888-f3ac-46f6-9532-cdd4d4e310c8-scripts\") pod \"ceilometer-0\" (UID: \"736b1888-f3ac-46f6-9532-cdd4d4e310c8\") " pod="openstack/ceilometer-0" Nov 28 13:39:44 crc kubenswrapper[4631]: I1128 13:39:44.542938 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/736b1888-f3ac-46f6-9532-cdd4d4e310c8-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"736b1888-f3ac-46f6-9532-cdd4d4e310c8\") " pod="openstack/ceilometer-0" Nov 28 13:39:44 crc kubenswrapper[4631]: I1128 13:39:44.542996 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/736b1888-f3ac-46f6-9532-cdd4d4e310c8-run-httpd\") pod \"ceilometer-0\" (UID: \"736b1888-f3ac-46f6-9532-cdd4d4e310c8\") " pod="openstack/ceilometer-0" Nov 28 13:39:44 crc kubenswrapper[4631]: I1128 13:39:44.543020 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/736b1888-f3ac-46f6-9532-cdd4d4e310c8-config-data\") pod \"ceilometer-0\" (UID: \"736b1888-f3ac-46f6-9532-cdd4d4e310c8\") " pod="openstack/ceilometer-0" Nov 28 13:39:44 crc kubenswrapper[4631]: I1128 13:39:44.543045 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/736b1888-f3ac-46f6-9532-cdd4d4e310c8-log-httpd\") pod \"ceilometer-0\" (UID: \"736b1888-f3ac-46f6-9532-cdd4d4e310c8\") " pod="openstack/ceilometer-0" Nov 28 13:39:44 crc kubenswrapper[4631]: I1128 13:39:44.647247 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hrdmr\" (UniqueName: \"kubernetes.io/projected/736b1888-f3ac-46f6-9532-cdd4d4e310c8-kube-api-access-hrdmr\") pod \"ceilometer-0\" (UID: \"736b1888-f3ac-46f6-9532-cdd4d4e310c8\") " pod="openstack/ceilometer-0" Nov 28 13:39:44 crc kubenswrapper[4631]: I1128 13:39:44.647677 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/736b1888-f3ac-46f6-9532-cdd4d4e310c8-scripts\") pod \"ceilometer-0\" (UID: \"736b1888-f3ac-46f6-9532-cdd4d4e310c8\") " pod="openstack/ceilometer-0" Nov 28 13:39:44 crc kubenswrapper[4631]: I1128 13:39:44.647952 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/736b1888-f3ac-46f6-9532-cdd4d4e310c8-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"736b1888-f3ac-46f6-9532-cdd4d4e310c8\") " pod="openstack/ceilometer-0" Nov 28 13:39:44 crc kubenswrapper[4631]: I1128 13:39:44.648115 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/736b1888-f3ac-46f6-9532-cdd4d4e310c8-run-httpd\") pod \"ceilometer-0\" (UID: \"736b1888-f3ac-46f6-9532-cdd4d4e310c8\") " pod="openstack/ceilometer-0" Nov 28 13:39:44 crc kubenswrapper[4631]: I1128 13:39:44.648227 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/736b1888-f3ac-46f6-9532-cdd4d4e310c8-config-data\") pod \"ceilometer-0\" (UID: \"736b1888-f3ac-46f6-9532-cdd4d4e310c8\") " pod="openstack/ceilometer-0" Nov 28 13:39:44 crc kubenswrapper[4631]: I1128 13:39:44.648381 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/736b1888-f3ac-46f6-9532-cdd4d4e310c8-log-httpd\") pod \"ceilometer-0\" (UID: \"736b1888-f3ac-46f6-9532-cdd4d4e310c8\") " pod="openstack/ceilometer-0" Nov 28 13:39:44 crc kubenswrapper[4631]: I1128 13:39:44.648580 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/736b1888-f3ac-46f6-9532-cdd4d4e310c8-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"736b1888-f3ac-46f6-9532-cdd4d4e310c8\") " pod="openstack/ceilometer-0" Nov 28 13:39:44 crc kubenswrapper[4631]: I1128 13:39:44.649132 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/736b1888-f3ac-46f6-9532-cdd4d4e310c8-run-httpd\") pod \"ceilometer-0\" (UID: \"736b1888-f3ac-46f6-9532-cdd4d4e310c8\") " pod="openstack/ceilometer-0" Nov 28 13:39:44 crc kubenswrapper[4631]: I1128 13:39:44.650189 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/736b1888-f3ac-46f6-9532-cdd4d4e310c8-log-httpd\") pod \"ceilometer-0\" (UID: \"736b1888-f3ac-46f6-9532-cdd4d4e310c8\") " pod="openstack/ceilometer-0" Nov 28 13:39:44 crc kubenswrapper[4631]: I1128 13:39:44.662516 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/736b1888-f3ac-46f6-9532-cdd4d4e310c8-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"736b1888-f3ac-46f6-9532-cdd4d4e310c8\") " pod="openstack/ceilometer-0" Nov 28 13:39:44 crc kubenswrapper[4631]: I1128 13:39:44.665288 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/736b1888-f3ac-46f6-9532-cdd4d4e310c8-scripts\") pod \"ceilometer-0\" (UID: \"736b1888-f3ac-46f6-9532-cdd4d4e310c8\") " pod="openstack/ceilometer-0" Nov 28 13:39:44 crc kubenswrapper[4631]: I1128 13:39:44.677056 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/736b1888-f3ac-46f6-9532-cdd4d4e310c8-config-data\") pod \"ceilometer-0\" (UID: \"736b1888-f3ac-46f6-9532-cdd4d4e310c8\") " pod="openstack/ceilometer-0" Nov 28 13:39:44 crc kubenswrapper[4631]: I1128 13:39:44.677811 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/736b1888-f3ac-46f6-9532-cdd4d4e310c8-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"736b1888-f3ac-46f6-9532-cdd4d4e310c8\") " pod="openstack/ceilometer-0" Nov 28 13:39:44 crc kubenswrapper[4631]: I1128 13:39:44.722519 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hrdmr\" (UniqueName: \"kubernetes.io/projected/736b1888-f3ac-46f6-9532-cdd4d4e310c8-kube-api-access-hrdmr\") pod \"ceilometer-0\" (UID: \"736b1888-f3ac-46f6-9532-cdd4d4e310c8\") " pod="openstack/ceilometer-0" Nov 28 13:39:44 crc kubenswrapper[4631]: I1128 13:39:44.851408 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 28 13:39:45 crc kubenswrapper[4631]: I1128 13:39:45.234561 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 28 13:39:45 crc kubenswrapper[4631]: I1128 13:39:45.369361 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"736b1888-f3ac-46f6-9532-cdd4d4e310c8","Type":"ContainerStarted","Data":"68b45d3ea18cc197789698afa6f17b0ce6989f5c4bf6441e21fd8482d8af8e3b"} Nov 28 13:39:45 crc kubenswrapper[4631]: I1128 13:39:45.541538 4631 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d511d3b7-eff1-45f3-ab81-0f71b4324dc7" path="/var/lib/kubelet/pods/d511d3b7-eff1-45f3-ab81-0f71b4324dc7/volumes" Nov 28 13:39:46 crc kubenswrapper[4631]: E1128 13:39:46.142701 4631 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0f61d189_17d8_4154_ac95_df78d4042055.slice/crio-0a32d7ae2c3a534184dba524af21eb67d2088a5219dc164fd35b2643f3e54ddd\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod641b1d24_f22a_4d63_8e9e_d76c3dde2135.slice\": RecentStats: unable to find data in memory cache]" Nov 28 13:39:46 crc kubenswrapper[4631]: I1128 13:39:46.385043 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"736b1888-f3ac-46f6-9532-cdd4d4e310c8","Type":"ContainerStarted","Data":"66fc9c2ad3bdcee59e76c4232c6d98384e6b554b3218f4acfe4ddbfc6e974aa5"} Nov 28 13:39:47 crc kubenswrapper[4631]: I1128 13:39:47.402658 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"736b1888-f3ac-46f6-9532-cdd4d4e310c8","Type":"ContainerStarted","Data":"1c14e41c0b95654690f096dcb82b6936bc505ab33653274e9bd24cce207423ec"} Nov 28 13:39:47 crc kubenswrapper[4631]: I1128 13:39:47.438364 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-tqj67"] Nov 28 13:39:47 crc kubenswrapper[4631]: I1128 13:39:47.439782 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-tqj67" Nov 28 13:39:47 crc kubenswrapper[4631]: I1128 13:39:47.442713 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Nov 28 13:39:47 crc kubenswrapper[4631]: I1128 13:39:47.442730 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Nov 28 13:39:47 crc kubenswrapper[4631]: I1128 13:39:47.449579 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-z77kh" Nov 28 13:39:47 crc kubenswrapper[4631]: I1128 13:39:47.480683 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-tqj67"] Nov 28 13:39:47 crc kubenswrapper[4631]: I1128 13:39:47.531854 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8a4fda29-ad66-4077-b2f9-796fe6fd961a-scripts\") pod \"nova-cell0-conductor-db-sync-tqj67\" (UID: \"8a4fda29-ad66-4077-b2f9-796fe6fd961a\") " pod="openstack/nova-cell0-conductor-db-sync-tqj67" Nov 28 13:39:47 crc kubenswrapper[4631]: I1128 13:39:47.531904 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a4fda29-ad66-4077-b2f9-796fe6fd961a-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-tqj67\" (UID: \"8a4fda29-ad66-4077-b2f9-796fe6fd961a\") " pod="openstack/nova-cell0-conductor-db-sync-tqj67" Nov 28 13:39:47 crc kubenswrapper[4631]: I1128 13:39:47.531995 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a4fda29-ad66-4077-b2f9-796fe6fd961a-config-data\") pod \"nova-cell0-conductor-db-sync-tqj67\" (UID: \"8a4fda29-ad66-4077-b2f9-796fe6fd961a\") " pod="openstack/nova-cell0-conductor-db-sync-tqj67" Nov 28 13:39:47 crc kubenswrapper[4631]: I1128 13:39:47.532017 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qwp6z\" (UniqueName: \"kubernetes.io/projected/8a4fda29-ad66-4077-b2f9-796fe6fd961a-kube-api-access-qwp6z\") pod \"nova-cell0-conductor-db-sync-tqj67\" (UID: \"8a4fda29-ad66-4077-b2f9-796fe6fd961a\") " pod="openstack/nova-cell0-conductor-db-sync-tqj67" Nov 28 13:39:47 crc kubenswrapper[4631]: I1128 13:39:47.633558 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a4fda29-ad66-4077-b2f9-796fe6fd961a-config-data\") pod \"nova-cell0-conductor-db-sync-tqj67\" (UID: \"8a4fda29-ad66-4077-b2f9-796fe6fd961a\") " pod="openstack/nova-cell0-conductor-db-sync-tqj67" Nov 28 13:39:47 crc kubenswrapper[4631]: I1128 13:39:47.633612 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qwp6z\" (UniqueName: \"kubernetes.io/projected/8a4fda29-ad66-4077-b2f9-796fe6fd961a-kube-api-access-qwp6z\") pod \"nova-cell0-conductor-db-sync-tqj67\" (UID: \"8a4fda29-ad66-4077-b2f9-796fe6fd961a\") " pod="openstack/nova-cell0-conductor-db-sync-tqj67" Nov 28 13:39:47 crc kubenswrapper[4631]: I1128 13:39:47.633687 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8a4fda29-ad66-4077-b2f9-796fe6fd961a-scripts\") pod \"nova-cell0-conductor-db-sync-tqj67\" (UID: \"8a4fda29-ad66-4077-b2f9-796fe6fd961a\") " pod="openstack/nova-cell0-conductor-db-sync-tqj67" Nov 28 13:39:47 crc kubenswrapper[4631]: I1128 13:39:47.633717 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a4fda29-ad66-4077-b2f9-796fe6fd961a-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-tqj67\" (UID: \"8a4fda29-ad66-4077-b2f9-796fe6fd961a\") " pod="openstack/nova-cell0-conductor-db-sync-tqj67" Nov 28 13:39:47 crc kubenswrapper[4631]: I1128 13:39:47.640355 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a4fda29-ad66-4077-b2f9-796fe6fd961a-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-tqj67\" (UID: \"8a4fda29-ad66-4077-b2f9-796fe6fd961a\") " pod="openstack/nova-cell0-conductor-db-sync-tqj67" Nov 28 13:39:47 crc kubenswrapper[4631]: I1128 13:39:47.644781 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a4fda29-ad66-4077-b2f9-796fe6fd961a-config-data\") pod \"nova-cell0-conductor-db-sync-tqj67\" (UID: \"8a4fda29-ad66-4077-b2f9-796fe6fd961a\") " pod="openstack/nova-cell0-conductor-db-sync-tqj67" Nov 28 13:39:47 crc kubenswrapper[4631]: I1128 13:39:47.658161 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8a4fda29-ad66-4077-b2f9-796fe6fd961a-scripts\") pod \"nova-cell0-conductor-db-sync-tqj67\" (UID: \"8a4fda29-ad66-4077-b2f9-796fe6fd961a\") " pod="openstack/nova-cell0-conductor-db-sync-tqj67" Nov 28 13:39:47 crc kubenswrapper[4631]: I1128 13:39:47.661511 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qwp6z\" (UniqueName: \"kubernetes.io/projected/8a4fda29-ad66-4077-b2f9-796fe6fd961a-kube-api-access-qwp6z\") pod \"nova-cell0-conductor-db-sync-tqj67\" (UID: \"8a4fda29-ad66-4077-b2f9-796fe6fd961a\") " pod="openstack/nova-cell0-conductor-db-sync-tqj67" Nov 28 13:39:47 crc kubenswrapper[4631]: I1128 13:39:47.760943 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-tqj67" Nov 28 13:39:48 crc kubenswrapper[4631]: I1128 13:39:48.274790 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-tqj67"] Nov 28 13:39:48 crc kubenswrapper[4631]: W1128 13:39:48.294875 4631 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8a4fda29_ad66_4077_b2f9_796fe6fd961a.slice/crio-1ea96f588712b0c65d6c2ca7037ca23fc287850635e3e2aa8b7beb94c25a076d WatchSource:0}: Error finding container 1ea96f588712b0c65d6c2ca7037ca23fc287850635e3e2aa8b7beb94c25a076d: Status 404 returned error can't find the container with id 1ea96f588712b0c65d6c2ca7037ca23fc287850635e3e2aa8b7beb94c25a076d Nov 28 13:39:48 crc kubenswrapper[4631]: I1128 13:39:48.413752 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"736b1888-f3ac-46f6-9532-cdd4d4e310c8","Type":"ContainerStarted","Data":"a9279936337bbc804fdbffba98e5a5ea1aa3d8c892c3df4653c010e0721d4ade"} Nov 28 13:39:48 crc kubenswrapper[4631]: I1128 13:39:48.415161 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-tqj67" event={"ID":"8a4fda29-ad66-4077-b2f9-796fe6fd961a","Type":"ContainerStarted","Data":"1ea96f588712b0c65d6c2ca7037ca23fc287850635e3e2aa8b7beb94c25a076d"} Nov 28 13:39:49 crc kubenswrapper[4631]: I1128 13:39:49.435312 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"736b1888-f3ac-46f6-9532-cdd4d4e310c8","Type":"ContainerStarted","Data":"38d193db6b1a62df309e687fdbd62664a103f1d506c32337a40560ad416f0368"} Nov 28 13:39:49 crc kubenswrapper[4631]: I1128 13:39:49.436245 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 28 13:39:49 crc kubenswrapper[4631]: I1128 13:39:49.488921 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.8730124479999999 podStartE2EDuration="5.48889336s" podCreationTimestamp="2025-11-28 13:39:44 +0000 UTC" firstStartedPulling="2025-11-28 13:39:45.23511729 +0000 UTC m=+1142.042420634" lastFinishedPulling="2025-11-28 13:39:48.850998212 +0000 UTC m=+1145.658301546" observedRunningTime="2025-11-28 13:39:49.464877575 +0000 UTC m=+1146.272180929" watchObservedRunningTime="2025-11-28 13:39:49.48889336 +0000 UTC m=+1146.296196704" Nov 28 13:39:55 crc kubenswrapper[4631]: I1128 13:39:55.559224 4631 generic.go:334] "Generic (PLEG): container finished" podID="111c5349-528d-4856-a3e4-1277e3b10889" containerID="ddca46c54a9ba33270a5c14213a03ffc4ee751ec03afaccc3b37cfa0eac87eb7" exitCode=137 Nov 28 13:39:55 crc kubenswrapper[4631]: I1128 13:39:55.559383 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-64f8fdf97b-jn5mg" event={"ID":"111c5349-528d-4856-a3e4-1277e3b10889","Type":"ContainerDied","Data":"ddca46c54a9ba33270a5c14213a03ffc4ee751ec03afaccc3b37cfa0eac87eb7"} Nov 28 13:39:55 crc kubenswrapper[4631]: I1128 13:39:55.560071 4631 scope.go:117] "RemoveContainer" containerID="ffafadc828320f83436d72910a3912cda7f6bb048d8c4c01528f4529bf55b772" Nov 28 13:39:56 crc kubenswrapper[4631]: E1128 13:39:56.533533 4631 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod641b1d24_f22a_4d63_8e9e_d76c3dde2135.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0f61d189_17d8_4154_ac95_df78d4042055.slice/crio-0a32d7ae2c3a534184dba524af21eb67d2088a5219dc164fd35b2643f3e54ddd\": RecentStats: unable to find data in memory cache]" Nov 28 13:39:56 crc kubenswrapper[4631]: I1128 13:39:56.578062 4631 generic.go:334] "Generic (PLEG): container finished" podID="cd607dae-568a-4b81-af81-3310c0e95854" containerID="26b4a5d749292bcec07819e5de34fe07da049e5be75efa94d19e13fdfcfb064d" exitCode=137 Nov 28 13:39:56 crc kubenswrapper[4631]: I1128 13:39:56.578393 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-544496c768-rzhsw" event={"ID":"cd607dae-568a-4b81-af81-3310c0e95854","Type":"ContainerDied","Data":"26b4a5d749292bcec07819e5de34fe07da049e5be75efa94d19e13fdfcfb064d"} Nov 28 13:40:01 crc kubenswrapper[4631]: I1128 13:40:01.691552 4631 scope.go:117] "RemoveContainer" containerID="bd4617d23aba6bbc47167ded3acd30280726cfe208e8bfc920d75fb684ec48e8" Nov 28 13:40:02 crc kubenswrapper[4631]: I1128 13:40:02.652373 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-tqj67" event={"ID":"8a4fda29-ad66-4077-b2f9-796fe6fd961a","Type":"ContainerStarted","Data":"d9cdc53db79944ad3a916e49b0da1cb467958dc12bea86bc7c817bf6487bf6ff"} Nov 28 13:40:02 crc kubenswrapper[4631]: I1128 13:40:02.660570 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-544496c768-rzhsw" event={"ID":"cd607dae-568a-4b81-af81-3310c0e95854","Type":"ContainerStarted","Data":"966640930a7d2630d4ef6bbacd3ac941e83df7e177040964f7f5e2cd4806f6a4"} Nov 28 13:40:02 crc kubenswrapper[4631]: I1128 13:40:02.665189 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-64f8fdf97b-jn5mg" event={"ID":"111c5349-528d-4856-a3e4-1277e3b10889","Type":"ContainerStarted","Data":"3b826a5556ea3b27f6357d4cc74220f4bce293db6dbcb49fa5131013bef0e517"} Nov 28 13:40:02 crc kubenswrapper[4631]: I1128 13:40:02.678797 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-tqj67" podStartSLOduration=2.261827522 podStartE2EDuration="15.678765904s" podCreationTimestamp="2025-11-28 13:39:47 +0000 UTC" firstStartedPulling="2025-11-28 13:39:48.297372723 +0000 UTC m=+1145.104676067" lastFinishedPulling="2025-11-28 13:40:01.714311105 +0000 UTC m=+1158.521614449" observedRunningTime="2025-11-28 13:40:02.670981969 +0000 UTC m=+1159.478285323" watchObservedRunningTime="2025-11-28 13:40:02.678765904 +0000 UTC m=+1159.486069248" Nov 28 13:40:05 crc kubenswrapper[4631]: I1128 13:40:05.331437 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-64f8fdf97b-jn5mg" Nov 28 13:40:05 crc kubenswrapper[4631]: I1128 13:40:05.331970 4631 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-64f8fdf97b-jn5mg" Nov 28 13:40:05 crc kubenswrapper[4631]: I1128 13:40:05.553543 4631 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-544496c768-rzhsw" Nov 28 13:40:05 crc kubenswrapper[4631]: I1128 13:40:05.553630 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-544496c768-rzhsw" Nov 28 13:40:06 crc kubenswrapper[4631]: E1128 13:40:06.846736 4631 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod641b1d24_f22a_4d63_8e9e_d76c3dde2135.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0f61d189_17d8_4154_ac95_df78d4042055.slice/crio-0a32d7ae2c3a534184dba524af21eb67d2088a5219dc164fd35b2643f3e54ddd\": RecentStats: unable to find data in memory cache]" Nov 28 13:40:14 crc kubenswrapper[4631]: I1128 13:40:14.858374 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Nov 28 13:40:15 crc kubenswrapper[4631]: I1128 13:40:15.333508 4631 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-64f8fdf97b-jn5mg" podUID="111c5349-528d-4856-a3e4-1277e3b10889" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.146:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.146:8443: connect: connection refused" Nov 28 13:40:15 crc kubenswrapper[4631]: I1128 13:40:15.556016 4631 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-544496c768-rzhsw" podUID="cd607dae-568a-4b81-af81-3310c0e95854" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.147:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.147:8443: connect: connection refused" Nov 28 13:40:17 crc kubenswrapper[4631]: E1128 13:40:17.106616 4631 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod641b1d24_f22a_4d63_8e9e_d76c3dde2135.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0f61d189_17d8_4154_ac95_df78d4042055.slice/crio-0a32d7ae2c3a534184dba524af21eb67d2088a5219dc164fd35b2643f3e54ddd\": RecentStats: unable to find data in memory cache]" Nov 28 13:40:19 crc kubenswrapper[4631]: I1128 13:40:19.635282 4631 patch_prober.go:28] interesting pod/machine-config-daemon-47sxw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 28 13:40:19 crc kubenswrapper[4631]: I1128 13:40:19.637404 4631 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" podUID="8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 28 13:40:20 crc kubenswrapper[4631]: I1128 13:40:20.515029 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 28 13:40:20 crc kubenswrapper[4631]: I1128 13:40:20.515342 4631 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="83d77da3-fbb7-434f-abaf-5be9b599a9bf" containerName="kube-state-metrics" containerID="cri-o://52d0c55ae2c92fdf222cd8f5c0b2b93b9073d9c66e613db020f87d5dd63419b7" gracePeriod=30 Nov 28 13:40:20 crc kubenswrapper[4631]: I1128 13:40:20.889628 4631 generic.go:334] "Generic (PLEG): container finished" podID="83d77da3-fbb7-434f-abaf-5be9b599a9bf" containerID="52d0c55ae2c92fdf222cd8f5c0b2b93b9073d9c66e613db020f87d5dd63419b7" exitCode=2 Nov 28 13:40:20 crc kubenswrapper[4631]: I1128 13:40:20.889786 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"83d77da3-fbb7-434f-abaf-5be9b599a9bf","Type":"ContainerDied","Data":"52d0c55ae2c92fdf222cd8f5c0b2b93b9073d9c66e613db020f87d5dd63419b7"} Nov 28 13:40:21 crc kubenswrapper[4631]: I1128 13:40:21.222771 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 28 13:40:21 crc kubenswrapper[4631]: I1128 13:40:21.344210 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xj59v\" (UniqueName: \"kubernetes.io/projected/83d77da3-fbb7-434f-abaf-5be9b599a9bf-kube-api-access-xj59v\") pod \"83d77da3-fbb7-434f-abaf-5be9b599a9bf\" (UID: \"83d77da3-fbb7-434f-abaf-5be9b599a9bf\") " Nov 28 13:40:21 crc kubenswrapper[4631]: I1128 13:40:21.371037 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/83d77da3-fbb7-434f-abaf-5be9b599a9bf-kube-api-access-xj59v" (OuterVolumeSpecName: "kube-api-access-xj59v") pod "83d77da3-fbb7-434f-abaf-5be9b599a9bf" (UID: "83d77da3-fbb7-434f-abaf-5be9b599a9bf"). InnerVolumeSpecName "kube-api-access-xj59v". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:40:21 crc kubenswrapper[4631]: I1128 13:40:21.450215 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xj59v\" (UniqueName: \"kubernetes.io/projected/83d77da3-fbb7-434f-abaf-5be9b599a9bf-kube-api-access-xj59v\") on node \"crc\" DevicePath \"\"" Nov 28 13:40:21 crc kubenswrapper[4631]: I1128 13:40:21.905251 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"83d77da3-fbb7-434f-abaf-5be9b599a9bf","Type":"ContainerDied","Data":"0feb5123c599b3af93212c76b682c0ee877de3076bacb746e6bad20ec87f297f"} Nov 28 13:40:21 crc kubenswrapper[4631]: I1128 13:40:21.906540 4631 scope.go:117] "RemoveContainer" containerID="52d0c55ae2c92fdf222cd8f5c0b2b93b9073d9c66e613db020f87d5dd63419b7" Nov 28 13:40:21 crc kubenswrapper[4631]: I1128 13:40:21.906812 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 28 13:40:21 crc kubenswrapper[4631]: I1128 13:40:21.964389 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 28 13:40:21 crc kubenswrapper[4631]: I1128 13:40:21.983216 4631 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 28 13:40:22 crc kubenswrapper[4631]: I1128 13:40:22.013626 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Nov 28 13:40:22 crc kubenswrapper[4631]: E1128 13:40:22.014582 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="83d77da3-fbb7-434f-abaf-5be9b599a9bf" containerName="kube-state-metrics" Nov 28 13:40:22 crc kubenswrapper[4631]: I1128 13:40:22.014675 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="83d77da3-fbb7-434f-abaf-5be9b599a9bf" containerName="kube-state-metrics" Nov 28 13:40:22 crc kubenswrapper[4631]: I1128 13:40:22.014945 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="83d77da3-fbb7-434f-abaf-5be9b599a9bf" containerName="kube-state-metrics" Nov 28 13:40:22 crc kubenswrapper[4631]: I1128 13:40:22.015816 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 28 13:40:22 crc kubenswrapper[4631]: I1128 13:40:22.019201 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Nov 28 13:40:22 crc kubenswrapper[4631]: I1128 13:40:22.020935 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Nov 28 13:40:22 crc kubenswrapper[4631]: I1128 13:40:22.037269 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 28 13:40:22 crc kubenswrapper[4631]: I1128 13:40:22.060599 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ljz8w\" (UniqueName: \"kubernetes.io/projected/c1f85a30-128c-487c-ac41-f1d7a64a52a1-kube-api-access-ljz8w\") pod \"kube-state-metrics-0\" (UID: \"c1f85a30-128c-487c-ac41-f1d7a64a52a1\") " pod="openstack/kube-state-metrics-0" Nov 28 13:40:22 crc kubenswrapper[4631]: I1128 13:40:22.060925 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c1f85a30-128c-487c-ac41-f1d7a64a52a1-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"c1f85a30-128c-487c-ac41-f1d7a64a52a1\") " pod="openstack/kube-state-metrics-0" Nov 28 13:40:22 crc kubenswrapper[4631]: I1128 13:40:22.061132 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/c1f85a30-128c-487c-ac41-f1d7a64a52a1-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"c1f85a30-128c-487c-ac41-f1d7a64a52a1\") " pod="openstack/kube-state-metrics-0" Nov 28 13:40:22 crc kubenswrapper[4631]: I1128 13:40:22.062670 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/c1f85a30-128c-487c-ac41-f1d7a64a52a1-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"c1f85a30-128c-487c-ac41-f1d7a64a52a1\") " pod="openstack/kube-state-metrics-0" Nov 28 13:40:22 crc kubenswrapper[4631]: I1128 13:40:22.164668 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/c1f85a30-128c-487c-ac41-f1d7a64a52a1-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"c1f85a30-128c-487c-ac41-f1d7a64a52a1\") " pod="openstack/kube-state-metrics-0" Nov 28 13:40:22 crc kubenswrapper[4631]: I1128 13:40:22.164765 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/c1f85a30-128c-487c-ac41-f1d7a64a52a1-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"c1f85a30-128c-487c-ac41-f1d7a64a52a1\") " pod="openstack/kube-state-metrics-0" Nov 28 13:40:22 crc kubenswrapper[4631]: I1128 13:40:22.164789 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ljz8w\" (UniqueName: \"kubernetes.io/projected/c1f85a30-128c-487c-ac41-f1d7a64a52a1-kube-api-access-ljz8w\") pod \"kube-state-metrics-0\" (UID: \"c1f85a30-128c-487c-ac41-f1d7a64a52a1\") " pod="openstack/kube-state-metrics-0" Nov 28 13:40:22 crc kubenswrapper[4631]: I1128 13:40:22.164824 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c1f85a30-128c-487c-ac41-f1d7a64a52a1-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"c1f85a30-128c-487c-ac41-f1d7a64a52a1\") " pod="openstack/kube-state-metrics-0" Nov 28 13:40:22 crc kubenswrapper[4631]: I1128 13:40:22.173148 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/c1f85a30-128c-487c-ac41-f1d7a64a52a1-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"c1f85a30-128c-487c-ac41-f1d7a64a52a1\") " pod="openstack/kube-state-metrics-0" Nov 28 13:40:22 crc kubenswrapper[4631]: I1128 13:40:22.188050 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/c1f85a30-128c-487c-ac41-f1d7a64a52a1-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"c1f85a30-128c-487c-ac41-f1d7a64a52a1\") " pod="openstack/kube-state-metrics-0" Nov 28 13:40:22 crc kubenswrapper[4631]: I1128 13:40:22.202117 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c1f85a30-128c-487c-ac41-f1d7a64a52a1-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"c1f85a30-128c-487c-ac41-f1d7a64a52a1\") " pod="openstack/kube-state-metrics-0" Nov 28 13:40:22 crc kubenswrapper[4631]: I1128 13:40:22.248123 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ljz8w\" (UniqueName: \"kubernetes.io/projected/c1f85a30-128c-487c-ac41-f1d7a64a52a1-kube-api-access-ljz8w\") pod \"kube-state-metrics-0\" (UID: \"c1f85a30-128c-487c-ac41-f1d7a64a52a1\") " pod="openstack/kube-state-metrics-0" Nov 28 13:40:22 crc kubenswrapper[4631]: I1128 13:40:22.336205 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 28 13:40:23 crc kubenswrapper[4631]: W1128 13:40:23.016854 4631 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc1f85a30_128c_487c_ac41_f1d7a64a52a1.slice/crio-ad6a28d0e08182e03394715495edc04446c125e8b47b934f240a92bd5cec0b22 WatchSource:0}: Error finding container ad6a28d0e08182e03394715495edc04446c125e8b47b934f240a92bd5cec0b22: Status 404 returned error can't find the container with id ad6a28d0e08182e03394715495edc04446c125e8b47b934f240a92bd5cec0b22 Nov 28 13:40:23 crc kubenswrapper[4631]: I1128 13:40:23.020133 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 28 13:40:23 crc kubenswrapper[4631]: I1128 13:40:23.531371 4631 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="83d77da3-fbb7-434f-abaf-5be9b599a9bf" path="/var/lib/kubelet/pods/83d77da3-fbb7-434f-abaf-5be9b599a9bf/volumes" Nov 28 13:40:23 crc kubenswrapper[4631]: I1128 13:40:23.818569 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 28 13:40:23 crc kubenswrapper[4631]: I1128 13:40:23.818894 4631 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="736b1888-f3ac-46f6-9532-cdd4d4e310c8" containerName="ceilometer-central-agent" containerID="cri-o://66fc9c2ad3bdcee59e76c4232c6d98384e6b554b3218f4acfe4ddbfc6e974aa5" gracePeriod=30 Nov 28 13:40:23 crc kubenswrapper[4631]: I1128 13:40:23.819497 4631 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="736b1888-f3ac-46f6-9532-cdd4d4e310c8" containerName="proxy-httpd" containerID="cri-o://38d193db6b1a62df309e687fdbd62664a103f1d506c32337a40560ad416f0368" gracePeriod=30 Nov 28 13:40:23 crc kubenswrapper[4631]: I1128 13:40:23.819571 4631 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="736b1888-f3ac-46f6-9532-cdd4d4e310c8" containerName="sg-core" containerID="cri-o://a9279936337bbc804fdbffba98e5a5ea1aa3d8c892c3df4653c010e0721d4ade" gracePeriod=30 Nov 28 13:40:23 crc kubenswrapper[4631]: I1128 13:40:23.819617 4631 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="736b1888-f3ac-46f6-9532-cdd4d4e310c8" containerName="ceilometer-notification-agent" containerID="cri-o://1c14e41c0b95654690f096dcb82b6936bc505ab33653274e9bd24cce207423ec" gracePeriod=30 Nov 28 13:40:23 crc kubenswrapper[4631]: I1128 13:40:23.930945 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"c1f85a30-128c-487c-ac41-f1d7a64a52a1","Type":"ContainerStarted","Data":"ad6a28d0e08182e03394715495edc04446c125e8b47b934f240a92bd5cec0b22"} Nov 28 13:40:24 crc kubenswrapper[4631]: I1128 13:40:24.944507 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"c1f85a30-128c-487c-ac41-f1d7a64a52a1","Type":"ContainerStarted","Data":"8e77e77a9f8d062c4dce49da3ad0418a7c28086a0d28db243dce4b8bcea35c47"} Nov 28 13:40:24 crc kubenswrapper[4631]: I1128 13:40:24.945408 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Nov 28 13:40:24 crc kubenswrapper[4631]: I1128 13:40:24.950068 4631 generic.go:334] "Generic (PLEG): container finished" podID="736b1888-f3ac-46f6-9532-cdd4d4e310c8" containerID="38d193db6b1a62df309e687fdbd62664a103f1d506c32337a40560ad416f0368" exitCode=0 Nov 28 13:40:24 crc kubenswrapper[4631]: I1128 13:40:24.950276 4631 generic.go:334] "Generic (PLEG): container finished" podID="736b1888-f3ac-46f6-9532-cdd4d4e310c8" containerID="a9279936337bbc804fdbffba98e5a5ea1aa3d8c892c3df4653c010e0721d4ade" exitCode=2 Nov 28 13:40:24 crc kubenswrapper[4631]: I1128 13:40:24.950407 4631 generic.go:334] "Generic (PLEG): container finished" podID="736b1888-f3ac-46f6-9532-cdd4d4e310c8" containerID="66fc9c2ad3bdcee59e76c4232c6d98384e6b554b3218f4acfe4ddbfc6e974aa5" exitCode=0 Nov 28 13:40:24 crc kubenswrapper[4631]: I1128 13:40:24.950122 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"736b1888-f3ac-46f6-9532-cdd4d4e310c8","Type":"ContainerDied","Data":"38d193db6b1a62df309e687fdbd62664a103f1d506c32337a40560ad416f0368"} Nov 28 13:40:24 crc kubenswrapper[4631]: I1128 13:40:24.950557 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"736b1888-f3ac-46f6-9532-cdd4d4e310c8","Type":"ContainerDied","Data":"a9279936337bbc804fdbffba98e5a5ea1aa3d8c892c3df4653c010e0721d4ade"} Nov 28 13:40:24 crc kubenswrapper[4631]: I1128 13:40:24.950668 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"736b1888-f3ac-46f6-9532-cdd4d4e310c8","Type":"ContainerDied","Data":"66fc9c2ad3bdcee59e76c4232c6d98384e6b554b3218f4acfe4ddbfc6e974aa5"} Nov 28 13:40:25 crc kubenswrapper[4631]: I1128 13:40:25.332163 4631 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-64f8fdf97b-jn5mg" podUID="111c5349-528d-4856-a3e4-1277e3b10889" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.146:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.146:8443: connect: connection refused" Nov 28 13:40:25 crc kubenswrapper[4631]: I1128 13:40:25.554332 4631 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-544496c768-rzhsw" podUID="cd607dae-568a-4b81-af81-3310c0e95854" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.147:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.147:8443: connect: connection refused" Nov 28 13:40:25 crc kubenswrapper[4631]: I1128 13:40:25.963001 4631 generic.go:334] "Generic (PLEG): container finished" podID="736b1888-f3ac-46f6-9532-cdd4d4e310c8" containerID="1c14e41c0b95654690f096dcb82b6936bc505ab33653274e9bd24cce207423ec" exitCode=0 Nov 28 13:40:25 crc kubenswrapper[4631]: I1128 13:40:25.964249 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"736b1888-f3ac-46f6-9532-cdd4d4e310c8","Type":"ContainerDied","Data":"1c14e41c0b95654690f096dcb82b6936bc505ab33653274e9bd24cce207423ec"} Nov 28 13:40:26 crc kubenswrapper[4631]: I1128 13:40:26.735227 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 28 13:40:26 crc kubenswrapper[4631]: I1128 13:40:26.766871 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=5.030771827 podStartE2EDuration="5.766851777s" podCreationTimestamp="2025-11-28 13:40:21 +0000 UTC" firstStartedPulling="2025-11-28 13:40:23.022782664 +0000 UTC m=+1179.830086008" lastFinishedPulling="2025-11-28 13:40:23.758862624 +0000 UTC m=+1180.566165958" observedRunningTime="2025-11-28 13:40:24.969588542 +0000 UTC m=+1181.776891886" watchObservedRunningTime="2025-11-28 13:40:26.766851777 +0000 UTC m=+1183.574155121" Nov 28 13:40:26 crc kubenswrapper[4631]: I1128 13:40:26.790227 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/736b1888-f3ac-46f6-9532-cdd4d4e310c8-log-httpd\") pod \"736b1888-f3ac-46f6-9532-cdd4d4e310c8\" (UID: \"736b1888-f3ac-46f6-9532-cdd4d4e310c8\") " Nov 28 13:40:26 crc kubenswrapper[4631]: I1128 13:40:26.790398 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/736b1888-f3ac-46f6-9532-cdd4d4e310c8-combined-ca-bundle\") pod \"736b1888-f3ac-46f6-9532-cdd4d4e310c8\" (UID: \"736b1888-f3ac-46f6-9532-cdd4d4e310c8\") " Nov 28 13:40:26 crc kubenswrapper[4631]: I1128 13:40:26.790609 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/736b1888-f3ac-46f6-9532-cdd4d4e310c8-run-httpd\") pod \"736b1888-f3ac-46f6-9532-cdd4d4e310c8\" (UID: \"736b1888-f3ac-46f6-9532-cdd4d4e310c8\") " Nov 28 13:40:26 crc kubenswrapper[4631]: I1128 13:40:26.790728 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/736b1888-f3ac-46f6-9532-cdd4d4e310c8-scripts\") pod \"736b1888-f3ac-46f6-9532-cdd4d4e310c8\" (UID: \"736b1888-f3ac-46f6-9532-cdd4d4e310c8\") " Nov 28 13:40:26 crc kubenswrapper[4631]: I1128 13:40:26.790765 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hrdmr\" (UniqueName: \"kubernetes.io/projected/736b1888-f3ac-46f6-9532-cdd4d4e310c8-kube-api-access-hrdmr\") pod \"736b1888-f3ac-46f6-9532-cdd4d4e310c8\" (UID: \"736b1888-f3ac-46f6-9532-cdd4d4e310c8\") " Nov 28 13:40:26 crc kubenswrapper[4631]: I1128 13:40:26.790797 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/736b1888-f3ac-46f6-9532-cdd4d4e310c8-sg-core-conf-yaml\") pod \"736b1888-f3ac-46f6-9532-cdd4d4e310c8\" (UID: \"736b1888-f3ac-46f6-9532-cdd4d4e310c8\") " Nov 28 13:40:26 crc kubenswrapper[4631]: I1128 13:40:26.790833 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/736b1888-f3ac-46f6-9532-cdd4d4e310c8-config-data\") pod \"736b1888-f3ac-46f6-9532-cdd4d4e310c8\" (UID: \"736b1888-f3ac-46f6-9532-cdd4d4e310c8\") " Nov 28 13:40:26 crc kubenswrapper[4631]: I1128 13:40:26.799500 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/736b1888-f3ac-46f6-9532-cdd4d4e310c8-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "736b1888-f3ac-46f6-9532-cdd4d4e310c8" (UID: "736b1888-f3ac-46f6-9532-cdd4d4e310c8"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 13:40:26 crc kubenswrapper[4631]: I1128 13:40:26.799834 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/736b1888-f3ac-46f6-9532-cdd4d4e310c8-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "736b1888-f3ac-46f6-9532-cdd4d4e310c8" (UID: "736b1888-f3ac-46f6-9532-cdd4d4e310c8"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 13:40:26 crc kubenswrapper[4631]: I1128 13:40:26.805712 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/736b1888-f3ac-46f6-9532-cdd4d4e310c8-scripts" (OuterVolumeSpecName: "scripts") pod "736b1888-f3ac-46f6-9532-cdd4d4e310c8" (UID: "736b1888-f3ac-46f6-9532-cdd4d4e310c8"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:40:26 crc kubenswrapper[4631]: I1128 13:40:26.828746 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/736b1888-f3ac-46f6-9532-cdd4d4e310c8-kube-api-access-hrdmr" (OuterVolumeSpecName: "kube-api-access-hrdmr") pod "736b1888-f3ac-46f6-9532-cdd4d4e310c8" (UID: "736b1888-f3ac-46f6-9532-cdd4d4e310c8"). InnerVolumeSpecName "kube-api-access-hrdmr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:40:26 crc kubenswrapper[4631]: I1128 13:40:26.854127 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/736b1888-f3ac-46f6-9532-cdd4d4e310c8-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "736b1888-f3ac-46f6-9532-cdd4d4e310c8" (UID: "736b1888-f3ac-46f6-9532-cdd4d4e310c8"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:40:26 crc kubenswrapper[4631]: I1128 13:40:26.896250 4631 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/736b1888-f3ac-46f6-9532-cdd4d4e310c8-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 28 13:40:26 crc kubenswrapper[4631]: I1128 13:40:26.896303 4631 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/736b1888-f3ac-46f6-9532-cdd4d4e310c8-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 13:40:26 crc kubenswrapper[4631]: I1128 13:40:26.896314 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hrdmr\" (UniqueName: \"kubernetes.io/projected/736b1888-f3ac-46f6-9532-cdd4d4e310c8-kube-api-access-hrdmr\") on node \"crc\" DevicePath \"\"" Nov 28 13:40:26 crc kubenswrapper[4631]: I1128 13:40:26.896327 4631 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/736b1888-f3ac-46f6-9532-cdd4d4e310c8-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 28 13:40:26 crc kubenswrapper[4631]: I1128 13:40:26.896336 4631 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/736b1888-f3ac-46f6-9532-cdd4d4e310c8-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 28 13:40:26 crc kubenswrapper[4631]: I1128 13:40:26.967534 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/736b1888-f3ac-46f6-9532-cdd4d4e310c8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "736b1888-f3ac-46f6-9532-cdd4d4e310c8" (UID: "736b1888-f3ac-46f6-9532-cdd4d4e310c8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:40:26 crc kubenswrapper[4631]: I1128 13:40:26.984804 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"736b1888-f3ac-46f6-9532-cdd4d4e310c8","Type":"ContainerDied","Data":"68b45d3ea18cc197789698afa6f17b0ce6989f5c4bf6441e21fd8482d8af8e3b"} Nov 28 13:40:26 crc kubenswrapper[4631]: I1128 13:40:26.984841 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 28 13:40:26 crc kubenswrapper[4631]: I1128 13:40:26.985057 4631 scope.go:117] "RemoveContainer" containerID="38d193db6b1a62df309e687fdbd62664a103f1d506c32337a40560ad416f0368" Nov 28 13:40:26 crc kubenswrapper[4631]: I1128 13:40:26.998123 4631 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/736b1888-f3ac-46f6-9532-cdd4d4e310c8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 13:40:27 crc kubenswrapper[4631]: I1128 13:40:27.024152 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/736b1888-f3ac-46f6-9532-cdd4d4e310c8-config-data" (OuterVolumeSpecName: "config-data") pod "736b1888-f3ac-46f6-9532-cdd4d4e310c8" (UID: "736b1888-f3ac-46f6-9532-cdd4d4e310c8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:40:27 crc kubenswrapper[4631]: I1128 13:40:27.032736 4631 scope.go:117] "RemoveContainer" containerID="a9279936337bbc804fdbffba98e5a5ea1aa3d8c892c3df4653c010e0721d4ade" Nov 28 13:40:27 crc kubenswrapper[4631]: I1128 13:40:27.062070 4631 scope.go:117] "RemoveContainer" containerID="1c14e41c0b95654690f096dcb82b6936bc505ab33653274e9bd24cce207423ec" Nov 28 13:40:27 crc kubenswrapper[4631]: I1128 13:40:27.092659 4631 scope.go:117] "RemoveContainer" containerID="66fc9c2ad3bdcee59e76c4232c6d98384e6b554b3218f4acfe4ddbfc6e974aa5" Nov 28 13:40:27 crc kubenswrapper[4631]: I1128 13:40:27.100122 4631 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/736b1888-f3ac-46f6-9532-cdd4d4e310c8-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 13:40:27 crc kubenswrapper[4631]: I1128 13:40:27.350347 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 28 13:40:27 crc kubenswrapper[4631]: E1128 13:40:27.367563 4631 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod641b1d24_f22a_4d63_8e9e_d76c3dde2135.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0f61d189_17d8_4154_ac95_df78d4042055.slice/crio-0a32d7ae2c3a534184dba524af21eb67d2088a5219dc164fd35b2643f3e54ddd\": RecentStats: unable to find data in memory cache]" Nov 28 13:40:27 crc kubenswrapper[4631]: I1128 13:40:27.375584 4631 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 28 13:40:27 crc kubenswrapper[4631]: I1128 13:40:27.409168 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 28 13:40:27 crc kubenswrapper[4631]: E1128 13:40:27.410260 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="736b1888-f3ac-46f6-9532-cdd4d4e310c8" containerName="sg-core" Nov 28 13:40:27 crc kubenswrapper[4631]: I1128 13:40:27.410280 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="736b1888-f3ac-46f6-9532-cdd4d4e310c8" containerName="sg-core" Nov 28 13:40:27 crc kubenswrapper[4631]: E1128 13:40:27.410388 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="736b1888-f3ac-46f6-9532-cdd4d4e310c8" containerName="proxy-httpd" Nov 28 13:40:27 crc kubenswrapper[4631]: I1128 13:40:27.410398 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="736b1888-f3ac-46f6-9532-cdd4d4e310c8" containerName="proxy-httpd" Nov 28 13:40:27 crc kubenswrapper[4631]: E1128 13:40:27.410416 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="736b1888-f3ac-46f6-9532-cdd4d4e310c8" containerName="ceilometer-central-agent" Nov 28 13:40:27 crc kubenswrapper[4631]: I1128 13:40:27.410463 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="736b1888-f3ac-46f6-9532-cdd4d4e310c8" containerName="ceilometer-central-agent" Nov 28 13:40:27 crc kubenswrapper[4631]: E1128 13:40:27.410478 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="736b1888-f3ac-46f6-9532-cdd4d4e310c8" containerName="ceilometer-notification-agent" Nov 28 13:40:27 crc kubenswrapper[4631]: I1128 13:40:27.410487 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="736b1888-f3ac-46f6-9532-cdd4d4e310c8" containerName="ceilometer-notification-agent" Nov 28 13:40:27 crc kubenswrapper[4631]: I1128 13:40:27.411075 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="736b1888-f3ac-46f6-9532-cdd4d4e310c8" containerName="ceilometer-central-agent" Nov 28 13:40:27 crc kubenswrapper[4631]: I1128 13:40:27.411104 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="736b1888-f3ac-46f6-9532-cdd4d4e310c8" containerName="ceilometer-notification-agent" Nov 28 13:40:27 crc kubenswrapper[4631]: I1128 13:40:27.411122 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="736b1888-f3ac-46f6-9532-cdd4d4e310c8" containerName="sg-core" Nov 28 13:40:27 crc kubenswrapper[4631]: I1128 13:40:27.411144 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="736b1888-f3ac-46f6-9532-cdd4d4e310c8" containerName="proxy-httpd" Nov 28 13:40:27 crc kubenswrapper[4631]: I1128 13:40:27.413689 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 28 13:40:27 crc kubenswrapper[4631]: I1128 13:40:27.420824 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 28 13:40:27 crc kubenswrapper[4631]: I1128 13:40:27.421065 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 28 13:40:27 crc kubenswrapper[4631]: I1128 13:40:27.421226 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Nov 28 13:40:27 crc kubenswrapper[4631]: I1128 13:40:27.424976 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 28 13:40:27 crc kubenswrapper[4631]: I1128 13:40:27.527761 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a00de8c7-a886-4851-b7cd-c8784a3c6bd9-scripts\") pod \"ceilometer-0\" (UID: \"a00de8c7-a886-4851-b7cd-c8784a3c6bd9\") " pod="openstack/ceilometer-0" Nov 28 13:40:27 crc kubenswrapper[4631]: I1128 13:40:27.528006 4631 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="736b1888-f3ac-46f6-9532-cdd4d4e310c8" path="/var/lib/kubelet/pods/736b1888-f3ac-46f6-9532-cdd4d4e310c8/volumes" Nov 28 13:40:27 crc kubenswrapper[4631]: I1128 13:40:27.531208 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a00de8c7-a886-4851-b7cd-c8784a3c6bd9-log-httpd\") pod \"ceilometer-0\" (UID: \"a00de8c7-a886-4851-b7cd-c8784a3c6bd9\") " pod="openstack/ceilometer-0" Nov 28 13:40:27 crc kubenswrapper[4631]: I1128 13:40:27.531344 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a00de8c7-a886-4851-b7cd-c8784a3c6bd9-config-data\") pod \"ceilometer-0\" (UID: \"a00de8c7-a886-4851-b7cd-c8784a3c6bd9\") " pod="openstack/ceilometer-0" Nov 28 13:40:27 crc kubenswrapper[4631]: I1128 13:40:27.531420 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a00de8c7-a886-4851-b7cd-c8784a3c6bd9-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a00de8c7-a886-4851-b7cd-c8784a3c6bd9\") " pod="openstack/ceilometer-0" Nov 28 13:40:27 crc kubenswrapper[4631]: I1128 13:40:27.531527 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xnzmq\" (UniqueName: \"kubernetes.io/projected/a00de8c7-a886-4851-b7cd-c8784a3c6bd9-kube-api-access-xnzmq\") pod \"ceilometer-0\" (UID: \"a00de8c7-a886-4851-b7cd-c8784a3c6bd9\") " pod="openstack/ceilometer-0" Nov 28 13:40:27 crc kubenswrapper[4631]: I1128 13:40:27.531573 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/a00de8c7-a886-4851-b7cd-c8784a3c6bd9-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"a00de8c7-a886-4851-b7cd-c8784a3c6bd9\") " pod="openstack/ceilometer-0" Nov 28 13:40:27 crc kubenswrapper[4631]: I1128 13:40:27.531642 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a00de8c7-a886-4851-b7cd-c8784a3c6bd9-run-httpd\") pod \"ceilometer-0\" (UID: \"a00de8c7-a886-4851-b7cd-c8784a3c6bd9\") " pod="openstack/ceilometer-0" Nov 28 13:40:27 crc kubenswrapper[4631]: I1128 13:40:27.531666 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a00de8c7-a886-4851-b7cd-c8784a3c6bd9-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a00de8c7-a886-4851-b7cd-c8784a3c6bd9\") " pod="openstack/ceilometer-0" Nov 28 13:40:27 crc kubenswrapper[4631]: I1128 13:40:27.634121 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xnzmq\" (UniqueName: \"kubernetes.io/projected/a00de8c7-a886-4851-b7cd-c8784a3c6bd9-kube-api-access-xnzmq\") pod \"ceilometer-0\" (UID: \"a00de8c7-a886-4851-b7cd-c8784a3c6bd9\") " pod="openstack/ceilometer-0" Nov 28 13:40:27 crc kubenswrapper[4631]: I1128 13:40:27.634195 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/a00de8c7-a886-4851-b7cd-c8784a3c6bd9-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"a00de8c7-a886-4851-b7cd-c8784a3c6bd9\") " pod="openstack/ceilometer-0" Nov 28 13:40:27 crc kubenswrapper[4631]: I1128 13:40:27.634239 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a00de8c7-a886-4851-b7cd-c8784a3c6bd9-run-httpd\") pod \"ceilometer-0\" (UID: \"a00de8c7-a886-4851-b7cd-c8784a3c6bd9\") " pod="openstack/ceilometer-0" Nov 28 13:40:27 crc kubenswrapper[4631]: I1128 13:40:27.634263 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a00de8c7-a886-4851-b7cd-c8784a3c6bd9-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a00de8c7-a886-4851-b7cd-c8784a3c6bd9\") " pod="openstack/ceilometer-0" Nov 28 13:40:27 crc kubenswrapper[4631]: I1128 13:40:27.634328 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a00de8c7-a886-4851-b7cd-c8784a3c6bd9-scripts\") pod \"ceilometer-0\" (UID: \"a00de8c7-a886-4851-b7cd-c8784a3c6bd9\") " pod="openstack/ceilometer-0" Nov 28 13:40:27 crc kubenswrapper[4631]: I1128 13:40:27.634393 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a00de8c7-a886-4851-b7cd-c8784a3c6bd9-log-httpd\") pod \"ceilometer-0\" (UID: \"a00de8c7-a886-4851-b7cd-c8784a3c6bd9\") " pod="openstack/ceilometer-0" Nov 28 13:40:27 crc kubenswrapper[4631]: I1128 13:40:27.634413 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a00de8c7-a886-4851-b7cd-c8784a3c6bd9-config-data\") pod \"ceilometer-0\" (UID: \"a00de8c7-a886-4851-b7cd-c8784a3c6bd9\") " pod="openstack/ceilometer-0" Nov 28 13:40:27 crc kubenswrapper[4631]: I1128 13:40:27.634436 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a00de8c7-a886-4851-b7cd-c8784a3c6bd9-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a00de8c7-a886-4851-b7cd-c8784a3c6bd9\") " pod="openstack/ceilometer-0" Nov 28 13:40:27 crc kubenswrapper[4631]: I1128 13:40:27.635967 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a00de8c7-a886-4851-b7cd-c8784a3c6bd9-run-httpd\") pod \"ceilometer-0\" (UID: \"a00de8c7-a886-4851-b7cd-c8784a3c6bd9\") " pod="openstack/ceilometer-0" Nov 28 13:40:27 crc kubenswrapper[4631]: I1128 13:40:27.636343 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a00de8c7-a886-4851-b7cd-c8784a3c6bd9-log-httpd\") pod \"ceilometer-0\" (UID: \"a00de8c7-a886-4851-b7cd-c8784a3c6bd9\") " pod="openstack/ceilometer-0" Nov 28 13:40:27 crc kubenswrapper[4631]: I1128 13:40:27.641276 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a00de8c7-a886-4851-b7cd-c8784a3c6bd9-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a00de8c7-a886-4851-b7cd-c8784a3c6bd9\") " pod="openstack/ceilometer-0" Nov 28 13:40:27 crc kubenswrapper[4631]: I1128 13:40:27.643786 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a00de8c7-a886-4851-b7cd-c8784a3c6bd9-config-data\") pod \"ceilometer-0\" (UID: \"a00de8c7-a886-4851-b7cd-c8784a3c6bd9\") " pod="openstack/ceilometer-0" Nov 28 13:40:27 crc kubenswrapper[4631]: I1128 13:40:27.644756 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a00de8c7-a886-4851-b7cd-c8784a3c6bd9-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a00de8c7-a886-4851-b7cd-c8784a3c6bd9\") " pod="openstack/ceilometer-0" Nov 28 13:40:27 crc kubenswrapper[4631]: I1128 13:40:27.648399 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/a00de8c7-a886-4851-b7cd-c8784a3c6bd9-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"a00de8c7-a886-4851-b7cd-c8784a3c6bd9\") " pod="openstack/ceilometer-0" Nov 28 13:40:27 crc kubenswrapper[4631]: I1128 13:40:27.661214 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a00de8c7-a886-4851-b7cd-c8784a3c6bd9-scripts\") pod \"ceilometer-0\" (UID: \"a00de8c7-a886-4851-b7cd-c8784a3c6bd9\") " pod="openstack/ceilometer-0" Nov 28 13:40:27 crc kubenswrapper[4631]: I1128 13:40:27.667219 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xnzmq\" (UniqueName: \"kubernetes.io/projected/a00de8c7-a886-4851-b7cd-c8784a3c6bd9-kube-api-access-xnzmq\") pod \"ceilometer-0\" (UID: \"a00de8c7-a886-4851-b7cd-c8784a3c6bd9\") " pod="openstack/ceilometer-0" Nov 28 13:40:27 crc kubenswrapper[4631]: I1128 13:40:27.755160 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 28 13:40:28 crc kubenswrapper[4631]: I1128 13:40:28.286066 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 28 13:40:28 crc kubenswrapper[4631]: W1128 13:40:28.308465 4631 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda00de8c7_a886_4851_b7cd_c8784a3c6bd9.slice/crio-83a90711a3c7c51a24a1c070bf63ee69b6f7698a291854e4ceffb37c59d5f4f5 WatchSource:0}: Error finding container 83a90711a3c7c51a24a1c070bf63ee69b6f7698a291854e4ceffb37c59d5f4f5: Status 404 returned error can't find the container with id 83a90711a3c7c51a24a1c070bf63ee69b6f7698a291854e4ceffb37c59d5f4f5 Nov 28 13:40:29 crc kubenswrapper[4631]: I1128 13:40:29.025586 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a00de8c7-a886-4851-b7cd-c8784a3c6bd9","Type":"ContainerStarted","Data":"83a90711a3c7c51a24a1c070bf63ee69b6f7698a291854e4ceffb37c59d5f4f5"} Nov 28 13:40:30 crc kubenswrapper[4631]: I1128 13:40:30.036097 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a00de8c7-a886-4851-b7cd-c8784a3c6bd9","Type":"ContainerStarted","Data":"3e70f5937aa64c1f6b17c239bba41cef0aa54ddf458ce5b8cafeccbb3e212581"} Nov 28 13:40:31 crc kubenswrapper[4631]: I1128 13:40:31.052565 4631 generic.go:334] "Generic (PLEG): container finished" podID="8a4fda29-ad66-4077-b2f9-796fe6fd961a" containerID="d9cdc53db79944ad3a916e49b0da1cb467958dc12bea86bc7c817bf6487bf6ff" exitCode=0 Nov 28 13:40:31 crc kubenswrapper[4631]: I1128 13:40:31.052682 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-tqj67" event={"ID":"8a4fda29-ad66-4077-b2f9-796fe6fd961a","Type":"ContainerDied","Data":"d9cdc53db79944ad3a916e49b0da1cb467958dc12bea86bc7c817bf6487bf6ff"} Nov 28 13:40:31 crc kubenswrapper[4631]: I1128 13:40:31.060141 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a00de8c7-a886-4851-b7cd-c8784a3c6bd9","Type":"ContainerStarted","Data":"0f6318a7db020712586bb2e56cdd6f8046b9a725a41e45e0a31daac51781c62c"} Nov 28 13:40:32 crc kubenswrapper[4631]: I1128 13:40:32.074379 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a00de8c7-a886-4851-b7cd-c8784a3c6bd9","Type":"ContainerStarted","Data":"11e2e2d4a8aff3ef637fdd4f4485a2cb4c252744fa8b2147478739303369e0e2"} Nov 28 13:40:32 crc kubenswrapper[4631]: I1128 13:40:32.500107 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Nov 28 13:40:32 crc kubenswrapper[4631]: I1128 13:40:32.536822 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-tqj67" Nov 28 13:40:32 crc kubenswrapper[4631]: I1128 13:40:32.668818 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8a4fda29-ad66-4077-b2f9-796fe6fd961a-scripts\") pod \"8a4fda29-ad66-4077-b2f9-796fe6fd961a\" (UID: \"8a4fda29-ad66-4077-b2f9-796fe6fd961a\") " Nov 28 13:40:32 crc kubenswrapper[4631]: I1128 13:40:32.668900 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qwp6z\" (UniqueName: \"kubernetes.io/projected/8a4fda29-ad66-4077-b2f9-796fe6fd961a-kube-api-access-qwp6z\") pod \"8a4fda29-ad66-4077-b2f9-796fe6fd961a\" (UID: \"8a4fda29-ad66-4077-b2f9-796fe6fd961a\") " Nov 28 13:40:32 crc kubenswrapper[4631]: I1128 13:40:32.669052 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a4fda29-ad66-4077-b2f9-796fe6fd961a-combined-ca-bundle\") pod \"8a4fda29-ad66-4077-b2f9-796fe6fd961a\" (UID: \"8a4fda29-ad66-4077-b2f9-796fe6fd961a\") " Nov 28 13:40:32 crc kubenswrapper[4631]: I1128 13:40:32.669130 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a4fda29-ad66-4077-b2f9-796fe6fd961a-config-data\") pod \"8a4fda29-ad66-4077-b2f9-796fe6fd961a\" (UID: \"8a4fda29-ad66-4077-b2f9-796fe6fd961a\") " Nov 28 13:40:32 crc kubenswrapper[4631]: I1128 13:40:32.710522 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8a4fda29-ad66-4077-b2f9-796fe6fd961a-kube-api-access-qwp6z" (OuterVolumeSpecName: "kube-api-access-qwp6z") pod "8a4fda29-ad66-4077-b2f9-796fe6fd961a" (UID: "8a4fda29-ad66-4077-b2f9-796fe6fd961a"). InnerVolumeSpecName "kube-api-access-qwp6z". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:40:32 crc kubenswrapper[4631]: I1128 13:40:32.710734 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a4fda29-ad66-4077-b2f9-796fe6fd961a-scripts" (OuterVolumeSpecName: "scripts") pod "8a4fda29-ad66-4077-b2f9-796fe6fd961a" (UID: "8a4fda29-ad66-4077-b2f9-796fe6fd961a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:40:32 crc kubenswrapper[4631]: I1128 13:40:32.766540 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a4fda29-ad66-4077-b2f9-796fe6fd961a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8a4fda29-ad66-4077-b2f9-796fe6fd961a" (UID: "8a4fda29-ad66-4077-b2f9-796fe6fd961a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:40:32 crc kubenswrapper[4631]: I1128 13:40:32.768200 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a4fda29-ad66-4077-b2f9-796fe6fd961a-config-data" (OuterVolumeSpecName: "config-data") pod "8a4fda29-ad66-4077-b2f9-796fe6fd961a" (UID: "8a4fda29-ad66-4077-b2f9-796fe6fd961a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:40:32 crc kubenswrapper[4631]: I1128 13:40:32.771765 4631 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a4fda29-ad66-4077-b2f9-796fe6fd961a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 13:40:32 crc kubenswrapper[4631]: I1128 13:40:32.771798 4631 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a4fda29-ad66-4077-b2f9-796fe6fd961a-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 13:40:32 crc kubenswrapper[4631]: I1128 13:40:32.771808 4631 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8a4fda29-ad66-4077-b2f9-796fe6fd961a-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 13:40:32 crc kubenswrapper[4631]: I1128 13:40:32.771820 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qwp6z\" (UniqueName: \"kubernetes.io/projected/8a4fda29-ad66-4077-b2f9-796fe6fd961a-kube-api-access-qwp6z\") on node \"crc\" DevicePath \"\"" Nov 28 13:40:33 crc kubenswrapper[4631]: I1128 13:40:33.087133 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-tqj67" Nov 28 13:40:33 crc kubenswrapper[4631]: I1128 13:40:33.087152 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-tqj67" event={"ID":"8a4fda29-ad66-4077-b2f9-796fe6fd961a","Type":"ContainerDied","Data":"1ea96f588712b0c65d6c2ca7037ca23fc287850635e3e2aa8b7beb94c25a076d"} Nov 28 13:40:33 crc kubenswrapper[4631]: I1128 13:40:33.088614 4631 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1ea96f588712b0c65d6c2ca7037ca23fc287850635e3e2aa8b7beb94c25a076d" Nov 28 13:40:33 crc kubenswrapper[4631]: I1128 13:40:33.090934 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a00de8c7-a886-4851-b7cd-c8784a3c6bd9","Type":"ContainerStarted","Data":"d1a7db8ddb7853f6005959d4900a1ada50c4962ce089b37695da99fd552f8573"} Nov 28 13:40:33 crc kubenswrapper[4631]: I1128 13:40:33.091834 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 28 13:40:33 crc kubenswrapper[4631]: I1128 13:40:33.127143 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.155803493 podStartE2EDuration="6.127114129s" podCreationTimestamp="2025-11-28 13:40:27 +0000 UTC" firstStartedPulling="2025-11-28 13:40:28.311327138 +0000 UTC m=+1185.118630482" lastFinishedPulling="2025-11-28 13:40:32.282637774 +0000 UTC m=+1189.089941118" observedRunningTime="2025-11-28 13:40:33.118608036 +0000 UTC m=+1189.925911380" watchObservedRunningTime="2025-11-28 13:40:33.127114129 +0000 UTC m=+1189.934417473" Nov 28 13:40:33 crc kubenswrapper[4631]: I1128 13:40:33.257753 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Nov 28 13:40:33 crc kubenswrapper[4631]: E1128 13:40:33.258316 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a4fda29-ad66-4077-b2f9-796fe6fd961a" containerName="nova-cell0-conductor-db-sync" Nov 28 13:40:33 crc kubenswrapper[4631]: I1128 13:40:33.258344 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a4fda29-ad66-4077-b2f9-796fe6fd961a" containerName="nova-cell0-conductor-db-sync" Nov 28 13:40:33 crc kubenswrapper[4631]: I1128 13:40:33.258564 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="8a4fda29-ad66-4077-b2f9-796fe6fd961a" containerName="nova-cell0-conductor-db-sync" Nov 28 13:40:33 crc kubenswrapper[4631]: I1128 13:40:33.259332 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Nov 28 13:40:33 crc kubenswrapper[4631]: I1128 13:40:33.262110 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Nov 28 13:40:33 crc kubenswrapper[4631]: I1128 13:40:33.262415 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-z77kh" Nov 28 13:40:33 crc kubenswrapper[4631]: I1128 13:40:33.289161 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Nov 28 13:40:33 crc kubenswrapper[4631]: I1128 13:40:33.385759 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cc2a0580-660d-4a19-a755-70256a4e4abe-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"cc2a0580-660d-4a19-a755-70256a4e4abe\") " pod="openstack/nova-cell0-conductor-0" Nov 28 13:40:33 crc kubenswrapper[4631]: I1128 13:40:33.385906 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc2a0580-660d-4a19-a755-70256a4e4abe-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"cc2a0580-660d-4a19-a755-70256a4e4abe\") " pod="openstack/nova-cell0-conductor-0" Nov 28 13:40:33 crc kubenswrapper[4631]: I1128 13:40:33.386863 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zfx92\" (UniqueName: \"kubernetes.io/projected/cc2a0580-660d-4a19-a755-70256a4e4abe-kube-api-access-zfx92\") pod \"nova-cell0-conductor-0\" (UID: \"cc2a0580-660d-4a19-a755-70256a4e4abe\") " pod="openstack/nova-cell0-conductor-0" Nov 28 13:40:33 crc kubenswrapper[4631]: I1128 13:40:33.489199 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zfx92\" (UniqueName: \"kubernetes.io/projected/cc2a0580-660d-4a19-a755-70256a4e4abe-kube-api-access-zfx92\") pod \"nova-cell0-conductor-0\" (UID: \"cc2a0580-660d-4a19-a755-70256a4e4abe\") " pod="openstack/nova-cell0-conductor-0" Nov 28 13:40:33 crc kubenswrapper[4631]: I1128 13:40:33.489278 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cc2a0580-660d-4a19-a755-70256a4e4abe-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"cc2a0580-660d-4a19-a755-70256a4e4abe\") " pod="openstack/nova-cell0-conductor-0" Nov 28 13:40:33 crc kubenswrapper[4631]: I1128 13:40:33.489354 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc2a0580-660d-4a19-a755-70256a4e4abe-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"cc2a0580-660d-4a19-a755-70256a4e4abe\") " pod="openstack/nova-cell0-conductor-0" Nov 28 13:40:33 crc kubenswrapper[4631]: I1128 13:40:33.497778 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc2a0580-660d-4a19-a755-70256a4e4abe-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"cc2a0580-660d-4a19-a755-70256a4e4abe\") " pod="openstack/nova-cell0-conductor-0" Nov 28 13:40:33 crc kubenswrapper[4631]: I1128 13:40:33.502152 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cc2a0580-660d-4a19-a755-70256a4e4abe-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"cc2a0580-660d-4a19-a755-70256a4e4abe\") " pod="openstack/nova-cell0-conductor-0" Nov 28 13:40:33 crc kubenswrapper[4631]: I1128 13:40:33.525125 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zfx92\" (UniqueName: \"kubernetes.io/projected/cc2a0580-660d-4a19-a755-70256a4e4abe-kube-api-access-zfx92\") pod \"nova-cell0-conductor-0\" (UID: \"cc2a0580-660d-4a19-a755-70256a4e4abe\") " pod="openstack/nova-cell0-conductor-0" Nov 28 13:40:33 crc kubenswrapper[4631]: I1128 13:40:33.579874 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Nov 28 13:40:34 crc kubenswrapper[4631]: I1128 13:40:34.214841 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Nov 28 13:40:34 crc kubenswrapper[4631]: W1128 13:40:34.250009 4631 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcc2a0580_660d_4a19_a755_70256a4e4abe.slice/crio-551823f5a56a07b9b7d23efe0331afd0e253946448e4d36abee52f7f69e9ec09 WatchSource:0}: Error finding container 551823f5a56a07b9b7d23efe0331afd0e253946448e4d36abee52f7f69e9ec09: Status 404 returned error can't find the container with id 551823f5a56a07b9b7d23efe0331afd0e253946448e4d36abee52f7f69e9ec09 Nov 28 13:40:35 crc kubenswrapper[4631]: I1128 13:40:35.135207 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"cc2a0580-660d-4a19-a755-70256a4e4abe","Type":"ContainerStarted","Data":"db256804573e3b398de98f0ca594a048c43bad7d1eead634ae231a267cbb6ea0"} Nov 28 13:40:35 crc kubenswrapper[4631]: I1128 13:40:35.135752 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"cc2a0580-660d-4a19-a755-70256a4e4abe","Type":"ContainerStarted","Data":"551823f5a56a07b9b7d23efe0331afd0e253946448e4d36abee52f7f69e9ec09"} Nov 28 13:40:35 crc kubenswrapper[4631]: I1128 13:40:35.135784 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Nov 28 13:40:35 crc kubenswrapper[4631]: I1128 13:40:35.172144 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.172119918 podStartE2EDuration="2.172119918s" podCreationTimestamp="2025-11-28 13:40:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 13:40:35.159042805 +0000 UTC m=+1191.966346149" watchObservedRunningTime="2025-11-28 13:40:35.172119918 +0000 UTC m=+1191.979423262" Nov 28 13:40:37 crc kubenswrapper[4631]: E1128 13:40:37.652238 4631 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod641b1d24_f22a_4d63_8e9e_d76c3dde2135.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0f61d189_17d8_4154_ac95_df78d4042055.slice/crio-0a32d7ae2c3a534184dba524af21eb67d2088a5219dc164fd35b2643f3e54ddd\": RecentStats: unable to find data in memory cache]" Nov 28 13:40:38 crc kubenswrapper[4631]: I1128 13:40:38.136384 4631 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-544496c768-rzhsw" Nov 28 13:40:38 crc kubenswrapper[4631]: I1128 13:40:38.178408 4631 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-64f8fdf97b-jn5mg" Nov 28 13:40:39 crc kubenswrapper[4631]: I1128 13:40:39.974134 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-544496c768-rzhsw" Nov 28 13:40:40 crc kubenswrapper[4631]: I1128 13:40:40.094937 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-64f8fdf97b-jn5mg"] Nov 28 13:40:40 crc kubenswrapper[4631]: I1128 13:40:40.095224 4631 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-64f8fdf97b-jn5mg" podUID="111c5349-528d-4856-a3e4-1277e3b10889" containerName="horizon-log" containerID="cri-o://7bec69392b7f0e2c28b3311f3ec81e5bb4480735e00154f9e67c85a33543bc29" gracePeriod=30 Nov 28 13:40:40 crc kubenswrapper[4631]: I1128 13:40:40.095507 4631 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-64f8fdf97b-jn5mg" podUID="111c5349-528d-4856-a3e4-1277e3b10889" containerName="horizon" containerID="cri-o://3b826a5556ea3b27f6357d4cc74220f4bce293db6dbcb49fa5131013bef0e517" gracePeriod=30 Nov 28 13:40:40 crc kubenswrapper[4631]: I1128 13:40:40.105365 4631 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-64f8fdf97b-jn5mg" podUID="111c5349-528d-4856-a3e4-1277e3b10889" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.146:8443/dashboard/auth/login/?next=/dashboard/\": EOF" Nov 28 13:40:43 crc kubenswrapper[4631]: I1128 13:40:43.529138 4631 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-64f8fdf97b-jn5mg" podUID="111c5349-528d-4856-a3e4-1277e3b10889" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.146:8443/dashboard/auth/login/?next=/dashboard/\": read tcp 10.217.0.2:56584->10.217.0.146:8443: read: connection reset by peer" Nov 28 13:40:43 crc kubenswrapper[4631]: I1128 13:40:43.647273 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Nov 28 13:40:44 crc kubenswrapper[4631]: I1128 13:40:44.216504 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-cs6bm"] Nov 28 13:40:44 crc kubenswrapper[4631]: I1128 13:40:44.218956 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-cs6bm" Nov 28 13:40:44 crc kubenswrapper[4631]: I1128 13:40:44.223596 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Nov 28 13:40:44 crc kubenswrapper[4631]: I1128 13:40:44.224273 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Nov 28 13:40:44 crc kubenswrapper[4631]: I1128 13:40:44.273861 4631 generic.go:334] "Generic (PLEG): container finished" podID="111c5349-528d-4856-a3e4-1277e3b10889" containerID="3b826a5556ea3b27f6357d4cc74220f4bce293db6dbcb49fa5131013bef0e517" exitCode=0 Nov 28 13:40:44 crc kubenswrapper[4631]: I1128 13:40:44.274336 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-64f8fdf97b-jn5mg" event={"ID":"111c5349-528d-4856-a3e4-1277e3b10889","Type":"ContainerDied","Data":"3b826a5556ea3b27f6357d4cc74220f4bce293db6dbcb49fa5131013bef0e517"} Nov 28 13:40:44 crc kubenswrapper[4631]: I1128 13:40:44.274391 4631 scope.go:117] "RemoveContainer" containerID="ddca46c54a9ba33270a5c14213a03ffc4ee751ec03afaccc3b37cfa0eac87eb7" Nov 28 13:40:44 crc kubenswrapper[4631]: I1128 13:40:44.282021 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-cs6bm"] Nov 28 13:40:44 crc kubenswrapper[4631]: I1128 13:40:44.347805 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s7fj8\" (UniqueName: \"kubernetes.io/projected/476e6d13-87a1-4155-ae6b-f6453e9c2044-kube-api-access-s7fj8\") pod \"nova-cell0-cell-mapping-cs6bm\" (UID: \"476e6d13-87a1-4155-ae6b-f6453e9c2044\") " pod="openstack/nova-cell0-cell-mapping-cs6bm" Nov 28 13:40:44 crc kubenswrapper[4631]: I1128 13:40:44.347917 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/476e6d13-87a1-4155-ae6b-f6453e9c2044-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-cs6bm\" (UID: \"476e6d13-87a1-4155-ae6b-f6453e9c2044\") " pod="openstack/nova-cell0-cell-mapping-cs6bm" Nov 28 13:40:44 crc kubenswrapper[4631]: I1128 13:40:44.347948 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/476e6d13-87a1-4155-ae6b-f6453e9c2044-config-data\") pod \"nova-cell0-cell-mapping-cs6bm\" (UID: \"476e6d13-87a1-4155-ae6b-f6453e9c2044\") " pod="openstack/nova-cell0-cell-mapping-cs6bm" Nov 28 13:40:44 crc kubenswrapper[4631]: I1128 13:40:44.347970 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/476e6d13-87a1-4155-ae6b-f6453e9c2044-scripts\") pod \"nova-cell0-cell-mapping-cs6bm\" (UID: \"476e6d13-87a1-4155-ae6b-f6453e9c2044\") " pod="openstack/nova-cell0-cell-mapping-cs6bm" Nov 28 13:40:44 crc kubenswrapper[4631]: I1128 13:40:44.450552 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Nov 28 13:40:44 crc kubenswrapper[4631]: I1128 13:40:44.452645 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 28 13:40:44 crc kubenswrapper[4631]: I1128 13:40:44.452893 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/476e6d13-87a1-4155-ae6b-f6453e9c2044-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-cs6bm\" (UID: \"476e6d13-87a1-4155-ae6b-f6453e9c2044\") " pod="openstack/nova-cell0-cell-mapping-cs6bm" Nov 28 13:40:44 crc kubenswrapper[4631]: I1128 13:40:44.452965 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/476e6d13-87a1-4155-ae6b-f6453e9c2044-config-data\") pod \"nova-cell0-cell-mapping-cs6bm\" (UID: \"476e6d13-87a1-4155-ae6b-f6453e9c2044\") " pod="openstack/nova-cell0-cell-mapping-cs6bm" Nov 28 13:40:44 crc kubenswrapper[4631]: I1128 13:40:44.452987 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/476e6d13-87a1-4155-ae6b-f6453e9c2044-scripts\") pod \"nova-cell0-cell-mapping-cs6bm\" (UID: \"476e6d13-87a1-4155-ae6b-f6453e9c2044\") " pod="openstack/nova-cell0-cell-mapping-cs6bm" Nov 28 13:40:44 crc kubenswrapper[4631]: I1128 13:40:44.453105 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s7fj8\" (UniqueName: \"kubernetes.io/projected/476e6d13-87a1-4155-ae6b-f6453e9c2044-kube-api-access-s7fj8\") pod \"nova-cell0-cell-mapping-cs6bm\" (UID: \"476e6d13-87a1-4155-ae6b-f6453e9c2044\") " pod="openstack/nova-cell0-cell-mapping-cs6bm" Nov 28 13:40:44 crc kubenswrapper[4631]: I1128 13:40:44.463150 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Nov 28 13:40:44 crc kubenswrapper[4631]: I1128 13:40:44.465214 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/476e6d13-87a1-4155-ae6b-f6453e9c2044-scripts\") pod \"nova-cell0-cell-mapping-cs6bm\" (UID: \"476e6d13-87a1-4155-ae6b-f6453e9c2044\") " pod="openstack/nova-cell0-cell-mapping-cs6bm" Nov 28 13:40:44 crc kubenswrapper[4631]: I1128 13:40:44.469107 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/476e6d13-87a1-4155-ae6b-f6453e9c2044-config-data\") pod \"nova-cell0-cell-mapping-cs6bm\" (UID: \"476e6d13-87a1-4155-ae6b-f6453e9c2044\") " pod="openstack/nova-cell0-cell-mapping-cs6bm" Nov 28 13:40:44 crc kubenswrapper[4631]: I1128 13:40:44.492124 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/476e6d13-87a1-4155-ae6b-f6453e9c2044-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-cs6bm\" (UID: \"476e6d13-87a1-4155-ae6b-f6453e9c2044\") " pod="openstack/nova-cell0-cell-mapping-cs6bm" Nov 28 13:40:44 crc kubenswrapper[4631]: I1128 13:40:44.504175 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s7fj8\" (UniqueName: \"kubernetes.io/projected/476e6d13-87a1-4155-ae6b-f6453e9c2044-kube-api-access-s7fj8\") pod \"nova-cell0-cell-mapping-cs6bm\" (UID: \"476e6d13-87a1-4155-ae6b-f6453e9c2044\") " pod="openstack/nova-cell0-cell-mapping-cs6bm" Nov 28 13:40:44 crc kubenswrapper[4631]: I1128 13:40:44.509338 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 28 13:40:44 crc kubenswrapper[4631]: I1128 13:40:44.511024 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 28 13:40:44 crc kubenswrapper[4631]: I1128 13:40:44.515373 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Nov 28 13:40:44 crc kubenswrapper[4631]: I1128 13:40:44.555143 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c7ac2d63-abe0-4862-a3a2-5d2b1dc87ac9-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"c7ac2d63-abe0-4862-a3a2-5d2b1dc87ac9\") " pod="openstack/nova-api-0" Nov 28 13:40:44 crc kubenswrapper[4631]: I1128 13:40:44.555242 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xbsqb\" (UniqueName: \"kubernetes.io/projected/c7ac2d63-abe0-4862-a3a2-5d2b1dc87ac9-kube-api-access-xbsqb\") pod \"nova-api-0\" (UID: \"c7ac2d63-abe0-4862-a3a2-5d2b1dc87ac9\") " pod="openstack/nova-api-0" Nov 28 13:40:44 crc kubenswrapper[4631]: I1128 13:40:44.555306 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c7ac2d63-abe0-4862-a3a2-5d2b1dc87ac9-logs\") pod \"nova-api-0\" (UID: \"c7ac2d63-abe0-4862-a3a2-5d2b1dc87ac9\") " pod="openstack/nova-api-0" Nov 28 13:40:44 crc kubenswrapper[4631]: I1128 13:40:44.555369 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c7ac2d63-abe0-4862-a3a2-5d2b1dc87ac9-config-data\") pod \"nova-api-0\" (UID: \"c7ac2d63-abe0-4862-a3a2-5d2b1dc87ac9\") " pod="openstack/nova-api-0" Nov 28 13:40:44 crc kubenswrapper[4631]: I1128 13:40:44.563601 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-cs6bm" Nov 28 13:40:44 crc kubenswrapper[4631]: I1128 13:40:44.573439 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 28 13:40:44 crc kubenswrapper[4631]: I1128 13:40:44.633340 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 28 13:40:44 crc kubenswrapper[4631]: I1128 13:40:44.664218 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c7ac2d63-abe0-4862-a3a2-5d2b1dc87ac9-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"c7ac2d63-abe0-4862-a3a2-5d2b1dc87ac9\") " pod="openstack/nova-api-0" Nov 28 13:40:44 crc kubenswrapper[4631]: I1128 13:40:44.664306 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e8a60bc-6e73-499a-959a-c0f104014641-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"5e8a60bc-6e73-499a-959a-c0f104014641\") " pod="openstack/nova-cell1-novncproxy-0" Nov 28 13:40:44 crc kubenswrapper[4631]: I1128 13:40:44.664341 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xbsqb\" (UniqueName: \"kubernetes.io/projected/c7ac2d63-abe0-4862-a3a2-5d2b1dc87ac9-kube-api-access-xbsqb\") pod \"nova-api-0\" (UID: \"c7ac2d63-abe0-4862-a3a2-5d2b1dc87ac9\") " pod="openstack/nova-api-0" Nov 28 13:40:44 crc kubenswrapper[4631]: I1128 13:40:44.664358 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5e8a60bc-6e73-499a-959a-c0f104014641-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"5e8a60bc-6e73-499a-959a-c0f104014641\") " pod="openstack/nova-cell1-novncproxy-0" Nov 28 13:40:44 crc kubenswrapper[4631]: I1128 13:40:44.664404 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c7ac2d63-abe0-4862-a3a2-5d2b1dc87ac9-logs\") pod \"nova-api-0\" (UID: \"c7ac2d63-abe0-4862-a3a2-5d2b1dc87ac9\") " pod="openstack/nova-api-0" Nov 28 13:40:44 crc kubenswrapper[4631]: I1128 13:40:44.664431 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f2fhf\" (UniqueName: \"kubernetes.io/projected/5e8a60bc-6e73-499a-959a-c0f104014641-kube-api-access-f2fhf\") pod \"nova-cell1-novncproxy-0\" (UID: \"5e8a60bc-6e73-499a-959a-c0f104014641\") " pod="openstack/nova-cell1-novncproxy-0" Nov 28 13:40:44 crc kubenswrapper[4631]: I1128 13:40:44.664480 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c7ac2d63-abe0-4862-a3a2-5d2b1dc87ac9-config-data\") pod \"nova-api-0\" (UID: \"c7ac2d63-abe0-4862-a3a2-5d2b1dc87ac9\") " pod="openstack/nova-api-0" Nov 28 13:40:44 crc kubenswrapper[4631]: I1128 13:40:44.665646 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c7ac2d63-abe0-4862-a3a2-5d2b1dc87ac9-logs\") pod \"nova-api-0\" (UID: \"c7ac2d63-abe0-4862-a3a2-5d2b1dc87ac9\") " pod="openstack/nova-api-0" Nov 28 13:40:44 crc kubenswrapper[4631]: I1128 13:40:44.671387 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c7ac2d63-abe0-4862-a3a2-5d2b1dc87ac9-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"c7ac2d63-abe0-4862-a3a2-5d2b1dc87ac9\") " pod="openstack/nova-api-0" Nov 28 13:40:44 crc kubenswrapper[4631]: I1128 13:40:44.673054 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c7ac2d63-abe0-4862-a3a2-5d2b1dc87ac9-config-data\") pod \"nova-api-0\" (UID: \"c7ac2d63-abe0-4862-a3a2-5d2b1dc87ac9\") " pod="openstack/nova-api-0" Nov 28 13:40:44 crc kubenswrapper[4631]: I1128 13:40:44.700063 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xbsqb\" (UniqueName: \"kubernetes.io/projected/c7ac2d63-abe0-4862-a3a2-5d2b1dc87ac9-kube-api-access-xbsqb\") pod \"nova-api-0\" (UID: \"c7ac2d63-abe0-4862-a3a2-5d2b1dc87ac9\") " pod="openstack/nova-api-0" Nov 28 13:40:44 crc kubenswrapper[4631]: I1128 13:40:44.744601 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Nov 28 13:40:44 crc kubenswrapper[4631]: I1128 13:40:44.746714 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 28 13:40:44 crc kubenswrapper[4631]: I1128 13:40:44.750774 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Nov 28 13:40:44 crc kubenswrapper[4631]: I1128 13:40:44.766750 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e8a60bc-6e73-499a-959a-c0f104014641-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"5e8a60bc-6e73-499a-959a-c0f104014641\") " pod="openstack/nova-cell1-novncproxy-0" Nov 28 13:40:44 crc kubenswrapper[4631]: I1128 13:40:44.766831 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5e8a60bc-6e73-499a-959a-c0f104014641-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"5e8a60bc-6e73-499a-959a-c0f104014641\") " pod="openstack/nova-cell1-novncproxy-0" Nov 28 13:40:44 crc kubenswrapper[4631]: I1128 13:40:44.766890 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f2fhf\" (UniqueName: \"kubernetes.io/projected/5e8a60bc-6e73-499a-959a-c0f104014641-kube-api-access-f2fhf\") pod \"nova-cell1-novncproxy-0\" (UID: \"5e8a60bc-6e73-499a-959a-c0f104014641\") " pod="openstack/nova-cell1-novncproxy-0" Nov 28 13:40:44 crc kubenswrapper[4631]: I1128 13:40:44.771568 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5e8a60bc-6e73-499a-959a-c0f104014641-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"5e8a60bc-6e73-499a-959a-c0f104014641\") " pod="openstack/nova-cell1-novncproxy-0" Nov 28 13:40:44 crc kubenswrapper[4631]: I1128 13:40:44.786255 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e8a60bc-6e73-499a-959a-c0f104014641-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"5e8a60bc-6e73-499a-959a-c0f104014641\") " pod="openstack/nova-cell1-novncproxy-0" Nov 28 13:40:44 crc kubenswrapper[4631]: I1128 13:40:44.789780 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 28 13:40:44 crc kubenswrapper[4631]: I1128 13:40:44.819498 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Nov 28 13:40:44 crc kubenswrapper[4631]: I1128 13:40:44.821566 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 28 13:40:44 crc kubenswrapper[4631]: I1128 13:40:44.835185 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Nov 28 13:40:44 crc kubenswrapper[4631]: I1128 13:40:44.840123 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f2fhf\" (UniqueName: \"kubernetes.io/projected/5e8a60bc-6e73-499a-959a-c0f104014641-kube-api-access-f2fhf\") pod \"nova-cell1-novncproxy-0\" (UID: \"5e8a60bc-6e73-499a-959a-c0f104014641\") " pod="openstack/nova-cell1-novncproxy-0" Nov 28 13:40:44 crc kubenswrapper[4631]: I1128 13:40:44.870450 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qqsml\" (UniqueName: \"kubernetes.io/projected/e9e6bbdb-eac5-47db-80e5-b438965fa8aa-kube-api-access-qqsml\") pod \"nova-scheduler-0\" (UID: \"e9e6bbdb-eac5-47db-80e5-b438965fa8aa\") " pod="openstack/nova-scheduler-0" Nov 28 13:40:44 crc kubenswrapper[4631]: I1128 13:40:44.870764 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e9e6bbdb-eac5-47db-80e5-b438965fa8aa-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"e9e6bbdb-eac5-47db-80e5-b438965fa8aa\") " pod="openstack/nova-scheduler-0" Nov 28 13:40:44 crc kubenswrapper[4631]: I1128 13:40:44.870861 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e9e6bbdb-eac5-47db-80e5-b438965fa8aa-config-data\") pod \"nova-scheduler-0\" (UID: \"e9e6bbdb-eac5-47db-80e5-b438965fa8aa\") " pod="openstack/nova-scheduler-0" Nov 28 13:40:44 crc kubenswrapper[4631]: I1128 13:40:44.884900 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 28 13:40:44 crc kubenswrapper[4631]: I1128 13:40:44.896248 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 28 13:40:44 crc kubenswrapper[4631]: I1128 13:40:44.974242 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qqsml\" (UniqueName: \"kubernetes.io/projected/e9e6bbdb-eac5-47db-80e5-b438965fa8aa-kube-api-access-qqsml\") pod \"nova-scheduler-0\" (UID: \"e9e6bbdb-eac5-47db-80e5-b438965fa8aa\") " pod="openstack/nova-scheduler-0" Nov 28 13:40:44 crc kubenswrapper[4631]: I1128 13:40:44.974440 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-57cmk\" (UniqueName: \"kubernetes.io/projected/cf5855aa-20c9-4b9c-8205-dbd3dcaa8917-kube-api-access-57cmk\") pod \"nova-metadata-0\" (UID: \"cf5855aa-20c9-4b9c-8205-dbd3dcaa8917\") " pod="openstack/nova-metadata-0" Nov 28 13:40:44 crc kubenswrapper[4631]: I1128 13:40:44.974530 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e9e6bbdb-eac5-47db-80e5-b438965fa8aa-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"e9e6bbdb-eac5-47db-80e5-b438965fa8aa\") " pod="openstack/nova-scheduler-0" Nov 28 13:40:44 crc kubenswrapper[4631]: I1128 13:40:44.974592 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e9e6bbdb-eac5-47db-80e5-b438965fa8aa-config-data\") pod \"nova-scheduler-0\" (UID: \"e9e6bbdb-eac5-47db-80e5-b438965fa8aa\") " pod="openstack/nova-scheduler-0" Nov 28 13:40:44 crc kubenswrapper[4631]: I1128 13:40:44.974767 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cf5855aa-20c9-4b9c-8205-dbd3dcaa8917-logs\") pod \"nova-metadata-0\" (UID: \"cf5855aa-20c9-4b9c-8205-dbd3dcaa8917\") " pod="openstack/nova-metadata-0" Nov 28 13:40:45 crc kubenswrapper[4631]: I1128 13:40:44.998068 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cf5855aa-20c9-4b9c-8205-dbd3dcaa8917-config-data\") pod \"nova-metadata-0\" (UID: \"cf5855aa-20c9-4b9c-8205-dbd3dcaa8917\") " pod="openstack/nova-metadata-0" Nov 28 13:40:45 crc kubenswrapper[4631]: I1128 13:40:45.007501 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 28 13:40:45 crc kubenswrapper[4631]: I1128 13:40:45.019946 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cf5855aa-20c9-4b9c-8205-dbd3dcaa8917-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"cf5855aa-20c9-4b9c-8205-dbd3dcaa8917\") " pod="openstack/nova-metadata-0" Nov 28 13:40:45 crc kubenswrapper[4631]: I1128 13:40:45.026091 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e9e6bbdb-eac5-47db-80e5-b438965fa8aa-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"e9e6bbdb-eac5-47db-80e5-b438965fa8aa\") " pod="openstack/nova-scheduler-0" Nov 28 13:40:45 crc kubenswrapper[4631]: I1128 13:40:45.027189 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e9e6bbdb-eac5-47db-80e5-b438965fa8aa-config-data\") pod \"nova-scheduler-0\" (UID: \"e9e6bbdb-eac5-47db-80e5-b438965fa8aa\") " pod="openstack/nova-scheduler-0" Nov 28 13:40:45 crc kubenswrapper[4631]: I1128 13:40:45.046067 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qqsml\" (UniqueName: \"kubernetes.io/projected/e9e6bbdb-eac5-47db-80e5-b438965fa8aa-kube-api-access-qqsml\") pod \"nova-scheduler-0\" (UID: \"e9e6bbdb-eac5-47db-80e5-b438965fa8aa\") " pod="openstack/nova-scheduler-0" Nov 28 13:40:45 crc kubenswrapper[4631]: I1128 13:40:45.061870 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-757b4f8459-j7gzn"] Nov 28 13:40:45 crc kubenswrapper[4631]: I1128 13:40:45.065319 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-757b4f8459-j7gzn" Nov 28 13:40:45 crc kubenswrapper[4631]: I1128 13:40:45.128941 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cf5855aa-20c9-4b9c-8205-dbd3dcaa8917-logs\") pod \"nova-metadata-0\" (UID: \"cf5855aa-20c9-4b9c-8205-dbd3dcaa8917\") " pod="openstack/nova-metadata-0" Nov 28 13:40:45 crc kubenswrapper[4631]: I1128 13:40:45.129423 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cf5855aa-20c9-4b9c-8205-dbd3dcaa8917-config-data\") pod \"nova-metadata-0\" (UID: \"cf5855aa-20c9-4b9c-8205-dbd3dcaa8917\") " pod="openstack/nova-metadata-0" Nov 28 13:40:45 crc kubenswrapper[4631]: I1128 13:40:45.129519 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cf5855aa-20c9-4b9c-8205-dbd3dcaa8917-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"cf5855aa-20c9-4b9c-8205-dbd3dcaa8917\") " pod="openstack/nova-metadata-0" Nov 28 13:40:45 crc kubenswrapper[4631]: I1128 13:40:45.129577 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-57cmk\" (UniqueName: \"kubernetes.io/projected/cf5855aa-20c9-4b9c-8205-dbd3dcaa8917-kube-api-access-57cmk\") pod \"nova-metadata-0\" (UID: \"cf5855aa-20c9-4b9c-8205-dbd3dcaa8917\") " pod="openstack/nova-metadata-0" Nov 28 13:40:45 crc kubenswrapper[4631]: I1128 13:40:45.129612 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cf5855aa-20c9-4b9c-8205-dbd3dcaa8917-logs\") pod \"nova-metadata-0\" (UID: \"cf5855aa-20c9-4b9c-8205-dbd3dcaa8917\") " pod="openstack/nova-metadata-0" Nov 28 13:40:45 crc kubenswrapper[4631]: I1128 13:40:45.130492 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-757b4f8459-j7gzn"] Nov 28 13:40:45 crc kubenswrapper[4631]: I1128 13:40:45.140276 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cf5855aa-20c9-4b9c-8205-dbd3dcaa8917-config-data\") pod \"nova-metadata-0\" (UID: \"cf5855aa-20c9-4b9c-8205-dbd3dcaa8917\") " pod="openstack/nova-metadata-0" Nov 28 13:40:45 crc kubenswrapper[4631]: I1128 13:40:45.158149 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cf5855aa-20c9-4b9c-8205-dbd3dcaa8917-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"cf5855aa-20c9-4b9c-8205-dbd3dcaa8917\") " pod="openstack/nova-metadata-0" Nov 28 13:40:45 crc kubenswrapper[4631]: I1128 13:40:45.173682 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-57cmk\" (UniqueName: \"kubernetes.io/projected/cf5855aa-20c9-4b9c-8205-dbd3dcaa8917-kube-api-access-57cmk\") pod \"nova-metadata-0\" (UID: \"cf5855aa-20c9-4b9c-8205-dbd3dcaa8917\") " pod="openstack/nova-metadata-0" Nov 28 13:40:45 crc kubenswrapper[4631]: I1128 13:40:45.231555 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e1644d33-bda9-437e-b10c-5c48042cb1f2-config\") pod \"dnsmasq-dns-757b4f8459-j7gzn\" (UID: \"e1644d33-bda9-437e-b10c-5c48042cb1f2\") " pod="openstack/dnsmasq-dns-757b4f8459-j7gzn" Nov 28 13:40:45 crc kubenswrapper[4631]: I1128 13:40:45.231651 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e1644d33-bda9-437e-b10c-5c48042cb1f2-ovsdbserver-nb\") pod \"dnsmasq-dns-757b4f8459-j7gzn\" (UID: \"e1644d33-bda9-437e-b10c-5c48042cb1f2\") " pod="openstack/dnsmasq-dns-757b4f8459-j7gzn" Nov 28 13:40:45 crc kubenswrapper[4631]: I1128 13:40:45.231727 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e1644d33-bda9-437e-b10c-5c48042cb1f2-dns-svc\") pod \"dnsmasq-dns-757b4f8459-j7gzn\" (UID: \"e1644d33-bda9-437e-b10c-5c48042cb1f2\") " pod="openstack/dnsmasq-dns-757b4f8459-j7gzn" Nov 28 13:40:45 crc kubenswrapper[4631]: I1128 13:40:45.231758 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e1644d33-bda9-437e-b10c-5c48042cb1f2-dns-swift-storage-0\") pod \"dnsmasq-dns-757b4f8459-j7gzn\" (UID: \"e1644d33-bda9-437e-b10c-5c48042cb1f2\") " pod="openstack/dnsmasq-dns-757b4f8459-j7gzn" Nov 28 13:40:45 crc kubenswrapper[4631]: I1128 13:40:45.231821 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zz4pn\" (UniqueName: \"kubernetes.io/projected/e1644d33-bda9-437e-b10c-5c48042cb1f2-kube-api-access-zz4pn\") pod \"dnsmasq-dns-757b4f8459-j7gzn\" (UID: \"e1644d33-bda9-437e-b10c-5c48042cb1f2\") " pod="openstack/dnsmasq-dns-757b4f8459-j7gzn" Nov 28 13:40:45 crc kubenswrapper[4631]: I1128 13:40:45.231873 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e1644d33-bda9-437e-b10c-5c48042cb1f2-ovsdbserver-sb\") pod \"dnsmasq-dns-757b4f8459-j7gzn\" (UID: \"e1644d33-bda9-437e-b10c-5c48042cb1f2\") " pod="openstack/dnsmasq-dns-757b4f8459-j7gzn" Nov 28 13:40:45 crc kubenswrapper[4631]: I1128 13:40:45.246414 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 28 13:40:45 crc kubenswrapper[4631]: I1128 13:40:45.288939 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 28 13:40:45 crc kubenswrapper[4631]: I1128 13:40:45.341297 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e1644d33-bda9-437e-b10c-5c48042cb1f2-config\") pod \"dnsmasq-dns-757b4f8459-j7gzn\" (UID: \"e1644d33-bda9-437e-b10c-5c48042cb1f2\") " pod="openstack/dnsmasq-dns-757b4f8459-j7gzn" Nov 28 13:40:45 crc kubenswrapper[4631]: I1128 13:40:45.341366 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e1644d33-bda9-437e-b10c-5c48042cb1f2-ovsdbserver-nb\") pod \"dnsmasq-dns-757b4f8459-j7gzn\" (UID: \"e1644d33-bda9-437e-b10c-5c48042cb1f2\") " pod="openstack/dnsmasq-dns-757b4f8459-j7gzn" Nov 28 13:40:45 crc kubenswrapper[4631]: I1128 13:40:45.341440 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e1644d33-bda9-437e-b10c-5c48042cb1f2-dns-svc\") pod \"dnsmasq-dns-757b4f8459-j7gzn\" (UID: \"e1644d33-bda9-437e-b10c-5c48042cb1f2\") " pod="openstack/dnsmasq-dns-757b4f8459-j7gzn" Nov 28 13:40:45 crc kubenswrapper[4631]: I1128 13:40:45.341473 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e1644d33-bda9-437e-b10c-5c48042cb1f2-dns-swift-storage-0\") pod \"dnsmasq-dns-757b4f8459-j7gzn\" (UID: \"e1644d33-bda9-437e-b10c-5c48042cb1f2\") " pod="openstack/dnsmasq-dns-757b4f8459-j7gzn" Nov 28 13:40:45 crc kubenswrapper[4631]: I1128 13:40:45.341529 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zz4pn\" (UniqueName: \"kubernetes.io/projected/e1644d33-bda9-437e-b10c-5c48042cb1f2-kube-api-access-zz4pn\") pod \"dnsmasq-dns-757b4f8459-j7gzn\" (UID: \"e1644d33-bda9-437e-b10c-5c48042cb1f2\") " pod="openstack/dnsmasq-dns-757b4f8459-j7gzn" Nov 28 13:40:45 crc kubenswrapper[4631]: I1128 13:40:45.341585 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e1644d33-bda9-437e-b10c-5c48042cb1f2-ovsdbserver-sb\") pod \"dnsmasq-dns-757b4f8459-j7gzn\" (UID: \"e1644d33-bda9-437e-b10c-5c48042cb1f2\") " pod="openstack/dnsmasq-dns-757b4f8459-j7gzn" Nov 28 13:40:45 crc kubenswrapper[4631]: I1128 13:40:45.342731 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e1644d33-bda9-437e-b10c-5c48042cb1f2-ovsdbserver-sb\") pod \"dnsmasq-dns-757b4f8459-j7gzn\" (UID: \"e1644d33-bda9-437e-b10c-5c48042cb1f2\") " pod="openstack/dnsmasq-dns-757b4f8459-j7gzn" Nov 28 13:40:45 crc kubenswrapper[4631]: I1128 13:40:45.342776 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e1644d33-bda9-437e-b10c-5c48042cb1f2-dns-svc\") pod \"dnsmasq-dns-757b4f8459-j7gzn\" (UID: \"e1644d33-bda9-437e-b10c-5c48042cb1f2\") " pod="openstack/dnsmasq-dns-757b4f8459-j7gzn" Nov 28 13:40:45 crc kubenswrapper[4631]: I1128 13:40:45.343553 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e1644d33-bda9-437e-b10c-5c48042cb1f2-config\") pod \"dnsmasq-dns-757b4f8459-j7gzn\" (UID: \"e1644d33-bda9-437e-b10c-5c48042cb1f2\") " pod="openstack/dnsmasq-dns-757b4f8459-j7gzn" Nov 28 13:40:45 crc kubenswrapper[4631]: I1128 13:40:45.344105 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e1644d33-bda9-437e-b10c-5c48042cb1f2-ovsdbserver-nb\") pod \"dnsmasq-dns-757b4f8459-j7gzn\" (UID: \"e1644d33-bda9-437e-b10c-5c48042cb1f2\") " pod="openstack/dnsmasq-dns-757b4f8459-j7gzn" Nov 28 13:40:45 crc kubenswrapper[4631]: I1128 13:40:45.344119 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e1644d33-bda9-437e-b10c-5c48042cb1f2-dns-swift-storage-0\") pod \"dnsmasq-dns-757b4f8459-j7gzn\" (UID: \"e1644d33-bda9-437e-b10c-5c48042cb1f2\") " pod="openstack/dnsmasq-dns-757b4f8459-j7gzn" Nov 28 13:40:45 crc kubenswrapper[4631]: I1128 13:40:45.347636 4631 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-64f8fdf97b-jn5mg" podUID="111c5349-528d-4856-a3e4-1277e3b10889" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.146:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.146:8443: connect: connection refused" Nov 28 13:40:45 crc kubenswrapper[4631]: I1128 13:40:45.367208 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zz4pn\" (UniqueName: \"kubernetes.io/projected/e1644d33-bda9-437e-b10c-5c48042cb1f2-kube-api-access-zz4pn\") pod \"dnsmasq-dns-757b4f8459-j7gzn\" (UID: \"e1644d33-bda9-437e-b10c-5c48042cb1f2\") " pod="openstack/dnsmasq-dns-757b4f8459-j7gzn" Nov 28 13:40:45 crc kubenswrapper[4631]: I1128 13:40:45.431545 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-757b4f8459-j7gzn" Nov 28 13:40:45 crc kubenswrapper[4631]: I1128 13:40:45.612642 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-cs6bm"] Nov 28 13:40:45 crc kubenswrapper[4631]: I1128 13:40:45.873683 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 28 13:40:45 crc kubenswrapper[4631]: I1128 13:40:45.885893 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 28 13:40:46 crc kubenswrapper[4631]: I1128 13:40:46.248197 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 28 13:40:46 crc kubenswrapper[4631]: I1128 13:40:46.394473 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 28 13:40:46 crc kubenswrapper[4631]: I1128 13:40:46.408517 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"cf5855aa-20c9-4b9c-8205-dbd3dcaa8917","Type":"ContainerStarted","Data":"08593af7191fdacda9ad743065d8b3b03c47681de484335e1159135eb68e6797"} Nov 28 13:40:46 crc kubenswrapper[4631]: I1128 13:40:46.420089 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"5e8a60bc-6e73-499a-959a-c0f104014641","Type":"ContainerStarted","Data":"131605dec7c38ec98c2deaa0be880784fabc50133f77ad682d68346450009164"} Nov 28 13:40:46 crc kubenswrapper[4631]: I1128 13:40:46.429804 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"e9e6bbdb-eac5-47db-80e5-b438965fa8aa","Type":"ContainerStarted","Data":"541285d216b670f6796f84f4d8b4baf48e4e3762fd84e3d165c8c29734d32012"} Nov 28 13:40:46 crc kubenswrapper[4631]: I1128 13:40:46.441189 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"c7ac2d63-abe0-4862-a3a2-5d2b1dc87ac9","Type":"ContainerStarted","Data":"ff28b82ee28ede9ce6b12af991af71dee06a6d4d8170bdf238238bf642c83a35"} Nov 28 13:40:46 crc kubenswrapper[4631]: I1128 13:40:46.458149 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-cs6bm" event={"ID":"476e6d13-87a1-4155-ae6b-f6453e9c2044","Type":"ContainerStarted","Data":"44bdbae8c83b2de9874d7d4a4580cd7185b4f75b99e0304faaeb9f093af2f0d2"} Nov 28 13:40:46 crc kubenswrapper[4631]: I1128 13:40:46.458224 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-cs6bm" event={"ID":"476e6d13-87a1-4155-ae6b-f6453e9c2044","Type":"ContainerStarted","Data":"581a79913f03d0080cc808bc6911303568135f1d477a940396f37410ddadfd83"} Nov 28 13:40:46 crc kubenswrapper[4631]: I1128 13:40:46.484782 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-cs6bm" podStartSLOduration=2.484755741 podStartE2EDuration="2.484755741s" podCreationTimestamp="2025-11-28 13:40:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 13:40:46.481932094 +0000 UTC m=+1203.289235458" watchObservedRunningTime="2025-11-28 13:40:46.484755741 +0000 UTC m=+1203.292059085" Nov 28 13:40:46 crc kubenswrapper[4631]: I1128 13:40:46.632428 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-757b4f8459-j7gzn"] Nov 28 13:40:46 crc kubenswrapper[4631]: I1128 13:40:46.690386 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-tn4hn"] Nov 28 13:40:46 crc kubenswrapper[4631]: I1128 13:40:46.693637 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-tn4hn" Nov 28 13:40:46 crc kubenswrapper[4631]: I1128 13:40:46.697355 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Nov 28 13:40:46 crc kubenswrapper[4631]: I1128 13:40:46.697808 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Nov 28 13:40:46 crc kubenswrapper[4631]: I1128 13:40:46.742244 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6t89z\" (UniqueName: \"kubernetes.io/projected/6b7cc6f7-1a01-4c53-818d-f40548209a13-kube-api-access-6t89z\") pod \"nova-cell1-conductor-db-sync-tn4hn\" (UID: \"6b7cc6f7-1a01-4c53-818d-f40548209a13\") " pod="openstack/nova-cell1-conductor-db-sync-tn4hn" Nov 28 13:40:46 crc kubenswrapper[4631]: I1128 13:40:46.746387 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-tn4hn"] Nov 28 13:40:46 crc kubenswrapper[4631]: I1128 13:40:46.757259 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6b7cc6f7-1a01-4c53-818d-f40548209a13-config-data\") pod \"nova-cell1-conductor-db-sync-tn4hn\" (UID: \"6b7cc6f7-1a01-4c53-818d-f40548209a13\") " pod="openstack/nova-cell1-conductor-db-sync-tn4hn" Nov 28 13:40:46 crc kubenswrapper[4631]: I1128 13:40:46.757536 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b7cc6f7-1a01-4c53-818d-f40548209a13-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-tn4hn\" (UID: \"6b7cc6f7-1a01-4c53-818d-f40548209a13\") " pod="openstack/nova-cell1-conductor-db-sync-tn4hn" Nov 28 13:40:46 crc kubenswrapper[4631]: I1128 13:40:46.757850 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6b7cc6f7-1a01-4c53-818d-f40548209a13-scripts\") pod \"nova-cell1-conductor-db-sync-tn4hn\" (UID: \"6b7cc6f7-1a01-4c53-818d-f40548209a13\") " pod="openstack/nova-cell1-conductor-db-sync-tn4hn" Nov 28 13:40:46 crc kubenswrapper[4631]: I1128 13:40:46.861242 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6b7cc6f7-1a01-4c53-818d-f40548209a13-config-data\") pod \"nova-cell1-conductor-db-sync-tn4hn\" (UID: \"6b7cc6f7-1a01-4c53-818d-f40548209a13\") " pod="openstack/nova-cell1-conductor-db-sync-tn4hn" Nov 28 13:40:46 crc kubenswrapper[4631]: I1128 13:40:46.861317 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b7cc6f7-1a01-4c53-818d-f40548209a13-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-tn4hn\" (UID: \"6b7cc6f7-1a01-4c53-818d-f40548209a13\") " pod="openstack/nova-cell1-conductor-db-sync-tn4hn" Nov 28 13:40:46 crc kubenswrapper[4631]: I1128 13:40:46.861396 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6b7cc6f7-1a01-4c53-818d-f40548209a13-scripts\") pod \"nova-cell1-conductor-db-sync-tn4hn\" (UID: \"6b7cc6f7-1a01-4c53-818d-f40548209a13\") " pod="openstack/nova-cell1-conductor-db-sync-tn4hn" Nov 28 13:40:46 crc kubenswrapper[4631]: I1128 13:40:46.861429 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6t89z\" (UniqueName: \"kubernetes.io/projected/6b7cc6f7-1a01-4c53-818d-f40548209a13-kube-api-access-6t89z\") pod \"nova-cell1-conductor-db-sync-tn4hn\" (UID: \"6b7cc6f7-1a01-4c53-818d-f40548209a13\") " pod="openstack/nova-cell1-conductor-db-sync-tn4hn" Nov 28 13:40:46 crc kubenswrapper[4631]: I1128 13:40:46.871534 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6b7cc6f7-1a01-4c53-818d-f40548209a13-config-data\") pod \"nova-cell1-conductor-db-sync-tn4hn\" (UID: \"6b7cc6f7-1a01-4c53-818d-f40548209a13\") " pod="openstack/nova-cell1-conductor-db-sync-tn4hn" Nov 28 13:40:46 crc kubenswrapper[4631]: I1128 13:40:46.878404 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6b7cc6f7-1a01-4c53-818d-f40548209a13-scripts\") pod \"nova-cell1-conductor-db-sync-tn4hn\" (UID: \"6b7cc6f7-1a01-4c53-818d-f40548209a13\") " pod="openstack/nova-cell1-conductor-db-sync-tn4hn" Nov 28 13:40:46 crc kubenswrapper[4631]: I1128 13:40:46.881123 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b7cc6f7-1a01-4c53-818d-f40548209a13-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-tn4hn\" (UID: \"6b7cc6f7-1a01-4c53-818d-f40548209a13\") " pod="openstack/nova-cell1-conductor-db-sync-tn4hn" Nov 28 13:40:46 crc kubenswrapper[4631]: I1128 13:40:46.889897 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6t89z\" (UniqueName: \"kubernetes.io/projected/6b7cc6f7-1a01-4c53-818d-f40548209a13-kube-api-access-6t89z\") pod \"nova-cell1-conductor-db-sync-tn4hn\" (UID: \"6b7cc6f7-1a01-4c53-818d-f40548209a13\") " pod="openstack/nova-cell1-conductor-db-sync-tn4hn" Nov 28 13:40:47 crc kubenswrapper[4631]: I1128 13:40:47.190240 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-tn4hn" Nov 28 13:40:47 crc kubenswrapper[4631]: I1128 13:40:47.501567 4631 generic.go:334] "Generic (PLEG): container finished" podID="e1644d33-bda9-437e-b10c-5c48042cb1f2" containerID="cb0fed8758d1f40a3a884391b7df6bf2761bdff294e9368089d1c7134ae186fe" exitCode=0 Nov 28 13:40:47 crc kubenswrapper[4631]: I1128 13:40:47.502089 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-757b4f8459-j7gzn" event={"ID":"e1644d33-bda9-437e-b10c-5c48042cb1f2","Type":"ContainerDied","Data":"cb0fed8758d1f40a3a884391b7df6bf2761bdff294e9368089d1c7134ae186fe"} Nov 28 13:40:47 crc kubenswrapper[4631]: I1128 13:40:47.502235 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-757b4f8459-j7gzn" event={"ID":"e1644d33-bda9-437e-b10c-5c48042cb1f2","Type":"ContainerStarted","Data":"800facb9428255a99f2c1e48e7b635193d89440e1be749b4236bd11ac22bf4fc"} Nov 28 13:40:47 crc kubenswrapper[4631]: I1128 13:40:47.971234 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-tn4hn"] Nov 28 13:40:48 crc kubenswrapper[4631]: E1128 13:40:48.309541 4631 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod111c5349_528d_4856_a3e4_1277e3b10889.slice/crio-3b826a5556ea3b27f6357d4cc74220f4bce293db6dbcb49fa5131013bef0e517.scope\": RecentStats: unable to find data in memory cache]" Nov 28 13:40:48 crc kubenswrapper[4631]: I1128 13:40:48.544694 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-tn4hn" event={"ID":"6b7cc6f7-1a01-4c53-818d-f40548209a13","Type":"ContainerStarted","Data":"5e760b5a84be2bd8352e4542c8d70a6aafdf38b37d9178d5a56071b495bfeab2"} Nov 28 13:40:48 crc kubenswrapper[4631]: I1128 13:40:48.545170 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-tn4hn" event={"ID":"6b7cc6f7-1a01-4c53-818d-f40548209a13","Type":"ContainerStarted","Data":"e9e2939de3cb8f012cdf07e85c04ccc0369e5628d09163745831e8dc63132c76"} Nov 28 13:40:48 crc kubenswrapper[4631]: I1128 13:40:48.558433 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-757b4f8459-j7gzn" event={"ID":"e1644d33-bda9-437e-b10c-5c48042cb1f2","Type":"ContainerStarted","Data":"33c781584549893b9d761b9f653c86dbdd728267f0f6a18b982048c3cdf700a6"} Nov 28 13:40:48 crc kubenswrapper[4631]: I1128 13:40:48.559568 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-757b4f8459-j7gzn" Nov 28 13:40:48 crc kubenswrapper[4631]: I1128 13:40:48.588543 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-tn4hn" podStartSLOduration=2.5885107080000003 podStartE2EDuration="2.588510708s" podCreationTimestamp="2025-11-28 13:40:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 13:40:48.568445667 +0000 UTC m=+1205.375749011" watchObservedRunningTime="2025-11-28 13:40:48.588510708 +0000 UTC m=+1205.395814062" Nov 28 13:40:48 crc kubenswrapper[4631]: I1128 13:40:48.599466 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-757b4f8459-j7gzn" podStartSLOduration=3.599436999 podStartE2EDuration="3.599436999s" podCreationTimestamp="2025-11-28 13:40:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 13:40:48.599079661 +0000 UTC m=+1205.406383005" watchObservedRunningTime="2025-11-28 13:40:48.599436999 +0000 UTC m=+1205.406740343" Nov 28 13:40:49 crc kubenswrapper[4631]: I1128 13:40:49.287082 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 28 13:40:49 crc kubenswrapper[4631]: I1128 13:40:49.361021 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 28 13:40:49 crc kubenswrapper[4631]: I1128 13:40:49.635015 4631 patch_prober.go:28] interesting pod/machine-config-daemon-47sxw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 28 13:40:49 crc kubenswrapper[4631]: I1128 13:40:49.635094 4631 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" podUID="8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 28 13:40:52 crc kubenswrapper[4631]: I1128 13:40:52.609886 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"e9e6bbdb-eac5-47db-80e5-b438965fa8aa","Type":"ContainerStarted","Data":"1979d0923441a642c786ab0b83653968fbe20f2eaa0e71ce9909faee2d665e9a"} Nov 28 13:40:52 crc kubenswrapper[4631]: I1128 13:40:52.614980 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"c7ac2d63-abe0-4862-a3a2-5d2b1dc87ac9","Type":"ContainerStarted","Data":"a3329f50bfdfc20c2c25e44e2b0cb05371b0dfebb65a54cee815ad41c7dae1f2"} Nov 28 13:40:52 crc kubenswrapper[4631]: I1128 13:40:52.616829 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"cf5855aa-20c9-4b9c-8205-dbd3dcaa8917","Type":"ContainerStarted","Data":"9147cb55c461919e4702bc5b774769da1274e05808b7435122f790ac1a96851e"} Nov 28 13:40:52 crc kubenswrapper[4631]: I1128 13:40:52.618767 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"5e8a60bc-6e73-499a-959a-c0f104014641","Type":"ContainerStarted","Data":"2955d8dedee901ef27b5cf6a4ef41a86cc3d2d1ea69ec7f71f29c857428aaadc"} Nov 28 13:40:52 crc kubenswrapper[4631]: I1128 13:40:52.618897 4631 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="5e8a60bc-6e73-499a-959a-c0f104014641" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://2955d8dedee901ef27b5cf6a4ef41a86cc3d2d1ea69ec7f71f29c857428aaadc" gracePeriod=30 Nov 28 13:40:52 crc kubenswrapper[4631]: I1128 13:40:52.670815 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=3.16359146 podStartE2EDuration="8.67076466s" podCreationTimestamp="2025-11-28 13:40:44 +0000 UTC" firstStartedPulling="2025-11-28 13:40:46.359684226 +0000 UTC m=+1203.166987570" lastFinishedPulling="2025-11-28 13:40:51.866857426 +0000 UTC m=+1208.674160770" observedRunningTime="2025-11-28 13:40:52.634805048 +0000 UTC m=+1209.442108392" watchObservedRunningTime="2025-11-28 13:40:52.67076466 +0000 UTC m=+1209.478068004" Nov 28 13:40:52 crc kubenswrapper[4631]: I1128 13:40:52.676764 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.69683152 podStartE2EDuration="8.676733492s" podCreationTimestamp="2025-11-28 13:40:44 +0000 UTC" firstStartedPulling="2025-11-28 13:40:45.915240001 +0000 UTC m=+1202.722543345" lastFinishedPulling="2025-11-28 13:40:51.895141973 +0000 UTC m=+1208.702445317" observedRunningTime="2025-11-28 13:40:52.659224823 +0000 UTC m=+1209.466528187" watchObservedRunningTime="2025-11-28 13:40:52.676733492 +0000 UTC m=+1209.484036836" Nov 28 13:40:53 crc kubenswrapper[4631]: I1128 13:40:53.639395 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"c7ac2d63-abe0-4862-a3a2-5d2b1dc87ac9","Type":"ContainerStarted","Data":"cc8137bbcd37eea46d3f7488a36a67c1f80733d941fae1fa3f4b4ad9d84ecbd2"} Nov 28 13:40:53 crc kubenswrapper[4631]: I1128 13:40:53.652177 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"cf5855aa-20c9-4b9c-8205-dbd3dcaa8917","Type":"ContainerStarted","Data":"27a67357cdcf2080cf92542d66e81f8e95e1f0e08864cc3abc26d3fe3ac28e7a"} Nov 28 13:40:53 crc kubenswrapper[4631]: I1128 13:40:53.652332 4631 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="cf5855aa-20c9-4b9c-8205-dbd3dcaa8917" containerName="nova-metadata-metadata" containerID="cri-o://27a67357cdcf2080cf92542d66e81f8e95e1f0e08864cc3abc26d3fe3ac28e7a" gracePeriod=30 Nov 28 13:40:53 crc kubenswrapper[4631]: I1128 13:40:53.652461 4631 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="cf5855aa-20c9-4b9c-8205-dbd3dcaa8917" containerName="nova-metadata-log" containerID="cri-o://9147cb55c461919e4702bc5b774769da1274e05808b7435122f790ac1a96851e" gracePeriod=30 Nov 28 13:40:53 crc kubenswrapper[4631]: I1128 13:40:53.688020 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=3.6858163470000003 podStartE2EDuration="9.687985373s" podCreationTimestamp="2025-11-28 13:40:44 +0000 UTC" firstStartedPulling="2025-11-28 13:40:45.896356439 +0000 UTC m=+1202.703659783" lastFinishedPulling="2025-11-28 13:40:51.898525465 +0000 UTC m=+1208.705828809" observedRunningTime="2025-11-28 13:40:53.681921757 +0000 UTC m=+1210.489225101" watchObservedRunningTime="2025-11-28 13:40:53.687985373 +0000 UTC m=+1210.495288717" Nov 28 13:40:53 crc kubenswrapper[4631]: I1128 13:40:53.720319 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=4.174457381 podStartE2EDuration="9.720281316s" podCreationTimestamp="2025-11-28 13:40:44 +0000 UTC" firstStartedPulling="2025-11-28 13:40:46.379088291 +0000 UTC m=+1203.186391635" lastFinishedPulling="2025-11-28 13:40:51.924912226 +0000 UTC m=+1208.732215570" observedRunningTime="2025-11-28 13:40:53.712731805 +0000 UTC m=+1210.520035169" watchObservedRunningTime="2025-11-28 13:40:53.720281316 +0000 UTC m=+1210.527584660" Nov 28 13:40:54 crc kubenswrapper[4631]: I1128 13:40:54.344116 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 28 13:40:54 crc kubenswrapper[4631]: I1128 13:40:54.495412 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cf5855aa-20c9-4b9c-8205-dbd3dcaa8917-combined-ca-bundle\") pod \"cf5855aa-20c9-4b9c-8205-dbd3dcaa8917\" (UID: \"cf5855aa-20c9-4b9c-8205-dbd3dcaa8917\") " Nov 28 13:40:54 crc kubenswrapper[4631]: I1128 13:40:54.495515 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-57cmk\" (UniqueName: \"kubernetes.io/projected/cf5855aa-20c9-4b9c-8205-dbd3dcaa8917-kube-api-access-57cmk\") pod \"cf5855aa-20c9-4b9c-8205-dbd3dcaa8917\" (UID: \"cf5855aa-20c9-4b9c-8205-dbd3dcaa8917\") " Nov 28 13:40:54 crc kubenswrapper[4631]: I1128 13:40:54.495636 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cf5855aa-20c9-4b9c-8205-dbd3dcaa8917-logs\") pod \"cf5855aa-20c9-4b9c-8205-dbd3dcaa8917\" (UID: \"cf5855aa-20c9-4b9c-8205-dbd3dcaa8917\") " Nov 28 13:40:54 crc kubenswrapper[4631]: I1128 13:40:54.495735 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cf5855aa-20c9-4b9c-8205-dbd3dcaa8917-config-data\") pod \"cf5855aa-20c9-4b9c-8205-dbd3dcaa8917\" (UID: \"cf5855aa-20c9-4b9c-8205-dbd3dcaa8917\") " Nov 28 13:40:54 crc kubenswrapper[4631]: I1128 13:40:54.498079 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cf5855aa-20c9-4b9c-8205-dbd3dcaa8917-logs" (OuterVolumeSpecName: "logs") pod "cf5855aa-20c9-4b9c-8205-dbd3dcaa8917" (UID: "cf5855aa-20c9-4b9c-8205-dbd3dcaa8917"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 13:40:54 crc kubenswrapper[4631]: I1128 13:40:54.521855 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cf5855aa-20c9-4b9c-8205-dbd3dcaa8917-kube-api-access-57cmk" (OuterVolumeSpecName: "kube-api-access-57cmk") pod "cf5855aa-20c9-4b9c-8205-dbd3dcaa8917" (UID: "cf5855aa-20c9-4b9c-8205-dbd3dcaa8917"). InnerVolumeSpecName "kube-api-access-57cmk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:40:54 crc kubenswrapper[4631]: I1128 13:40:54.575662 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cf5855aa-20c9-4b9c-8205-dbd3dcaa8917-config-data" (OuterVolumeSpecName: "config-data") pod "cf5855aa-20c9-4b9c-8205-dbd3dcaa8917" (UID: "cf5855aa-20c9-4b9c-8205-dbd3dcaa8917"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:40:54 crc kubenswrapper[4631]: I1128 13:40:54.598083 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-57cmk\" (UniqueName: \"kubernetes.io/projected/cf5855aa-20c9-4b9c-8205-dbd3dcaa8917-kube-api-access-57cmk\") on node \"crc\" DevicePath \"\"" Nov 28 13:40:54 crc kubenswrapper[4631]: I1128 13:40:54.598133 4631 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cf5855aa-20c9-4b9c-8205-dbd3dcaa8917-logs\") on node \"crc\" DevicePath \"\"" Nov 28 13:40:54 crc kubenswrapper[4631]: I1128 13:40:54.598146 4631 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cf5855aa-20c9-4b9c-8205-dbd3dcaa8917-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 13:40:54 crc kubenswrapper[4631]: I1128 13:40:54.598805 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cf5855aa-20c9-4b9c-8205-dbd3dcaa8917-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "cf5855aa-20c9-4b9c-8205-dbd3dcaa8917" (UID: "cf5855aa-20c9-4b9c-8205-dbd3dcaa8917"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:40:54 crc kubenswrapper[4631]: I1128 13:40:54.671328 4631 generic.go:334] "Generic (PLEG): container finished" podID="cf5855aa-20c9-4b9c-8205-dbd3dcaa8917" containerID="27a67357cdcf2080cf92542d66e81f8e95e1f0e08864cc3abc26d3fe3ac28e7a" exitCode=0 Nov 28 13:40:54 crc kubenswrapper[4631]: I1128 13:40:54.671770 4631 generic.go:334] "Generic (PLEG): container finished" podID="cf5855aa-20c9-4b9c-8205-dbd3dcaa8917" containerID="9147cb55c461919e4702bc5b774769da1274e05808b7435122f790ac1a96851e" exitCode=143 Nov 28 13:40:54 crc kubenswrapper[4631]: I1128 13:40:54.671903 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 28 13:40:54 crc kubenswrapper[4631]: I1128 13:40:54.671544 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"cf5855aa-20c9-4b9c-8205-dbd3dcaa8917","Type":"ContainerDied","Data":"27a67357cdcf2080cf92542d66e81f8e95e1f0e08864cc3abc26d3fe3ac28e7a"} Nov 28 13:40:54 crc kubenswrapper[4631]: I1128 13:40:54.672027 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"cf5855aa-20c9-4b9c-8205-dbd3dcaa8917","Type":"ContainerDied","Data":"9147cb55c461919e4702bc5b774769da1274e05808b7435122f790ac1a96851e"} Nov 28 13:40:54 crc kubenswrapper[4631]: I1128 13:40:54.672041 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"cf5855aa-20c9-4b9c-8205-dbd3dcaa8917","Type":"ContainerDied","Data":"08593af7191fdacda9ad743065d8b3b03c47681de484335e1159135eb68e6797"} Nov 28 13:40:54 crc kubenswrapper[4631]: I1128 13:40:54.672059 4631 scope.go:117] "RemoveContainer" containerID="27a67357cdcf2080cf92542d66e81f8e95e1f0e08864cc3abc26d3fe3ac28e7a" Nov 28 13:40:54 crc kubenswrapper[4631]: I1128 13:40:54.702900 4631 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cf5855aa-20c9-4b9c-8205-dbd3dcaa8917-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 13:40:54 crc kubenswrapper[4631]: I1128 13:40:54.751373 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 28 13:40:54 crc kubenswrapper[4631]: I1128 13:40:54.753914 4631 scope.go:117] "RemoveContainer" containerID="9147cb55c461919e4702bc5b774769da1274e05808b7435122f790ac1a96851e" Nov 28 13:40:54 crc kubenswrapper[4631]: I1128 13:40:54.760451 4631 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Nov 28 13:40:54 crc kubenswrapper[4631]: I1128 13:40:54.800882 4631 scope.go:117] "RemoveContainer" containerID="27a67357cdcf2080cf92542d66e81f8e95e1f0e08864cc3abc26d3fe3ac28e7a" Nov 28 13:40:54 crc kubenswrapper[4631]: E1128 13:40:54.801650 4631 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"27a67357cdcf2080cf92542d66e81f8e95e1f0e08864cc3abc26d3fe3ac28e7a\": container with ID starting with 27a67357cdcf2080cf92542d66e81f8e95e1f0e08864cc3abc26d3fe3ac28e7a not found: ID does not exist" containerID="27a67357cdcf2080cf92542d66e81f8e95e1f0e08864cc3abc26d3fe3ac28e7a" Nov 28 13:40:54 crc kubenswrapper[4631]: I1128 13:40:54.801718 4631 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"27a67357cdcf2080cf92542d66e81f8e95e1f0e08864cc3abc26d3fe3ac28e7a"} err="failed to get container status \"27a67357cdcf2080cf92542d66e81f8e95e1f0e08864cc3abc26d3fe3ac28e7a\": rpc error: code = NotFound desc = could not find container \"27a67357cdcf2080cf92542d66e81f8e95e1f0e08864cc3abc26d3fe3ac28e7a\": container with ID starting with 27a67357cdcf2080cf92542d66e81f8e95e1f0e08864cc3abc26d3fe3ac28e7a not found: ID does not exist" Nov 28 13:40:54 crc kubenswrapper[4631]: I1128 13:40:54.801756 4631 scope.go:117] "RemoveContainer" containerID="9147cb55c461919e4702bc5b774769da1274e05808b7435122f790ac1a96851e" Nov 28 13:40:54 crc kubenswrapper[4631]: E1128 13:40:54.802105 4631 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9147cb55c461919e4702bc5b774769da1274e05808b7435122f790ac1a96851e\": container with ID starting with 9147cb55c461919e4702bc5b774769da1274e05808b7435122f790ac1a96851e not found: ID does not exist" containerID="9147cb55c461919e4702bc5b774769da1274e05808b7435122f790ac1a96851e" Nov 28 13:40:54 crc kubenswrapper[4631]: I1128 13:40:54.802147 4631 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9147cb55c461919e4702bc5b774769da1274e05808b7435122f790ac1a96851e"} err="failed to get container status \"9147cb55c461919e4702bc5b774769da1274e05808b7435122f790ac1a96851e\": rpc error: code = NotFound desc = could not find container \"9147cb55c461919e4702bc5b774769da1274e05808b7435122f790ac1a96851e\": container with ID starting with 9147cb55c461919e4702bc5b774769da1274e05808b7435122f790ac1a96851e not found: ID does not exist" Nov 28 13:40:54 crc kubenswrapper[4631]: I1128 13:40:54.802163 4631 scope.go:117] "RemoveContainer" containerID="27a67357cdcf2080cf92542d66e81f8e95e1f0e08864cc3abc26d3fe3ac28e7a" Nov 28 13:40:54 crc kubenswrapper[4631]: I1128 13:40:54.802489 4631 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"27a67357cdcf2080cf92542d66e81f8e95e1f0e08864cc3abc26d3fe3ac28e7a"} err="failed to get container status \"27a67357cdcf2080cf92542d66e81f8e95e1f0e08864cc3abc26d3fe3ac28e7a\": rpc error: code = NotFound desc = could not find container \"27a67357cdcf2080cf92542d66e81f8e95e1f0e08864cc3abc26d3fe3ac28e7a\": container with ID starting with 27a67357cdcf2080cf92542d66e81f8e95e1f0e08864cc3abc26d3fe3ac28e7a not found: ID does not exist" Nov 28 13:40:54 crc kubenswrapper[4631]: I1128 13:40:54.802518 4631 scope.go:117] "RemoveContainer" containerID="9147cb55c461919e4702bc5b774769da1274e05808b7435122f790ac1a96851e" Nov 28 13:40:54 crc kubenswrapper[4631]: I1128 13:40:54.802756 4631 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9147cb55c461919e4702bc5b774769da1274e05808b7435122f790ac1a96851e"} err="failed to get container status \"9147cb55c461919e4702bc5b774769da1274e05808b7435122f790ac1a96851e\": rpc error: code = NotFound desc = could not find container \"9147cb55c461919e4702bc5b774769da1274e05808b7435122f790ac1a96851e\": container with ID starting with 9147cb55c461919e4702bc5b774769da1274e05808b7435122f790ac1a96851e not found: ID does not exist" Nov 28 13:40:54 crc kubenswrapper[4631]: I1128 13:40:54.805518 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Nov 28 13:40:54 crc kubenswrapper[4631]: E1128 13:40:54.806481 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cf5855aa-20c9-4b9c-8205-dbd3dcaa8917" containerName="nova-metadata-log" Nov 28 13:40:54 crc kubenswrapper[4631]: I1128 13:40:54.806502 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="cf5855aa-20c9-4b9c-8205-dbd3dcaa8917" containerName="nova-metadata-log" Nov 28 13:40:54 crc kubenswrapper[4631]: E1128 13:40:54.806559 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cf5855aa-20c9-4b9c-8205-dbd3dcaa8917" containerName="nova-metadata-metadata" Nov 28 13:40:54 crc kubenswrapper[4631]: I1128 13:40:54.806567 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="cf5855aa-20c9-4b9c-8205-dbd3dcaa8917" containerName="nova-metadata-metadata" Nov 28 13:40:54 crc kubenswrapper[4631]: I1128 13:40:54.807014 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="cf5855aa-20c9-4b9c-8205-dbd3dcaa8917" containerName="nova-metadata-log" Nov 28 13:40:54 crc kubenswrapper[4631]: I1128 13:40:54.807034 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="cf5855aa-20c9-4b9c-8205-dbd3dcaa8917" containerName="nova-metadata-metadata" Nov 28 13:40:54 crc kubenswrapper[4631]: I1128 13:40:54.817256 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 28 13:40:54 crc kubenswrapper[4631]: I1128 13:40:54.831583 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Nov 28 13:40:54 crc kubenswrapper[4631]: I1128 13:40:54.836412 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Nov 28 13:40:54 crc kubenswrapper[4631]: I1128 13:40:54.863549 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 28 13:40:54 crc kubenswrapper[4631]: I1128 13:40:54.885671 4631 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 28 13:40:54 crc kubenswrapper[4631]: I1128 13:40:54.886074 4631 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 28 13:40:54 crc kubenswrapper[4631]: I1128 13:40:54.897430 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Nov 28 13:40:54 crc kubenswrapper[4631]: I1128 13:40:54.933310 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f487af97-9737-48cf-a0ab-a442f143929b-logs\") pod \"nova-metadata-0\" (UID: \"f487af97-9737-48cf-a0ab-a442f143929b\") " pod="openstack/nova-metadata-0" Nov 28 13:40:54 crc kubenswrapper[4631]: I1128 13:40:54.933378 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/f487af97-9737-48cf-a0ab-a442f143929b-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"f487af97-9737-48cf-a0ab-a442f143929b\") " pod="openstack/nova-metadata-0" Nov 28 13:40:54 crc kubenswrapper[4631]: I1128 13:40:54.933435 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f487af97-9737-48cf-a0ab-a442f143929b-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"f487af97-9737-48cf-a0ab-a442f143929b\") " pod="openstack/nova-metadata-0" Nov 28 13:40:54 crc kubenswrapper[4631]: I1128 13:40:54.933456 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f487af97-9737-48cf-a0ab-a442f143929b-config-data\") pod \"nova-metadata-0\" (UID: \"f487af97-9737-48cf-a0ab-a442f143929b\") " pod="openstack/nova-metadata-0" Nov 28 13:40:54 crc kubenswrapper[4631]: I1128 13:40:54.933479 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ms9vf\" (UniqueName: \"kubernetes.io/projected/f487af97-9737-48cf-a0ab-a442f143929b-kube-api-access-ms9vf\") pod \"nova-metadata-0\" (UID: \"f487af97-9737-48cf-a0ab-a442f143929b\") " pod="openstack/nova-metadata-0" Nov 28 13:40:55 crc kubenswrapper[4631]: I1128 13:40:55.035848 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f487af97-9737-48cf-a0ab-a442f143929b-logs\") pod \"nova-metadata-0\" (UID: \"f487af97-9737-48cf-a0ab-a442f143929b\") " pod="openstack/nova-metadata-0" Nov 28 13:40:55 crc kubenswrapper[4631]: I1128 13:40:55.035932 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/f487af97-9737-48cf-a0ab-a442f143929b-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"f487af97-9737-48cf-a0ab-a442f143929b\") " pod="openstack/nova-metadata-0" Nov 28 13:40:55 crc kubenswrapper[4631]: I1128 13:40:55.035999 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f487af97-9737-48cf-a0ab-a442f143929b-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"f487af97-9737-48cf-a0ab-a442f143929b\") " pod="openstack/nova-metadata-0" Nov 28 13:40:55 crc kubenswrapper[4631]: I1128 13:40:55.036017 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f487af97-9737-48cf-a0ab-a442f143929b-config-data\") pod \"nova-metadata-0\" (UID: \"f487af97-9737-48cf-a0ab-a442f143929b\") " pod="openstack/nova-metadata-0" Nov 28 13:40:55 crc kubenswrapper[4631]: I1128 13:40:55.036042 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ms9vf\" (UniqueName: \"kubernetes.io/projected/f487af97-9737-48cf-a0ab-a442f143929b-kube-api-access-ms9vf\") pod \"nova-metadata-0\" (UID: \"f487af97-9737-48cf-a0ab-a442f143929b\") " pod="openstack/nova-metadata-0" Nov 28 13:40:55 crc kubenswrapper[4631]: I1128 13:40:55.036326 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f487af97-9737-48cf-a0ab-a442f143929b-logs\") pod \"nova-metadata-0\" (UID: \"f487af97-9737-48cf-a0ab-a442f143929b\") " pod="openstack/nova-metadata-0" Nov 28 13:40:55 crc kubenswrapper[4631]: I1128 13:40:55.045208 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/f487af97-9737-48cf-a0ab-a442f143929b-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"f487af97-9737-48cf-a0ab-a442f143929b\") " pod="openstack/nova-metadata-0" Nov 28 13:40:55 crc kubenswrapper[4631]: I1128 13:40:55.047046 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f487af97-9737-48cf-a0ab-a442f143929b-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"f487af97-9737-48cf-a0ab-a442f143929b\") " pod="openstack/nova-metadata-0" Nov 28 13:40:55 crc kubenswrapper[4631]: I1128 13:40:55.047845 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f487af97-9737-48cf-a0ab-a442f143929b-config-data\") pod \"nova-metadata-0\" (UID: \"f487af97-9737-48cf-a0ab-a442f143929b\") " pod="openstack/nova-metadata-0" Nov 28 13:40:55 crc kubenswrapper[4631]: I1128 13:40:55.072110 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ms9vf\" (UniqueName: \"kubernetes.io/projected/f487af97-9737-48cf-a0ab-a442f143929b-kube-api-access-ms9vf\") pod \"nova-metadata-0\" (UID: \"f487af97-9737-48cf-a0ab-a442f143929b\") " pod="openstack/nova-metadata-0" Nov 28 13:40:55 crc kubenswrapper[4631]: I1128 13:40:55.162347 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 28 13:40:55 crc kubenswrapper[4631]: I1128 13:40:55.246956 4631 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Nov 28 13:40:55 crc kubenswrapper[4631]: I1128 13:40:55.247021 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Nov 28 13:40:55 crc kubenswrapper[4631]: I1128 13:40:55.318438 4631 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Nov 28 13:40:55 crc kubenswrapper[4631]: I1128 13:40:55.331719 4631 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-64f8fdf97b-jn5mg" podUID="111c5349-528d-4856-a3e4-1277e3b10889" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.146:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.146:8443: connect: connection refused" Nov 28 13:40:55 crc kubenswrapper[4631]: I1128 13:40:55.434478 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-757b4f8459-j7gzn" Nov 28 13:40:55 crc kubenswrapper[4631]: I1128 13:40:55.539313 4631 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cf5855aa-20c9-4b9c-8205-dbd3dcaa8917" path="/var/lib/kubelet/pods/cf5855aa-20c9-4b9c-8205-dbd3dcaa8917/volumes" Nov 28 13:40:55 crc kubenswrapper[4631]: I1128 13:40:55.635271 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-9ctzb"] Nov 28 13:40:55 crc kubenswrapper[4631]: I1128 13:40:55.635619 4631 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5c9776ccc5-9ctzb" podUID="dc4f7c59-e6c6-4659-8d2d-8e584afbe2ab" containerName="dnsmasq-dns" containerID="cri-o://c7c4bb142691d554c4385d0a03fd36470b904945d17f19500072f4ec681c4907" gracePeriod=10 Nov 28 13:40:55 crc kubenswrapper[4631]: I1128 13:40:55.733678 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Nov 28 13:40:55 crc kubenswrapper[4631]: I1128 13:40:55.955002 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 28 13:40:55 crc kubenswrapper[4631]: I1128 13:40:55.972803 4631 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="c7ac2d63-abe0-4862-a3a2-5d2b1dc87ac9" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.186:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 28 13:40:55 crc kubenswrapper[4631]: I1128 13:40:55.973496 4631 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="c7ac2d63-abe0-4862-a3a2-5d2b1dc87ac9" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.186:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 28 13:40:56 crc kubenswrapper[4631]: W1128 13:40:56.008132 4631 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf487af97_9737_48cf_a0ab_a442f143929b.slice/crio-97e91ebed5dc0241ab9ba4c5f47ca6bc4f76583f75a4bb410f4d257b5efba497 WatchSource:0}: Error finding container 97e91ebed5dc0241ab9ba4c5f47ca6bc4f76583f75a4bb410f4d257b5efba497: Status 404 returned error can't find the container with id 97e91ebed5dc0241ab9ba4c5f47ca6bc4f76583f75a4bb410f4d257b5efba497 Nov 28 13:40:56 crc kubenswrapper[4631]: I1128 13:40:56.529503 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9776ccc5-9ctzb" Nov 28 13:40:56 crc kubenswrapper[4631]: I1128 13:40:56.621517 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/dc4f7c59-e6c6-4659-8d2d-8e584afbe2ab-ovsdbserver-nb\") pod \"dc4f7c59-e6c6-4659-8d2d-8e584afbe2ab\" (UID: \"dc4f7c59-e6c6-4659-8d2d-8e584afbe2ab\") " Nov 28 13:40:56 crc kubenswrapper[4631]: I1128 13:40:56.621625 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/dc4f7c59-e6c6-4659-8d2d-8e584afbe2ab-dns-swift-storage-0\") pod \"dc4f7c59-e6c6-4659-8d2d-8e584afbe2ab\" (UID: \"dc4f7c59-e6c6-4659-8d2d-8e584afbe2ab\") " Nov 28 13:40:56 crc kubenswrapper[4631]: I1128 13:40:56.622169 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/dc4f7c59-e6c6-4659-8d2d-8e584afbe2ab-ovsdbserver-sb\") pod \"dc4f7c59-e6c6-4659-8d2d-8e584afbe2ab\" (UID: \"dc4f7c59-e6c6-4659-8d2d-8e584afbe2ab\") " Nov 28 13:40:56 crc kubenswrapper[4631]: I1128 13:40:56.622245 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dc4f7c59-e6c6-4659-8d2d-8e584afbe2ab-config\") pod \"dc4f7c59-e6c6-4659-8d2d-8e584afbe2ab\" (UID: \"dc4f7c59-e6c6-4659-8d2d-8e584afbe2ab\") " Nov 28 13:40:56 crc kubenswrapper[4631]: I1128 13:40:56.622382 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nc9pv\" (UniqueName: \"kubernetes.io/projected/dc4f7c59-e6c6-4659-8d2d-8e584afbe2ab-kube-api-access-nc9pv\") pod \"dc4f7c59-e6c6-4659-8d2d-8e584afbe2ab\" (UID: \"dc4f7c59-e6c6-4659-8d2d-8e584afbe2ab\") " Nov 28 13:40:56 crc kubenswrapper[4631]: I1128 13:40:56.622425 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/dc4f7c59-e6c6-4659-8d2d-8e584afbe2ab-dns-svc\") pod \"dc4f7c59-e6c6-4659-8d2d-8e584afbe2ab\" (UID: \"dc4f7c59-e6c6-4659-8d2d-8e584afbe2ab\") " Nov 28 13:40:56 crc kubenswrapper[4631]: I1128 13:40:56.721168 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f487af97-9737-48cf-a0ab-a442f143929b","Type":"ContainerStarted","Data":"9d45d82ad9a1cab47bbcabf7ac8ec4a6f275787e84b79703b999c96b65a918c2"} Nov 28 13:40:56 crc kubenswrapper[4631]: I1128 13:40:56.722038 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f487af97-9737-48cf-a0ab-a442f143929b","Type":"ContainerStarted","Data":"97e91ebed5dc0241ab9ba4c5f47ca6bc4f76583f75a4bb410f4d257b5efba497"} Nov 28 13:40:56 crc kubenswrapper[4631]: I1128 13:40:56.724048 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dc4f7c59-e6c6-4659-8d2d-8e584afbe2ab-kube-api-access-nc9pv" (OuterVolumeSpecName: "kube-api-access-nc9pv") pod "dc4f7c59-e6c6-4659-8d2d-8e584afbe2ab" (UID: "dc4f7c59-e6c6-4659-8d2d-8e584afbe2ab"). InnerVolumeSpecName "kube-api-access-nc9pv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:40:56 crc kubenswrapper[4631]: I1128 13:40:56.728345 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nc9pv\" (UniqueName: \"kubernetes.io/projected/dc4f7c59-e6c6-4659-8d2d-8e584afbe2ab-kube-api-access-nc9pv\") on node \"crc\" DevicePath \"\"" Nov 28 13:40:56 crc kubenswrapper[4631]: I1128 13:40:56.768052 4631 generic.go:334] "Generic (PLEG): container finished" podID="dc4f7c59-e6c6-4659-8d2d-8e584afbe2ab" containerID="c7c4bb142691d554c4385d0a03fd36470b904945d17f19500072f4ec681c4907" exitCode=0 Nov 28 13:40:56 crc kubenswrapper[4631]: I1128 13:40:56.768527 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-9ctzb" event={"ID":"dc4f7c59-e6c6-4659-8d2d-8e584afbe2ab","Type":"ContainerDied","Data":"c7c4bb142691d554c4385d0a03fd36470b904945d17f19500072f4ec681c4907"} Nov 28 13:40:56 crc kubenswrapper[4631]: I1128 13:40:56.768599 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9776ccc5-9ctzb" Nov 28 13:40:56 crc kubenswrapper[4631]: I1128 13:40:56.768625 4631 scope.go:117] "RemoveContainer" containerID="c7c4bb142691d554c4385d0a03fd36470b904945d17f19500072f4ec681c4907" Nov 28 13:40:56 crc kubenswrapper[4631]: I1128 13:40:56.768604 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-9ctzb" event={"ID":"dc4f7c59-e6c6-4659-8d2d-8e584afbe2ab","Type":"ContainerDied","Data":"98ed656c4e7b1e35f2223f91dbc22e1e584d50bbde7da6b5daf1caed1d8eb034"} Nov 28 13:40:56 crc kubenswrapper[4631]: I1128 13:40:56.810469 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dc4f7c59-e6c6-4659-8d2d-8e584afbe2ab-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "dc4f7c59-e6c6-4659-8d2d-8e584afbe2ab" (UID: "dc4f7c59-e6c6-4659-8d2d-8e584afbe2ab"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 13:40:56 crc kubenswrapper[4631]: I1128 13:40:56.830645 4631 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/dc4f7c59-e6c6-4659-8d2d-8e584afbe2ab-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 28 13:40:56 crc kubenswrapper[4631]: I1128 13:40:56.850079 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dc4f7c59-e6c6-4659-8d2d-8e584afbe2ab-config" (OuterVolumeSpecName: "config") pod "dc4f7c59-e6c6-4659-8d2d-8e584afbe2ab" (UID: "dc4f7c59-e6c6-4659-8d2d-8e584afbe2ab"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 13:40:56 crc kubenswrapper[4631]: I1128 13:40:56.890065 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dc4f7c59-e6c6-4659-8d2d-8e584afbe2ab-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "dc4f7c59-e6c6-4659-8d2d-8e584afbe2ab" (UID: "dc4f7c59-e6c6-4659-8d2d-8e584afbe2ab"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 13:40:56 crc kubenswrapper[4631]: I1128 13:40:56.932706 4631 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/dc4f7c59-e6c6-4659-8d2d-8e584afbe2ab-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 28 13:40:56 crc kubenswrapper[4631]: I1128 13:40:56.932750 4631 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dc4f7c59-e6c6-4659-8d2d-8e584afbe2ab-config\") on node \"crc\" DevicePath \"\"" Nov 28 13:40:56 crc kubenswrapper[4631]: I1128 13:40:56.933438 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dc4f7c59-e6c6-4659-8d2d-8e584afbe2ab-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "dc4f7c59-e6c6-4659-8d2d-8e584afbe2ab" (UID: "dc4f7c59-e6c6-4659-8d2d-8e584afbe2ab"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 13:40:56 crc kubenswrapper[4631]: I1128 13:40:56.942541 4631 scope.go:117] "RemoveContainer" containerID="8162b6a5a69f910fbfd625bb8c521567d39c2ebd96fe6a3aec4958d8543d33bf" Nov 28 13:40:56 crc kubenswrapper[4631]: I1128 13:40:56.954697 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dc4f7c59-e6c6-4659-8d2d-8e584afbe2ab-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "dc4f7c59-e6c6-4659-8d2d-8e584afbe2ab" (UID: "dc4f7c59-e6c6-4659-8d2d-8e584afbe2ab"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 13:40:57 crc kubenswrapper[4631]: I1128 13:40:57.017398 4631 scope.go:117] "RemoveContainer" containerID="c7c4bb142691d554c4385d0a03fd36470b904945d17f19500072f4ec681c4907" Nov 28 13:40:57 crc kubenswrapper[4631]: E1128 13:40:57.025310 4631 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c7c4bb142691d554c4385d0a03fd36470b904945d17f19500072f4ec681c4907\": container with ID starting with c7c4bb142691d554c4385d0a03fd36470b904945d17f19500072f4ec681c4907 not found: ID does not exist" containerID="c7c4bb142691d554c4385d0a03fd36470b904945d17f19500072f4ec681c4907" Nov 28 13:40:57 crc kubenswrapper[4631]: I1128 13:40:57.025370 4631 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c7c4bb142691d554c4385d0a03fd36470b904945d17f19500072f4ec681c4907"} err="failed to get container status \"c7c4bb142691d554c4385d0a03fd36470b904945d17f19500072f4ec681c4907\": rpc error: code = NotFound desc = could not find container \"c7c4bb142691d554c4385d0a03fd36470b904945d17f19500072f4ec681c4907\": container with ID starting with c7c4bb142691d554c4385d0a03fd36470b904945d17f19500072f4ec681c4907 not found: ID does not exist" Nov 28 13:40:57 crc kubenswrapper[4631]: I1128 13:40:57.025433 4631 scope.go:117] "RemoveContainer" containerID="8162b6a5a69f910fbfd625bb8c521567d39c2ebd96fe6a3aec4958d8543d33bf" Nov 28 13:40:57 crc kubenswrapper[4631]: E1128 13:40:57.028193 4631 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8162b6a5a69f910fbfd625bb8c521567d39c2ebd96fe6a3aec4958d8543d33bf\": container with ID starting with 8162b6a5a69f910fbfd625bb8c521567d39c2ebd96fe6a3aec4958d8543d33bf not found: ID does not exist" containerID="8162b6a5a69f910fbfd625bb8c521567d39c2ebd96fe6a3aec4958d8543d33bf" Nov 28 13:40:57 crc kubenswrapper[4631]: I1128 13:40:57.028250 4631 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8162b6a5a69f910fbfd625bb8c521567d39c2ebd96fe6a3aec4958d8543d33bf"} err="failed to get container status \"8162b6a5a69f910fbfd625bb8c521567d39c2ebd96fe6a3aec4958d8543d33bf\": rpc error: code = NotFound desc = could not find container \"8162b6a5a69f910fbfd625bb8c521567d39c2ebd96fe6a3aec4958d8543d33bf\": container with ID starting with 8162b6a5a69f910fbfd625bb8c521567d39c2ebd96fe6a3aec4958d8543d33bf not found: ID does not exist" Nov 28 13:40:57 crc kubenswrapper[4631]: I1128 13:40:57.034395 4631 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/dc4f7c59-e6c6-4659-8d2d-8e584afbe2ab-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 28 13:40:57 crc kubenswrapper[4631]: I1128 13:40:57.034435 4631 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/dc4f7c59-e6c6-4659-8d2d-8e584afbe2ab-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 28 13:40:57 crc kubenswrapper[4631]: I1128 13:40:57.113824 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-9ctzb"] Nov 28 13:40:57 crc kubenswrapper[4631]: I1128 13:40:57.124613 4631 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-9ctzb"] Nov 28 13:40:57 crc kubenswrapper[4631]: I1128 13:40:57.543160 4631 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dc4f7c59-e6c6-4659-8d2d-8e584afbe2ab" path="/var/lib/kubelet/pods/dc4f7c59-e6c6-4659-8d2d-8e584afbe2ab/volumes" Nov 28 13:40:57 crc kubenswrapper[4631]: I1128 13:40:57.788829 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Nov 28 13:40:57 crc kubenswrapper[4631]: I1128 13:40:57.803535 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f487af97-9737-48cf-a0ab-a442f143929b","Type":"ContainerStarted","Data":"0176fdd3cef121c949afaf075e55b66120e88ef916c4a388add9afe002dcac8d"} Nov 28 13:40:57 crc kubenswrapper[4631]: I1128 13:40:57.904961 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=3.90492723 podStartE2EDuration="3.90492723s" podCreationTimestamp="2025-11-28 13:40:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 13:40:57.891413827 +0000 UTC m=+1214.698717171" watchObservedRunningTime="2025-11-28 13:40:57.90492723 +0000 UTC m=+1214.712230574" Nov 28 13:40:58 crc kubenswrapper[4631]: E1128 13:40:58.721581 4631 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod111c5349_528d_4856_a3e4_1277e3b10889.slice/crio-3b826a5556ea3b27f6357d4cc74220f4bce293db6dbcb49fa5131013bef0e517.scope\": RecentStats: unable to find data in memory cache]" Nov 28 13:41:00 crc kubenswrapper[4631]: I1128 13:41:00.163063 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 28 13:41:00 crc kubenswrapper[4631]: I1128 13:41:00.165220 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 28 13:41:00 crc kubenswrapper[4631]: I1128 13:41:00.842626 4631 generic.go:334] "Generic (PLEG): container finished" podID="476e6d13-87a1-4155-ae6b-f6453e9c2044" containerID="44bdbae8c83b2de9874d7d4a4580cd7185b4f75b99e0304faaeb9f093af2f0d2" exitCode=0 Nov 28 13:41:00 crc kubenswrapper[4631]: I1128 13:41:00.842700 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-cs6bm" event={"ID":"476e6d13-87a1-4155-ae6b-f6453e9c2044","Type":"ContainerDied","Data":"44bdbae8c83b2de9874d7d4a4580cd7185b4f75b99e0304faaeb9f093af2f0d2"} Nov 28 13:41:00 crc kubenswrapper[4631]: I1128 13:41:00.844764 4631 generic.go:334] "Generic (PLEG): container finished" podID="6b7cc6f7-1a01-4c53-818d-f40548209a13" containerID="5e760b5a84be2bd8352e4542c8d70a6aafdf38b37d9178d5a56071b495bfeab2" exitCode=0 Nov 28 13:41:00 crc kubenswrapper[4631]: I1128 13:41:00.844841 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-tn4hn" event={"ID":"6b7cc6f7-1a01-4c53-818d-f40548209a13","Type":"ContainerDied","Data":"5e760b5a84be2bd8352e4542c8d70a6aafdf38b37d9178d5a56071b495bfeab2"} Nov 28 13:41:02 crc kubenswrapper[4631]: I1128 13:41:02.604650 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-tn4hn" Nov 28 13:41:02 crc kubenswrapper[4631]: I1128 13:41:02.612427 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-cs6bm" Nov 28 13:41:02 crc kubenswrapper[4631]: I1128 13:41:02.724332 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6t89z\" (UniqueName: \"kubernetes.io/projected/6b7cc6f7-1a01-4c53-818d-f40548209a13-kube-api-access-6t89z\") pod \"6b7cc6f7-1a01-4c53-818d-f40548209a13\" (UID: \"6b7cc6f7-1a01-4c53-818d-f40548209a13\") " Nov 28 13:41:02 crc kubenswrapper[4631]: I1128 13:41:02.724459 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/476e6d13-87a1-4155-ae6b-f6453e9c2044-combined-ca-bundle\") pod \"476e6d13-87a1-4155-ae6b-f6453e9c2044\" (UID: \"476e6d13-87a1-4155-ae6b-f6453e9c2044\") " Nov 28 13:41:02 crc kubenswrapper[4631]: I1128 13:41:02.724517 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6b7cc6f7-1a01-4c53-818d-f40548209a13-config-data\") pod \"6b7cc6f7-1a01-4c53-818d-f40548209a13\" (UID: \"6b7cc6f7-1a01-4c53-818d-f40548209a13\") " Nov 28 13:41:02 crc kubenswrapper[4631]: I1128 13:41:02.724541 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b7cc6f7-1a01-4c53-818d-f40548209a13-combined-ca-bundle\") pod \"6b7cc6f7-1a01-4c53-818d-f40548209a13\" (UID: \"6b7cc6f7-1a01-4c53-818d-f40548209a13\") " Nov 28 13:41:02 crc kubenswrapper[4631]: I1128 13:41:02.724647 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s7fj8\" (UniqueName: \"kubernetes.io/projected/476e6d13-87a1-4155-ae6b-f6453e9c2044-kube-api-access-s7fj8\") pod \"476e6d13-87a1-4155-ae6b-f6453e9c2044\" (UID: \"476e6d13-87a1-4155-ae6b-f6453e9c2044\") " Nov 28 13:41:02 crc kubenswrapper[4631]: I1128 13:41:02.724792 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/476e6d13-87a1-4155-ae6b-f6453e9c2044-scripts\") pod \"476e6d13-87a1-4155-ae6b-f6453e9c2044\" (UID: \"476e6d13-87a1-4155-ae6b-f6453e9c2044\") " Nov 28 13:41:02 crc kubenswrapper[4631]: I1128 13:41:02.725248 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/476e6d13-87a1-4155-ae6b-f6453e9c2044-config-data\") pod \"476e6d13-87a1-4155-ae6b-f6453e9c2044\" (UID: \"476e6d13-87a1-4155-ae6b-f6453e9c2044\") " Nov 28 13:41:02 crc kubenswrapper[4631]: I1128 13:41:02.725335 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6b7cc6f7-1a01-4c53-818d-f40548209a13-scripts\") pod \"6b7cc6f7-1a01-4c53-818d-f40548209a13\" (UID: \"6b7cc6f7-1a01-4c53-818d-f40548209a13\") " Nov 28 13:41:02 crc kubenswrapper[4631]: I1128 13:41:02.732153 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6b7cc6f7-1a01-4c53-818d-f40548209a13-scripts" (OuterVolumeSpecName: "scripts") pod "6b7cc6f7-1a01-4c53-818d-f40548209a13" (UID: "6b7cc6f7-1a01-4c53-818d-f40548209a13"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:41:02 crc kubenswrapper[4631]: I1128 13:41:02.752145 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/476e6d13-87a1-4155-ae6b-f6453e9c2044-scripts" (OuterVolumeSpecName: "scripts") pod "476e6d13-87a1-4155-ae6b-f6453e9c2044" (UID: "476e6d13-87a1-4155-ae6b-f6453e9c2044"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:41:02 crc kubenswrapper[4631]: I1128 13:41:02.758019 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/476e6d13-87a1-4155-ae6b-f6453e9c2044-kube-api-access-s7fj8" (OuterVolumeSpecName: "kube-api-access-s7fj8") pod "476e6d13-87a1-4155-ae6b-f6453e9c2044" (UID: "476e6d13-87a1-4155-ae6b-f6453e9c2044"). InnerVolumeSpecName "kube-api-access-s7fj8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:41:02 crc kubenswrapper[4631]: I1128 13:41:02.758641 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6b7cc6f7-1a01-4c53-818d-f40548209a13-kube-api-access-6t89z" (OuterVolumeSpecName: "kube-api-access-6t89z") pod "6b7cc6f7-1a01-4c53-818d-f40548209a13" (UID: "6b7cc6f7-1a01-4c53-818d-f40548209a13"). InnerVolumeSpecName "kube-api-access-6t89z". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:41:02 crc kubenswrapper[4631]: I1128 13:41:02.767022 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6b7cc6f7-1a01-4c53-818d-f40548209a13-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6b7cc6f7-1a01-4c53-818d-f40548209a13" (UID: "6b7cc6f7-1a01-4c53-818d-f40548209a13"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:41:02 crc kubenswrapper[4631]: I1128 13:41:02.779412 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/476e6d13-87a1-4155-ae6b-f6453e9c2044-config-data" (OuterVolumeSpecName: "config-data") pod "476e6d13-87a1-4155-ae6b-f6453e9c2044" (UID: "476e6d13-87a1-4155-ae6b-f6453e9c2044"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:41:02 crc kubenswrapper[4631]: I1128 13:41:02.780505 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6b7cc6f7-1a01-4c53-818d-f40548209a13-config-data" (OuterVolumeSpecName: "config-data") pod "6b7cc6f7-1a01-4c53-818d-f40548209a13" (UID: "6b7cc6f7-1a01-4c53-818d-f40548209a13"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:41:02 crc kubenswrapper[4631]: I1128 13:41:02.783523 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/476e6d13-87a1-4155-ae6b-f6453e9c2044-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "476e6d13-87a1-4155-ae6b-f6453e9c2044" (UID: "476e6d13-87a1-4155-ae6b-f6453e9c2044"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:41:02 crc kubenswrapper[4631]: I1128 13:41:02.829000 4631 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/476e6d13-87a1-4155-ae6b-f6453e9c2044-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 13:41:02 crc kubenswrapper[4631]: I1128 13:41:02.829038 4631 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6b7cc6f7-1a01-4c53-818d-f40548209a13-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 13:41:02 crc kubenswrapper[4631]: I1128 13:41:02.829048 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6t89z\" (UniqueName: \"kubernetes.io/projected/6b7cc6f7-1a01-4c53-818d-f40548209a13-kube-api-access-6t89z\") on node \"crc\" DevicePath \"\"" Nov 28 13:41:02 crc kubenswrapper[4631]: I1128 13:41:02.829060 4631 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/476e6d13-87a1-4155-ae6b-f6453e9c2044-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 13:41:02 crc kubenswrapper[4631]: I1128 13:41:02.829069 4631 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6b7cc6f7-1a01-4c53-818d-f40548209a13-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 13:41:02 crc kubenswrapper[4631]: I1128 13:41:02.829077 4631 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b7cc6f7-1a01-4c53-818d-f40548209a13-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 13:41:02 crc kubenswrapper[4631]: I1128 13:41:02.829084 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s7fj8\" (UniqueName: \"kubernetes.io/projected/476e6d13-87a1-4155-ae6b-f6453e9c2044-kube-api-access-s7fj8\") on node \"crc\" DevicePath \"\"" Nov 28 13:41:02 crc kubenswrapper[4631]: I1128 13:41:02.829097 4631 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/476e6d13-87a1-4155-ae6b-f6453e9c2044-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 13:41:02 crc kubenswrapper[4631]: I1128 13:41:02.866419 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-tn4hn" event={"ID":"6b7cc6f7-1a01-4c53-818d-f40548209a13","Type":"ContainerDied","Data":"e9e2939de3cb8f012cdf07e85c04ccc0369e5628d09163745831e8dc63132c76"} Nov 28 13:41:02 crc kubenswrapper[4631]: I1128 13:41:02.866487 4631 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e9e2939de3cb8f012cdf07e85c04ccc0369e5628d09163745831e8dc63132c76" Nov 28 13:41:02 crc kubenswrapper[4631]: I1128 13:41:02.866572 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-tn4hn" Nov 28 13:41:02 crc kubenswrapper[4631]: I1128 13:41:02.874537 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-cs6bm" event={"ID":"476e6d13-87a1-4155-ae6b-f6453e9c2044","Type":"ContainerDied","Data":"581a79913f03d0080cc808bc6911303568135f1d477a940396f37410ddadfd83"} Nov 28 13:41:02 crc kubenswrapper[4631]: I1128 13:41:02.874600 4631 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="581a79913f03d0080cc808bc6911303568135f1d477a940396f37410ddadfd83" Nov 28 13:41:02 crc kubenswrapper[4631]: I1128 13:41:02.874691 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-cs6bm" Nov 28 13:41:02 crc kubenswrapper[4631]: I1128 13:41:02.986408 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Nov 28 13:41:02 crc kubenswrapper[4631]: E1128 13:41:02.986990 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dc4f7c59-e6c6-4659-8d2d-8e584afbe2ab" containerName="dnsmasq-dns" Nov 28 13:41:02 crc kubenswrapper[4631]: I1128 13:41:02.987011 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="dc4f7c59-e6c6-4659-8d2d-8e584afbe2ab" containerName="dnsmasq-dns" Nov 28 13:41:02 crc kubenswrapper[4631]: E1128 13:41:02.987040 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dc4f7c59-e6c6-4659-8d2d-8e584afbe2ab" containerName="init" Nov 28 13:41:02 crc kubenswrapper[4631]: I1128 13:41:02.987064 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="dc4f7c59-e6c6-4659-8d2d-8e584afbe2ab" containerName="init" Nov 28 13:41:02 crc kubenswrapper[4631]: E1128 13:41:02.987072 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b7cc6f7-1a01-4c53-818d-f40548209a13" containerName="nova-cell1-conductor-db-sync" Nov 28 13:41:02 crc kubenswrapper[4631]: I1128 13:41:02.987079 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b7cc6f7-1a01-4c53-818d-f40548209a13" containerName="nova-cell1-conductor-db-sync" Nov 28 13:41:02 crc kubenswrapper[4631]: E1128 13:41:02.987095 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="476e6d13-87a1-4155-ae6b-f6453e9c2044" containerName="nova-manage" Nov 28 13:41:02 crc kubenswrapper[4631]: I1128 13:41:02.987103 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="476e6d13-87a1-4155-ae6b-f6453e9c2044" containerName="nova-manage" Nov 28 13:41:02 crc kubenswrapper[4631]: I1128 13:41:02.987338 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="6b7cc6f7-1a01-4c53-818d-f40548209a13" containerName="nova-cell1-conductor-db-sync" Nov 28 13:41:02 crc kubenswrapper[4631]: I1128 13:41:02.987351 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="dc4f7c59-e6c6-4659-8d2d-8e584afbe2ab" containerName="dnsmasq-dns" Nov 28 13:41:02 crc kubenswrapper[4631]: I1128 13:41:02.987382 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="476e6d13-87a1-4155-ae6b-f6453e9c2044" containerName="nova-manage" Nov 28 13:41:02 crc kubenswrapper[4631]: I1128 13:41:02.988282 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Nov 28 13:41:02 crc kubenswrapper[4631]: I1128 13:41:02.992012 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Nov 28 13:41:03 crc kubenswrapper[4631]: I1128 13:41:03.009596 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Nov 28 13:41:03 crc kubenswrapper[4631]: I1128 13:41:03.038342 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0db994a3-5269-44c5-8a6f-8b7fdc798c3d-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"0db994a3-5269-44c5-8a6f-8b7fdc798c3d\") " pod="openstack/nova-cell1-conductor-0" Nov 28 13:41:03 crc kubenswrapper[4631]: I1128 13:41:03.038492 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0db994a3-5269-44c5-8a6f-8b7fdc798c3d-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"0db994a3-5269-44c5-8a6f-8b7fdc798c3d\") " pod="openstack/nova-cell1-conductor-0" Nov 28 13:41:03 crc kubenswrapper[4631]: I1128 13:41:03.038523 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z9m2z\" (UniqueName: \"kubernetes.io/projected/0db994a3-5269-44c5-8a6f-8b7fdc798c3d-kube-api-access-z9m2z\") pod \"nova-cell1-conductor-0\" (UID: \"0db994a3-5269-44c5-8a6f-8b7fdc798c3d\") " pod="openstack/nova-cell1-conductor-0" Nov 28 13:41:03 crc kubenswrapper[4631]: I1128 13:41:03.140145 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0db994a3-5269-44c5-8a6f-8b7fdc798c3d-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"0db994a3-5269-44c5-8a6f-8b7fdc798c3d\") " pod="openstack/nova-cell1-conductor-0" Nov 28 13:41:03 crc kubenswrapper[4631]: I1128 13:41:03.140640 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0db994a3-5269-44c5-8a6f-8b7fdc798c3d-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"0db994a3-5269-44c5-8a6f-8b7fdc798c3d\") " pod="openstack/nova-cell1-conductor-0" Nov 28 13:41:03 crc kubenswrapper[4631]: I1128 13:41:03.140749 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z9m2z\" (UniqueName: \"kubernetes.io/projected/0db994a3-5269-44c5-8a6f-8b7fdc798c3d-kube-api-access-z9m2z\") pod \"nova-cell1-conductor-0\" (UID: \"0db994a3-5269-44c5-8a6f-8b7fdc798c3d\") " pod="openstack/nova-cell1-conductor-0" Nov 28 13:41:03 crc kubenswrapper[4631]: I1128 13:41:03.144727 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0db994a3-5269-44c5-8a6f-8b7fdc798c3d-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"0db994a3-5269-44c5-8a6f-8b7fdc798c3d\") " pod="openstack/nova-cell1-conductor-0" Nov 28 13:41:03 crc kubenswrapper[4631]: I1128 13:41:03.144943 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0db994a3-5269-44c5-8a6f-8b7fdc798c3d-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"0db994a3-5269-44c5-8a6f-8b7fdc798c3d\") " pod="openstack/nova-cell1-conductor-0" Nov 28 13:41:03 crc kubenswrapper[4631]: I1128 13:41:03.166530 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 28 13:41:03 crc kubenswrapper[4631]: I1128 13:41:03.166827 4631 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="c7ac2d63-abe0-4862-a3a2-5d2b1dc87ac9" containerName="nova-api-log" containerID="cri-o://a3329f50bfdfc20c2c25e44e2b0cb05371b0dfebb65a54cee815ad41c7dae1f2" gracePeriod=30 Nov 28 13:41:03 crc kubenswrapper[4631]: I1128 13:41:03.167351 4631 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="c7ac2d63-abe0-4862-a3a2-5d2b1dc87ac9" containerName="nova-api-api" containerID="cri-o://cc8137bbcd37eea46d3f7488a36a67c1f80733d941fae1fa3f4b4ad9d84ecbd2" gracePeriod=30 Nov 28 13:41:03 crc kubenswrapper[4631]: I1128 13:41:03.186102 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z9m2z\" (UniqueName: \"kubernetes.io/projected/0db994a3-5269-44c5-8a6f-8b7fdc798c3d-kube-api-access-z9m2z\") pod \"nova-cell1-conductor-0\" (UID: \"0db994a3-5269-44c5-8a6f-8b7fdc798c3d\") " pod="openstack/nova-cell1-conductor-0" Nov 28 13:41:03 crc kubenswrapper[4631]: I1128 13:41:03.199086 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Nov 28 13:41:03 crc kubenswrapper[4631]: I1128 13:41:03.199838 4631 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="e9e6bbdb-eac5-47db-80e5-b438965fa8aa" containerName="nova-scheduler-scheduler" containerID="cri-o://1979d0923441a642c786ab0b83653968fbe20f2eaa0e71ce9909faee2d665e9a" gracePeriod=30 Nov 28 13:41:03 crc kubenswrapper[4631]: I1128 13:41:03.214497 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 28 13:41:03 crc kubenswrapper[4631]: I1128 13:41:03.214844 4631 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="f487af97-9737-48cf-a0ab-a442f143929b" containerName="nova-metadata-log" containerID="cri-o://9d45d82ad9a1cab47bbcabf7ac8ec4a6f275787e84b79703b999c96b65a918c2" gracePeriod=30 Nov 28 13:41:03 crc kubenswrapper[4631]: I1128 13:41:03.215052 4631 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="f487af97-9737-48cf-a0ab-a442f143929b" containerName="nova-metadata-metadata" containerID="cri-o://0176fdd3cef121c949afaf075e55b66120e88ef916c4a388add9afe002dcac8d" gracePeriod=30 Nov 28 13:41:03 crc kubenswrapper[4631]: I1128 13:41:03.305748 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Nov 28 13:41:03 crc kubenswrapper[4631]: I1128 13:41:03.886763 4631 generic.go:334] "Generic (PLEG): container finished" podID="f487af97-9737-48cf-a0ab-a442f143929b" containerID="0176fdd3cef121c949afaf075e55b66120e88ef916c4a388add9afe002dcac8d" exitCode=0 Nov 28 13:41:03 crc kubenswrapper[4631]: I1128 13:41:03.887503 4631 generic.go:334] "Generic (PLEG): container finished" podID="f487af97-9737-48cf-a0ab-a442f143929b" containerID="9d45d82ad9a1cab47bbcabf7ac8ec4a6f275787e84b79703b999c96b65a918c2" exitCode=143 Nov 28 13:41:03 crc kubenswrapper[4631]: I1128 13:41:03.886869 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f487af97-9737-48cf-a0ab-a442f143929b","Type":"ContainerDied","Data":"0176fdd3cef121c949afaf075e55b66120e88ef916c4a388add9afe002dcac8d"} Nov 28 13:41:03 crc kubenswrapper[4631]: I1128 13:41:03.887588 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f487af97-9737-48cf-a0ab-a442f143929b","Type":"ContainerDied","Data":"9d45d82ad9a1cab47bbcabf7ac8ec4a6f275787e84b79703b999c96b65a918c2"} Nov 28 13:41:03 crc kubenswrapper[4631]: I1128 13:41:03.887609 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f487af97-9737-48cf-a0ab-a442f143929b","Type":"ContainerDied","Data":"97e91ebed5dc0241ab9ba4c5f47ca6bc4f76583f75a4bb410f4d257b5efba497"} Nov 28 13:41:03 crc kubenswrapper[4631]: I1128 13:41:03.887624 4631 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="97e91ebed5dc0241ab9ba4c5f47ca6bc4f76583f75a4bb410f4d257b5efba497" Nov 28 13:41:03 crc kubenswrapper[4631]: I1128 13:41:03.890234 4631 generic.go:334] "Generic (PLEG): container finished" podID="c7ac2d63-abe0-4862-a3a2-5d2b1dc87ac9" containerID="a3329f50bfdfc20c2c25e44e2b0cb05371b0dfebb65a54cee815ad41c7dae1f2" exitCode=143 Nov 28 13:41:03 crc kubenswrapper[4631]: I1128 13:41:03.890302 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"c7ac2d63-abe0-4862-a3a2-5d2b1dc87ac9","Type":"ContainerDied","Data":"a3329f50bfdfc20c2c25e44e2b0cb05371b0dfebb65a54cee815ad41c7dae1f2"} Nov 28 13:41:03 crc kubenswrapper[4631]: I1128 13:41:03.929860 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Nov 28 13:41:04 crc kubenswrapper[4631]: I1128 13:41:04.032965 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 28 13:41:04 crc kubenswrapper[4631]: I1128 13:41:04.179614 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ms9vf\" (UniqueName: \"kubernetes.io/projected/f487af97-9737-48cf-a0ab-a442f143929b-kube-api-access-ms9vf\") pod \"f487af97-9737-48cf-a0ab-a442f143929b\" (UID: \"f487af97-9737-48cf-a0ab-a442f143929b\") " Nov 28 13:41:04 crc kubenswrapper[4631]: I1128 13:41:04.179694 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/f487af97-9737-48cf-a0ab-a442f143929b-nova-metadata-tls-certs\") pod \"f487af97-9737-48cf-a0ab-a442f143929b\" (UID: \"f487af97-9737-48cf-a0ab-a442f143929b\") " Nov 28 13:41:04 crc kubenswrapper[4631]: I1128 13:41:04.179929 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f487af97-9737-48cf-a0ab-a442f143929b-logs\") pod \"f487af97-9737-48cf-a0ab-a442f143929b\" (UID: \"f487af97-9737-48cf-a0ab-a442f143929b\") " Nov 28 13:41:04 crc kubenswrapper[4631]: I1128 13:41:04.179965 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f487af97-9737-48cf-a0ab-a442f143929b-config-data\") pod \"f487af97-9737-48cf-a0ab-a442f143929b\" (UID: \"f487af97-9737-48cf-a0ab-a442f143929b\") " Nov 28 13:41:04 crc kubenswrapper[4631]: I1128 13:41:04.180090 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f487af97-9737-48cf-a0ab-a442f143929b-combined-ca-bundle\") pod \"f487af97-9737-48cf-a0ab-a442f143929b\" (UID: \"f487af97-9737-48cf-a0ab-a442f143929b\") " Nov 28 13:41:04 crc kubenswrapper[4631]: I1128 13:41:04.180683 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f487af97-9737-48cf-a0ab-a442f143929b-logs" (OuterVolumeSpecName: "logs") pod "f487af97-9737-48cf-a0ab-a442f143929b" (UID: "f487af97-9737-48cf-a0ab-a442f143929b"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 13:41:04 crc kubenswrapper[4631]: I1128 13:41:04.197871 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f487af97-9737-48cf-a0ab-a442f143929b-kube-api-access-ms9vf" (OuterVolumeSpecName: "kube-api-access-ms9vf") pod "f487af97-9737-48cf-a0ab-a442f143929b" (UID: "f487af97-9737-48cf-a0ab-a442f143929b"). InnerVolumeSpecName "kube-api-access-ms9vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:41:04 crc kubenswrapper[4631]: I1128 13:41:04.238614 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f487af97-9737-48cf-a0ab-a442f143929b-config-data" (OuterVolumeSpecName: "config-data") pod "f487af97-9737-48cf-a0ab-a442f143929b" (UID: "f487af97-9737-48cf-a0ab-a442f143929b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:41:04 crc kubenswrapper[4631]: I1128 13:41:04.257303 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f487af97-9737-48cf-a0ab-a442f143929b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f487af97-9737-48cf-a0ab-a442f143929b" (UID: "f487af97-9737-48cf-a0ab-a442f143929b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:41:04 crc kubenswrapper[4631]: I1128 13:41:04.273497 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f487af97-9737-48cf-a0ab-a442f143929b-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "f487af97-9737-48cf-a0ab-a442f143929b" (UID: "f487af97-9737-48cf-a0ab-a442f143929b"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:41:04 crc kubenswrapper[4631]: I1128 13:41:04.288985 4631 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f487af97-9737-48cf-a0ab-a442f143929b-logs\") on node \"crc\" DevicePath \"\"" Nov 28 13:41:04 crc kubenswrapper[4631]: I1128 13:41:04.289527 4631 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f487af97-9737-48cf-a0ab-a442f143929b-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 13:41:04 crc kubenswrapper[4631]: I1128 13:41:04.289585 4631 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f487af97-9737-48cf-a0ab-a442f143929b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 13:41:04 crc kubenswrapper[4631]: I1128 13:41:04.289609 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ms9vf\" (UniqueName: \"kubernetes.io/projected/f487af97-9737-48cf-a0ab-a442f143929b-kube-api-access-ms9vf\") on node \"crc\" DevicePath \"\"" Nov 28 13:41:04 crc kubenswrapper[4631]: I1128 13:41:04.289621 4631 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/f487af97-9737-48cf-a0ab-a442f143929b-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 28 13:41:04 crc kubenswrapper[4631]: I1128 13:41:04.905971 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 28 13:41:04 crc kubenswrapper[4631]: I1128 13:41:04.906022 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"0db994a3-5269-44c5-8a6f-8b7fdc798c3d","Type":"ContainerStarted","Data":"6747e9269bf1cb32e19002485bc00609db031bab8ef0dba9102853699bb5187f"} Nov 28 13:41:04 crc kubenswrapper[4631]: I1128 13:41:04.906089 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"0db994a3-5269-44c5-8a6f-8b7fdc798c3d","Type":"ContainerStarted","Data":"a7492dd6d74529ea64f32c33d122cebd4d4c5beef5959552f7eb173b8c1f72f1"} Nov 28 13:41:04 crc kubenswrapper[4631]: I1128 13:41:04.906224 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Nov 28 13:41:04 crc kubenswrapper[4631]: I1128 13:41:04.936250 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.936219523 podStartE2EDuration="2.936219523s" podCreationTimestamp="2025-11-28 13:41:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 13:41:04.934090073 +0000 UTC m=+1221.741393417" watchObservedRunningTime="2025-11-28 13:41:04.936219523 +0000 UTC m=+1221.743522867" Nov 28 13:41:04 crc kubenswrapper[4631]: I1128 13:41:04.966574 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 28 13:41:04 crc kubenswrapper[4631]: I1128 13:41:04.983675 4631 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Nov 28 13:41:05 crc kubenswrapper[4631]: I1128 13:41:05.000979 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Nov 28 13:41:05 crc kubenswrapper[4631]: E1128 13:41:05.005606 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f487af97-9737-48cf-a0ab-a442f143929b" containerName="nova-metadata-log" Nov 28 13:41:05 crc kubenswrapper[4631]: I1128 13:41:05.005645 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="f487af97-9737-48cf-a0ab-a442f143929b" containerName="nova-metadata-log" Nov 28 13:41:05 crc kubenswrapper[4631]: E1128 13:41:05.005666 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f487af97-9737-48cf-a0ab-a442f143929b" containerName="nova-metadata-metadata" Nov 28 13:41:05 crc kubenswrapper[4631]: I1128 13:41:05.005673 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="f487af97-9737-48cf-a0ab-a442f143929b" containerName="nova-metadata-metadata" Nov 28 13:41:05 crc kubenswrapper[4631]: I1128 13:41:05.005959 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="f487af97-9737-48cf-a0ab-a442f143929b" containerName="nova-metadata-metadata" Nov 28 13:41:05 crc kubenswrapper[4631]: I1128 13:41:05.005993 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="f487af97-9737-48cf-a0ab-a442f143929b" containerName="nova-metadata-log" Nov 28 13:41:05 crc kubenswrapper[4631]: I1128 13:41:05.007326 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 28 13:41:05 crc kubenswrapper[4631]: I1128 13:41:05.010198 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Nov 28 13:41:05 crc kubenswrapper[4631]: I1128 13:41:05.012395 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Nov 28 13:41:05 crc kubenswrapper[4631]: I1128 13:41:05.025278 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 28 13:41:05 crc kubenswrapper[4631]: I1128 13:41:05.107795 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/fb5b797e-a328-4f2e-a239-333213537f2a-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"fb5b797e-a328-4f2e-a239-333213537f2a\") " pod="openstack/nova-metadata-0" Nov 28 13:41:05 crc kubenswrapper[4631]: I1128 13:41:05.107845 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v5wjs\" (UniqueName: \"kubernetes.io/projected/fb5b797e-a328-4f2e-a239-333213537f2a-kube-api-access-v5wjs\") pod \"nova-metadata-0\" (UID: \"fb5b797e-a328-4f2e-a239-333213537f2a\") " pod="openstack/nova-metadata-0" Nov 28 13:41:05 crc kubenswrapper[4631]: I1128 13:41:05.107913 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fb5b797e-a328-4f2e-a239-333213537f2a-logs\") pod \"nova-metadata-0\" (UID: \"fb5b797e-a328-4f2e-a239-333213537f2a\") " pod="openstack/nova-metadata-0" Nov 28 13:41:05 crc kubenswrapper[4631]: I1128 13:41:05.108077 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fb5b797e-a328-4f2e-a239-333213537f2a-config-data\") pod \"nova-metadata-0\" (UID: \"fb5b797e-a328-4f2e-a239-333213537f2a\") " pod="openstack/nova-metadata-0" Nov 28 13:41:05 crc kubenswrapper[4631]: I1128 13:41:05.108103 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb5b797e-a328-4f2e-a239-333213537f2a-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"fb5b797e-a328-4f2e-a239-333213537f2a\") " pod="openstack/nova-metadata-0" Nov 28 13:41:05 crc kubenswrapper[4631]: I1128 13:41:05.210421 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fb5b797e-a328-4f2e-a239-333213537f2a-config-data\") pod \"nova-metadata-0\" (UID: \"fb5b797e-a328-4f2e-a239-333213537f2a\") " pod="openstack/nova-metadata-0" Nov 28 13:41:05 crc kubenswrapper[4631]: I1128 13:41:05.210524 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb5b797e-a328-4f2e-a239-333213537f2a-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"fb5b797e-a328-4f2e-a239-333213537f2a\") " pod="openstack/nova-metadata-0" Nov 28 13:41:05 crc kubenswrapper[4631]: I1128 13:41:05.211395 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/fb5b797e-a328-4f2e-a239-333213537f2a-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"fb5b797e-a328-4f2e-a239-333213537f2a\") " pod="openstack/nova-metadata-0" Nov 28 13:41:05 crc kubenswrapper[4631]: I1128 13:41:05.211422 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v5wjs\" (UniqueName: \"kubernetes.io/projected/fb5b797e-a328-4f2e-a239-333213537f2a-kube-api-access-v5wjs\") pod \"nova-metadata-0\" (UID: \"fb5b797e-a328-4f2e-a239-333213537f2a\") " pod="openstack/nova-metadata-0" Nov 28 13:41:05 crc kubenswrapper[4631]: I1128 13:41:05.211454 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fb5b797e-a328-4f2e-a239-333213537f2a-logs\") pod \"nova-metadata-0\" (UID: \"fb5b797e-a328-4f2e-a239-333213537f2a\") " pod="openstack/nova-metadata-0" Nov 28 13:41:05 crc kubenswrapper[4631]: I1128 13:41:05.211871 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fb5b797e-a328-4f2e-a239-333213537f2a-logs\") pod \"nova-metadata-0\" (UID: \"fb5b797e-a328-4f2e-a239-333213537f2a\") " pod="openstack/nova-metadata-0" Nov 28 13:41:05 crc kubenswrapper[4631]: I1128 13:41:05.216048 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb5b797e-a328-4f2e-a239-333213537f2a-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"fb5b797e-a328-4f2e-a239-333213537f2a\") " pod="openstack/nova-metadata-0" Nov 28 13:41:05 crc kubenswrapper[4631]: I1128 13:41:05.225049 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fb5b797e-a328-4f2e-a239-333213537f2a-config-data\") pod \"nova-metadata-0\" (UID: \"fb5b797e-a328-4f2e-a239-333213537f2a\") " pod="openstack/nova-metadata-0" Nov 28 13:41:05 crc kubenswrapper[4631]: I1128 13:41:05.226015 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/fb5b797e-a328-4f2e-a239-333213537f2a-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"fb5b797e-a328-4f2e-a239-333213537f2a\") " pod="openstack/nova-metadata-0" Nov 28 13:41:05 crc kubenswrapper[4631]: I1128 13:41:05.252143 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v5wjs\" (UniqueName: \"kubernetes.io/projected/fb5b797e-a328-4f2e-a239-333213537f2a-kube-api-access-v5wjs\") pod \"nova-metadata-0\" (UID: \"fb5b797e-a328-4f2e-a239-333213537f2a\") " pod="openstack/nova-metadata-0" Nov 28 13:41:05 crc kubenswrapper[4631]: E1128 13:41:05.259484 4631 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="1979d0923441a642c786ab0b83653968fbe20f2eaa0e71ce9909faee2d665e9a" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Nov 28 13:41:05 crc kubenswrapper[4631]: E1128 13:41:05.262026 4631 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="1979d0923441a642c786ab0b83653968fbe20f2eaa0e71ce9909faee2d665e9a" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Nov 28 13:41:05 crc kubenswrapper[4631]: E1128 13:41:05.281450 4631 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="1979d0923441a642c786ab0b83653968fbe20f2eaa0e71ce9909faee2d665e9a" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Nov 28 13:41:05 crc kubenswrapper[4631]: E1128 13:41:05.281548 4631 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="e9e6bbdb-eac5-47db-80e5-b438965fa8aa" containerName="nova-scheduler-scheduler" Nov 28 13:41:05 crc kubenswrapper[4631]: I1128 13:41:05.329331 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 28 13:41:05 crc kubenswrapper[4631]: I1128 13:41:05.333138 4631 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-64f8fdf97b-jn5mg" podUID="111c5349-528d-4856-a3e4-1277e3b10889" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.146:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.146:8443: connect: connection refused" Nov 28 13:41:05 crc kubenswrapper[4631]: I1128 13:41:05.527922 4631 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f487af97-9737-48cf-a0ab-a442f143929b" path="/var/lib/kubelet/pods/f487af97-9737-48cf-a0ab-a442f143929b/volumes" Nov 28 13:41:05 crc kubenswrapper[4631]: I1128 13:41:05.974768 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 28 13:41:05 crc kubenswrapper[4631]: W1128 13:41:05.982512 4631 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfb5b797e_a328_4f2e_a239_333213537f2a.slice/crio-c8ba24d666dddd639a797bd7984d0503b1da79c33affc620b326fe5ae1efcdad WatchSource:0}: Error finding container c8ba24d666dddd639a797bd7984d0503b1da79c33affc620b326fe5ae1efcdad: Status 404 returned error can't find the container with id c8ba24d666dddd639a797bd7984d0503b1da79c33affc620b326fe5ae1efcdad Nov 28 13:41:06 crc kubenswrapper[4631]: I1128 13:41:06.941238 4631 generic.go:334] "Generic (PLEG): container finished" podID="c7ac2d63-abe0-4862-a3a2-5d2b1dc87ac9" containerID="cc8137bbcd37eea46d3f7488a36a67c1f80733d941fae1fa3f4b4ad9d84ecbd2" exitCode=0 Nov 28 13:41:06 crc kubenswrapper[4631]: I1128 13:41:06.941339 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"c7ac2d63-abe0-4862-a3a2-5d2b1dc87ac9","Type":"ContainerDied","Data":"cc8137bbcd37eea46d3f7488a36a67c1f80733d941fae1fa3f4b4ad9d84ecbd2"} Nov 28 13:41:06 crc kubenswrapper[4631]: I1128 13:41:06.953335 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"fb5b797e-a328-4f2e-a239-333213537f2a","Type":"ContainerStarted","Data":"6b05c3bfd6fcbcd36095c9145b56050d4042f642dc03e2d34a74e87996ecbe6b"} Nov 28 13:41:06 crc kubenswrapper[4631]: I1128 13:41:06.953423 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"fb5b797e-a328-4f2e-a239-333213537f2a","Type":"ContainerStarted","Data":"3932276d04ef4e1355b3d36c0f83cfb37b4724d2af10ecd90370a7c5ae212aa6"} Nov 28 13:41:06 crc kubenswrapper[4631]: I1128 13:41:06.953441 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"fb5b797e-a328-4f2e-a239-333213537f2a","Type":"ContainerStarted","Data":"c8ba24d666dddd639a797bd7984d0503b1da79c33affc620b326fe5ae1efcdad"} Nov 28 13:41:06 crc kubenswrapper[4631]: I1128 13:41:06.965364 4631 generic.go:334] "Generic (PLEG): container finished" podID="e9e6bbdb-eac5-47db-80e5-b438965fa8aa" containerID="1979d0923441a642c786ab0b83653968fbe20f2eaa0e71ce9909faee2d665e9a" exitCode=0 Nov 28 13:41:06 crc kubenswrapper[4631]: I1128 13:41:06.965418 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"e9e6bbdb-eac5-47db-80e5-b438965fa8aa","Type":"ContainerDied","Data":"1979d0923441a642c786ab0b83653968fbe20f2eaa0e71ce9909faee2d665e9a"} Nov 28 13:41:07 crc kubenswrapper[4631]: I1128 13:41:07.062138 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 28 13:41:07 crc kubenswrapper[4631]: I1128 13:41:07.104627 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=3.104602938 podStartE2EDuration="3.104602938s" podCreationTimestamp="2025-11-28 13:41:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 13:41:06.984561882 +0000 UTC m=+1223.791865216" watchObservedRunningTime="2025-11-28 13:41:07.104602938 +0000 UTC m=+1223.911906282" Nov 28 13:41:07 crc kubenswrapper[4631]: I1128 13:41:07.167054 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qqsml\" (UniqueName: \"kubernetes.io/projected/e9e6bbdb-eac5-47db-80e5-b438965fa8aa-kube-api-access-qqsml\") pod \"e9e6bbdb-eac5-47db-80e5-b438965fa8aa\" (UID: \"e9e6bbdb-eac5-47db-80e5-b438965fa8aa\") " Nov 28 13:41:07 crc kubenswrapper[4631]: I1128 13:41:07.167184 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e9e6bbdb-eac5-47db-80e5-b438965fa8aa-config-data\") pod \"e9e6bbdb-eac5-47db-80e5-b438965fa8aa\" (UID: \"e9e6bbdb-eac5-47db-80e5-b438965fa8aa\") " Nov 28 13:41:07 crc kubenswrapper[4631]: I1128 13:41:07.167622 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e9e6bbdb-eac5-47db-80e5-b438965fa8aa-combined-ca-bundle\") pod \"e9e6bbdb-eac5-47db-80e5-b438965fa8aa\" (UID: \"e9e6bbdb-eac5-47db-80e5-b438965fa8aa\") " Nov 28 13:41:07 crc kubenswrapper[4631]: I1128 13:41:07.170102 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 28 13:41:07 crc kubenswrapper[4631]: I1128 13:41:07.183017 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e9e6bbdb-eac5-47db-80e5-b438965fa8aa-kube-api-access-qqsml" (OuterVolumeSpecName: "kube-api-access-qqsml") pod "e9e6bbdb-eac5-47db-80e5-b438965fa8aa" (UID: "e9e6bbdb-eac5-47db-80e5-b438965fa8aa"). InnerVolumeSpecName "kube-api-access-qqsml". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:41:07 crc kubenswrapper[4631]: I1128 13:41:07.230532 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e9e6bbdb-eac5-47db-80e5-b438965fa8aa-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e9e6bbdb-eac5-47db-80e5-b438965fa8aa" (UID: "e9e6bbdb-eac5-47db-80e5-b438965fa8aa"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:41:07 crc kubenswrapper[4631]: I1128 13:41:07.247007 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e9e6bbdb-eac5-47db-80e5-b438965fa8aa-config-data" (OuterVolumeSpecName: "config-data") pod "e9e6bbdb-eac5-47db-80e5-b438965fa8aa" (UID: "e9e6bbdb-eac5-47db-80e5-b438965fa8aa"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:41:07 crc kubenswrapper[4631]: I1128 13:41:07.272847 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c7ac2d63-abe0-4862-a3a2-5d2b1dc87ac9-logs\") pod \"c7ac2d63-abe0-4862-a3a2-5d2b1dc87ac9\" (UID: \"c7ac2d63-abe0-4862-a3a2-5d2b1dc87ac9\") " Nov 28 13:41:07 crc kubenswrapper[4631]: I1128 13:41:07.272914 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xbsqb\" (UniqueName: \"kubernetes.io/projected/c7ac2d63-abe0-4862-a3a2-5d2b1dc87ac9-kube-api-access-xbsqb\") pod \"c7ac2d63-abe0-4862-a3a2-5d2b1dc87ac9\" (UID: \"c7ac2d63-abe0-4862-a3a2-5d2b1dc87ac9\") " Nov 28 13:41:07 crc kubenswrapper[4631]: I1128 13:41:07.272983 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c7ac2d63-abe0-4862-a3a2-5d2b1dc87ac9-combined-ca-bundle\") pod \"c7ac2d63-abe0-4862-a3a2-5d2b1dc87ac9\" (UID: \"c7ac2d63-abe0-4862-a3a2-5d2b1dc87ac9\") " Nov 28 13:41:07 crc kubenswrapper[4631]: I1128 13:41:07.273450 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c7ac2d63-abe0-4862-a3a2-5d2b1dc87ac9-config-data\") pod \"c7ac2d63-abe0-4862-a3a2-5d2b1dc87ac9\" (UID: \"c7ac2d63-abe0-4862-a3a2-5d2b1dc87ac9\") " Nov 28 13:41:07 crc kubenswrapper[4631]: I1128 13:41:07.273694 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c7ac2d63-abe0-4862-a3a2-5d2b1dc87ac9-logs" (OuterVolumeSpecName: "logs") pod "c7ac2d63-abe0-4862-a3a2-5d2b1dc87ac9" (UID: "c7ac2d63-abe0-4862-a3a2-5d2b1dc87ac9"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 13:41:07 crc kubenswrapper[4631]: I1128 13:41:07.274977 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qqsml\" (UniqueName: \"kubernetes.io/projected/e9e6bbdb-eac5-47db-80e5-b438965fa8aa-kube-api-access-qqsml\") on node \"crc\" DevicePath \"\"" Nov 28 13:41:07 crc kubenswrapper[4631]: I1128 13:41:07.275007 4631 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e9e6bbdb-eac5-47db-80e5-b438965fa8aa-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 13:41:07 crc kubenswrapper[4631]: I1128 13:41:07.275021 4631 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c7ac2d63-abe0-4862-a3a2-5d2b1dc87ac9-logs\") on node \"crc\" DevicePath \"\"" Nov 28 13:41:07 crc kubenswrapper[4631]: I1128 13:41:07.275034 4631 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e9e6bbdb-eac5-47db-80e5-b438965fa8aa-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 13:41:07 crc kubenswrapper[4631]: I1128 13:41:07.284711 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c7ac2d63-abe0-4862-a3a2-5d2b1dc87ac9-kube-api-access-xbsqb" (OuterVolumeSpecName: "kube-api-access-xbsqb") pod "c7ac2d63-abe0-4862-a3a2-5d2b1dc87ac9" (UID: "c7ac2d63-abe0-4862-a3a2-5d2b1dc87ac9"). InnerVolumeSpecName "kube-api-access-xbsqb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:41:07 crc kubenswrapper[4631]: I1128 13:41:07.307836 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c7ac2d63-abe0-4862-a3a2-5d2b1dc87ac9-config-data" (OuterVolumeSpecName: "config-data") pod "c7ac2d63-abe0-4862-a3a2-5d2b1dc87ac9" (UID: "c7ac2d63-abe0-4862-a3a2-5d2b1dc87ac9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:41:07 crc kubenswrapper[4631]: I1128 13:41:07.311546 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c7ac2d63-abe0-4862-a3a2-5d2b1dc87ac9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c7ac2d63-abe0-4862-a3a2-5d2b1dc87ac9" (UID: "c7ac2d63-abe0-4862-a3a2-5d2b1dc87ac9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:41:07 crc kubenswrapper[4631]: I1128 13:41:07.376973 4631 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c7ac2d63-abe0-4862-a3a2-5d2b1dc87ac9-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 13:41:07 crc kubenswrapper[4631]: I1128 13:41:07.377329 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xbsqb\" (UniqueName: \"kubernetes.io/projected/c7ac2d63-abe0-4862-a3a2-5d2b1dc87ac9-kube-api-access-xbsqb\") on node \"crc\" DevicePath \"\"" Nov 28 13:41:07 crc kubenswrapper[4631]: I1128 13:41:07.377343 4631 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c7ac2d63-abe0-4862-a3a2-5d2b1dc87ac9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 13:41:07 crc kubenswrapper[4631]: I1128 13:41:07.981133 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"c7ac2d63-abe0-4862-a3a2-5d2b1dc87ac9","Type":"ContainerDied","Data":"ff28b82ee28ede9ce6b12af991af71dee06a6d4d8170bdf238238bf642c83a35"} Nov 28 13:41:07 crc kubenswrapper[4631]: I1128 13:41:07.981229 4631 scope.go:117] "RemoveContainer" containerID="cc8137bbcd37eea46d3f7488a36a67c1f80733d941fae1fa3f4b4ad9d84ecbd2" Nov 28 13:41:07 crc kubenswrapper[4631]: I1128 13:41:07.981155 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 28 13:41:07 crc kubenswrapper[4631]: I1128 13:41:07.983690 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"e9e6bbdb-eac5-47db-80e5-b438965fa8aa","Type":"ContainerDied","Data":"541285d216b670f6796f84f4d8b4baf48e4e3762fd84e3d165c8c29734d32012"} Nov 28 13:41:07 crc kubenswrapper[4631]: I1128 13:41:07.983746 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 28 13:41:08 crc kubenswrapper[4631]: I1128 13:41:08.076508 4631 scope.go:117] "RemoveContainer" containerID="a3329f50bfdfc20c2c25e44e2b0cb05371b0dfebb65a54cee815ad41c7dae1f2" Nov 28 13:41:08 crc kubenswrapper[4631]: I1128 13:41:08.092233 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 28 13:41:08 crc kubenswrapper[4631]: I1128 13:41:08.114559 4631 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Nov 28 13:41:08 crc kubenswrapper[4631]: I1128 13:41:08.138405 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Nov 28 13:41:08 crc kubenswrapper[4631]: I1128 13:41:08.152444 4631 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Nov 28 13:41:08 crc kubenswrapper[4631]: I1128 13:41:08.152812 4631 scope.go:117] "RemoveContainer" containerID="1979d0923441a642c786ab0b83653968fbe20f2eaa0e71ce9909faee2d665e9a" Nov 28 13:41:08 crc kubenswrapper[4631]: I1128 13:41:08.171494 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Nov 28 13:41:08 crc kubenswrapper[4631]: E1128 13:41:08.172152 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c7ac2d63-abe0-4862-a3a2-5d2b1dc87ac9" containerName="nova-api-api" Nov 28 13:41:08 crc kubenswrapper[4631]: I1128 13:41:08.172235 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="c7ac2d63-abe0-4862-a3a2-5d2b1dc87ac9" containerName="nova-api-api" Nov 28 13:41:08 crc kubenswrapper[4631]: E1128 13:41:08.172312 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e9e6bbdb-eac5-47db-80e5-b438965fa8aa" containerName="nova-scheduler-scheduler" Nov 28 13:41:08 crc kubenswrapper[4631]: I1128 13:41:08.172373 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="e9e6bbdb-eac5-47db-80e5-b438965fa8aa" containerName="nova-scheduler-scheduler" Nov 28 13:41:08 crc kubenswrapper[4631]: E1128 13:41:08.172446 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c7ac2d63-abe0-4862-a3a2-5d2b1dc87ac9" containerName="nova-api-log" Nov 28 13:41:08 crc kubenswrapper[4631]: I1128 13:41:08.172505 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="c7ac2d63-abe0-4862-a3a2-5d2b1dc87ac9" containerName="nova-api-log" Nov 28 13:41:08 crc kubenswrapper[4631]: I1128 13:41:08.172784 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="e9e6bbdb-eac5-47db-80e5-b438965fa8aa" containerName="nova-scheduler-scheduler" Nov 28 13:41:08 crc kubenswrapper[4631]: I1128 13:41:08.172878 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="c7ac2d63-abe0-4862-a3a2-5d2b1dc87ac9" containerName="nova-api-api" Nov 28 13:41:08 crc kubenswrapper[4631]: I1128 13:41:08.172957 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="c7ac2d63-abe0-4862-a3a2-5d2b1dc87ac9" containerName="nova-api-log" Nov 28 13:41:08 crc kubenswrapper[4631]: I1128 13:41:08.174237 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 28 13:41:08 crc kubenswrapper[4631]: I1128 13:41:08.185648 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Nov 28 13:41:08 crc kubenswrapper[4631]: I1128 13:41:08.199161 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Nov 28 13:41:08 crc kubenswrapper[4631]: I1128 13:41:08.207541 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 28 13:41:08 crc kubenswrapper[4631]: I1128 13:41:08.213655 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Nov 28 13:41:08 crc kubenswrapper[4631]: I1128 13:41:08.218708 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 28 13:41:08 crc kubenswrapper[4631]: I1128 13:41:08.238082 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 28 13:41:08 crc kubenswrapper[4631]: I1128 13:41:08.300586 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8gjg5\" (UniqueName: \"kubernetes.io/projected/a52eb82e-ac7b-4a48-9021-4c450da3ecae-kube-api-access-8gjg5\") pod \"nova-scheduler-0\" (UID: \"a52eb82e-ac7b-4a48-9021-4c450da3ecae\") " pod="openstack/nova-scheduler-0" Nov 28 13:41:08 crc kubenswrapper[4631]: I1128 13:41:08.300668 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a52eb82e-ac7b-4a48-9021-4c450da3ecae-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"a52eb82e-ac7b-4a48-9021-4c450da3ecae\") " pod="openstack/nova-scheduler-0" Nov 28 13:41:08 crc kubenswrapper[4631]: I1128 13:41:08.300728 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gb2lq\" (UniqueName: \"kubernetes.io/projected/44c10c63-05fe-47e8-acc6-c927c9c3db7e-kube-api-access-gb2lq\") pod \"nova-api-0\" (UID: \"44c10c63-05fe-47e8-acc6-c927c9c3db7e\") " pod="openstack/nova-api-0" Nov 28 13:41:08 crc kubenswrapper[4631]: I1128 13:41:08.300747 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/44c10c63-05fe-47e8-acc6-c927c9c3db7e-config-data\") pod \"nova-api-0\" (UID: \"44c10c63-05fe-47e8-acc6-c927c9c3db7e\") " pod="openstack/nova-api-0" Nov 28 13:41:08 crc kubenswrapper[4631]: I1128 13:41:08.300774 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a52eb82e-ac7b-4a48-9021-4c450da3ecae-config-data\") pod \"nova-scheduler-0\" (UID: \"a52eb82e-ac7b-4a48-9021-4c450da3ecae\") " pod="openstack/nova-scheduler-0" Nov 28 13:41:08 crc kubenswrapper[4631]: I1128 13:41:08.300925 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/44c10c63-05fe-47e8-acc6-c927c9c3db7e-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"44c10c63-05fe-47e8-acc6-c927c9c3db7e\") " pod="openstack/nova-api-0" Nov 28 13:41:08 crc kubenswrapper[4631]: I1128 13:41:08.300989 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/44c10c63-05fe-47e8-acc6-c927c9c3db7e-logs\") pod \"nova-api-0\" (UID: \"44c10c63-05fe-47e8-acc6-c927c9c3db7e\") " pod="openstack/nova-api-0" Nov 28 13:41:08 crc kubenswrapper[4631]: I1128 13:41:08.403510 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a52eb82e-ac7b-4a48-9021-4c450da3ecae-config-data\") pod \"nova-scheduler-0\" (UID: \"a52eb82e-ac7b-4a48-9021-4c450da3ecae\") " pod="openstack/nova-scheduler-0" Nov 28 13:41:08 crc kubenswrapper[4631]: I1128 13:41:08.403570 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/44c10c63-05fe-47e8-acc6-c927c9c3db7e-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"44c10c63-05fe-47e8-acc6-c927c9c3db7e\") " pod="openstack/nova-api-0" Nov 28 13:41:08 crc kubenswrapper[4631]: I1128 13:41:08.403592 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/44c10c63-05fe-47e8-acc6-c927c9c3db7e-logs\") pod \"nova-api-0\" (UID: \"44c10c63-05fe-47e8-acc6-c927c9c3db7e\") " pod="openstack/nova-api-0" Nov 28 13:41:08 crc kubenswrapper[4631]: I1128 13:41:08.403693 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8gjg5\" (UniqueName: \"kubernetes.io/projected/a52eb82e-ac7b-4a48-9021-4c450da3ecae-kube-api-access-8gjg5\") pod \"nova-scheduler-0\" (UID: \"a52eb82e-ac7b-4a48-9021-4c450da3ecae\") " pod="openstack/nova-scheduler-0" Nov 28 13:41:08 crc kubenswrapper[4631]: I1128 13:41:08.403759 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a52eb82e-ac7b-4a48-9021-4c450da3ecae-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"a52eb82e-ac7b-4a48-9021-4c450da3ecae\") " pod="openstack/nova-scheduler-0" Nov 28 13:41:08 crc kubenswrapper[4631]: I1128 13:41:08.403822 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gb2lq\" (UniqueName: \"kubernetes.io/projected/44c10c63-05fe-47e8-acc6-c927c9c3db7e-kube-api-access-gb2lq\") pod \"nova-api-0\" (UID: \"44c10c63-05fe-47e8-acc6-c927c9c3db7e\") " pod="openstack/nova-api-0" Nov 28 13:41:08 crc kubenswrapper[4631]: I1128 13:41:08.403845 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/44c10c63-05fe-47e8-acc6-c927c9c3db7e-config-data\") pod \"nova-api-0\" (UID: \"44c10c63-05fe-47e8-acc6-c927c9c3db7e\") " pod="openstack/nova-api-0" Nov 28 13:41:08 crc kubenswrapper[4631]: I1128 13:41:08.404854 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/44c10c63-05fe-47e8-acc6-c927c9c3db7e-logs\") pod \"nova-api-0\" (UID: \"44c10c63-05fe-47e8-acc6-c927c9c3db7e\") " pod="openstack/nova-api-0" Nov 28 13:41:08 crc kubenswrapper[4631]: I1128 13:41:08.410667 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a52eb82e-ac7b-4a48-9021-4c450da3ecae-config-data\") pod \"nova-scheduler-0\" (UID: \"a52eb82e-ac7b-4a48-9021-4c450da3ecae\") " pod="openstack/nova-scheduler-0" Nov 28 13:41:08 crc kubenswrapper[4631]: I1128 13:41:08.410785 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/44c10c63-05fe-47e8-acc6-c927c9c3db7e-config-data\") pod \"nova-api-0\" (UID: \"44c10c63-05fe-47e8-acc6-c927c9c3db7e\") " pod="openstack/nova-api-0" Nov 28 13:41:08 crc kubenswrapper[4631]: I1128 13:41:08.410957 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a52eb82e-ac7b-4a48-9021-4c450da3ecae-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"a52eb82e-ac7b-4a48-9021-4c450da3ecae\") " pod="openstack/nova-scheduler-0" Nov 28 13:41:08 crc kubenswrapper[4631]: I1128 13:41:08.411640 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/44c10c63-05fe-47e8-acc6-c927c9c3db7e-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"44c10c63-05fe-47e8-acc6-c927c9c3db7e\") " pod="openstack/nova-api-0" Nov 28 13:41:08 crc kubenswrapper[4631]: I1128 13:41:08.441889 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gb2lq\" (UniqueName: \"kubernetes.io/projected/44c10c63-05fe-47e8-acc6-c927c9c3db7e-kube-api-access-gb2lq\") pod \"nova-api-0\" (UID: \"44c10c63-05fe-47e8-acc6-c927c9c3db7e\") " pod="openstack/nova-api-0" Nov 28 13:41:08 crc kubenswrapper[4631]: I1128 13:41:08.442977 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8gjg5\" (UniqueName: \"kubernetes.io/projected/a52eb82e-ac7b-4a48-9021-4c450da3ecae-kube-api-access-8gjg5\") pod \"nova-scheduler-0\" (UID: \"a52eb82e-ac7b-4a48-9021-4c450da3ecae\") " pod="openstack/nova-scheduler-0" Nov 28 13:41:08 crc kubenswrapper[4631]: I1128 13:41:08.504013 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 28 13:41:08 crc kubenswrapper[4631]: I1128 13:41:08.531799 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 28 13:41:09 crc kubenswrapper[4631]: E1128 13:41:09.020831 4631 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod111c5349_528d_4856_a3e4_1277e3b10889.slice/crio-3b826a5556ea3b27f6357d4cc74220f4bce293db6dbcb49fa5131013bef0e517.scope\": RecentStats: unable to find data in memory cache]" Nov 28 13:41:09 crc kubenswrapper[4631]: I1128 13:41:09.104663 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 28 13:41:09 crc kubenswrapper[4631]: I1128 13:41:09.143163 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 28 13:41:09 crc kubenswrapper[4631]: I1128 13:41:09.531655 4631 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c7ac2d63-abe0-4862-a3a2-5d2b1dc87ac9" path="/var/lib/kubelet/pods/c7ac2d63-abe0-4862-a3a2-5d2b1dc87ac9/volumes" Nov 28 13:41:09 crc kubenswrapper[4631]: I1128 13:41:09.532544 4631 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e9e6bbdb-eac5-47db-80e5-b438965fa8aa" path="/var/lib/kubelet/pods/e9e6bbdb-eac5-47db-80e5-b438965fa8aa/volumes" Nov 28 13:41:10 crc kubenswrapper[4631]: I1128 13:41:10.011614 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"a52eb82e-ac7b-4a48-9021-4c450da3ecae","Type":"ContainerStarted","Data":"9806c18b8805bf90bd761a4c5a6a35fa5f6a85fabc73ecead82d8bd9898ac626"} Nov 28 13:41:10 crc kubenswrapper[4631]: I1128 13:41:10.012107 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"a52eb82e-ac7b-4a48-9021-4c450da3ecae","Type":"ContainerStarted","Data":"398d71554c3839952b9b273209cc0a7a898e31e280d5458196d0361360d1bdf9"} Nov 28 13:41:10 crc kubenswrapper[4631]: I1128 13:41:10.014251 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"44c10c63-05fe-47e8-acc6-c927c9c3db7e","Type":"ContainerStarted","Data":"cf453b65de3fb82eee03c5027c805c5311f5c1efd14c3e31bbfb557b4dcd04a8"} Nov 28 13:41:10 crc kubenswrapper[4631]: I1128 13:41:10.014277 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"44c10c63-05fe-47e8-acc6-c927c9c3db7e","Type":"ContainerStarted","Data":"4103b4cf6b619a45021786723e84b8ad4e2fc43a3f9eab4611393b14c989a39e"} Nov 28 13:41:10 crc kubenswrapper[4631]: I1128 13:41:10.014317 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"44c10c63-05fe-47e8-acc6-c927c9c3db7e","Type":"ContainerStarted","Data":"8c94b35aaaf92a53e543b23309fc94a5193c4352d5814a7f17d376531709b814"} Nov 28 13:41:10 crc kubenswrapper[4631]: I1128 13:41:10.047192 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.047144293 podStartE2EDuration="2.047144293s" podCreationTimestamp="2025-11-28 13:41:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 13:41:10.03072414 +0000 UTC m=+1226.838027484" watchObservedRunningTime="2025-11-28 13:41:10.047144293 +0000 UTC m=+1226.854447637" Nov 28 13:41:10 crc kubenswrapper[4631]: I1128 13:41:10.330022 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 28 13:41:10 crc kubenswrapper[4631]: I1128 13:41:10.331332 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 28 13:41:10 crc kubenswrapper[4631]: I1128 13:41:10.593011 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-64f8fdf97b-jn5mg" Nov 28 13:41:10 crc kubenswrapper[4631]: I1128 13:41:10.634385 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.634336646 podStartE2EDuration="2.634336646s" podCreationTimestamp="2025-11-28 13:41:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 13:41:10.051866226 +0000 UTC m=+1226.859169570" watchObservedRunningTime="2025-11-28 13:41:10.634336646 +0000 UTC m=+1227.441639990" Nov 28 13:41:10 crc kubenswrapper[4631]: I1128 13:41:10.659214 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/111c5349-528d-4856-a3e4-1277e3b10889-horizon-secret-key\") pod \"111c5349-528d-4856-a3e4-1277e3b10889\" (UID: \"111c5349-528d-4856-a3e4-1277e3b10889\") " Nov 28 13:41:10 crc kubenswrapper[4631]: I1128 13:41:10.659414 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/111c5349-528d-4856-a3e4-1277e3b10889-logs\") pod \"111c5349-528d-4856-a3e4-1277e3b10889\" (UID: \"111c5349-528d-4856-a3e4-1277e3b10889\") " Nov 28 13:41:10 crc kubenswrapper[4631]: I1128 13:41:10.659626 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/111c5349-528d-4856-a3e4-1277e3b10889-combined-ca-bundle\") pod \"111c5349-528d-4856-a3e4-1277e3b10889\" (UID: \"111c5349-528d-4856-a3e4-1277e3b10889\") " Nov 28 13:41:10 crc kubenswrapper[4631]: I1128 13:41:10.659825 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k8tns\" (UniqueName: \"kubernetes.io/projected/111c5349-528d-4856-a3e4-1277e3b10889-kube-api-access-k8tns\") pod \"111c5349-528d-4856-a3e4-1277e3b10889\" (UID: \"111c5349-528d-4856-a3e4-1277e3b10889\") " Nov 28 13:41:10 crc kubenswrapper[4631]: I1128 13:41:10.659960 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/111c5349-528d-4856-a3e4-1277e3b10889-config-data\") pod \"111c5349-528d-4856-a3e4-1277e3b10889\" (UID: \"111c5349-528d-4856-a3e4-1277e3b10889\") " Nov 28 13:41:10 crc kubenswrapper[4631]: I1128 13:41:10.660179 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/111c5349-528d-4856-a3e4-1277e3b10889-horizon-tls-certs\") pod \"111c5349-528d-4856-a3e4-1277e3b10889\" (UID: \"111c5349-528d-4856-a3e4-1277e3b10889\") " Nov 28 13:41:10 crc kubenswrapper[4631]: I1128 13:41:10.660277 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/111c5349-528d-4856-a3e4-1277e3b10889-scripts\") pod \"111c5349-528d-4856-a3e4-1277e3b10889\" (UID: \"111c5349-528d-4856-a3e4-1277e3b10889\") " Nov 28 13:41:10 crc kubenswrapper[4631]: I1128 13:41:10.663545 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/111c5349-528d-4856-a3e4-1277e3b10889-logs" (OuterVolumeSpecName: "logs") pod "111c5349-528d-4856-a3e4-1277e3b10889" (UID: "111c5349-528d-4856-a3e4-1277e3b10889"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 13:41:10 crc kubenswrapper[4631]: I1128 13:41:10.683205 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/111c5349-528d-4856-a3e4-1277e3b10889-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "111c5349-528d-4856-a3e4-1277e3b10889" (UID: "111c5349-528d-4856-a3e4-1277e3b10889"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:41:10 crc kubenswrapper[4631]: I1128 13:41:10.687019 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/111c5349-528d-4856-a3e4-1277e3b10889-kube-api-access-k8tns" (OuterVolumeSpecName: "kube-api-access-k8tns") pod "111c5349-528d-4856-a3e4-1277e3b10889" (UID: "111c5349-528d-4856-a3e4-1277e3b10889"). InnerVolumeSpecName "kube-api-access-k8tns". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:41:10 crc kubenswrapper[4631]: I1128 13:41:10.689942 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/111c5349-528d-4856-a3e4-1277e3b10889-scripts" (OuterVolumeSpecName: "scripts") pod "111c5349-528d-4856-a3e4-1277e3b10889" (UID: "111c5349-528d-4856-a3e4-1277e3b10889"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 13:41:10 crc kubenswrapper[4631]: I1128 13:41:10.693429 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/111c5349-528d-4856-a3e4-1277e3b10889-config-data" (OuterVolumeSpecName: "config-data") pod "111c5349-528d-4856-a3e4-1277e3b10889" (UID: "111c5349-528d-4856-a3e4-1277e3b10889"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 13:41:10 crc kubenswrapper[4631]: I1128 13:41:10.707558 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/111c5349-528d-4856-a3e4-1277e3b10889-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "111c5349-528d-4856-a3e4-1277e3b10889" (UID: "111c5349-528d-4856-a3e4-1277e3b10889"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:41:10 crc kubenswrapper[4631]: I1128 13:41:10.745506 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/111c5349-528d-4856-a3e4-1277e3b10889-horizon-tls-certs" (OuterVolumeSpecName: "horizon-tls-certs") pod "111c5349-528d-4856-a3e4-1277e3b10889" (UID: "111c5349-528d-4856-a3e4-1277e3b10889"). InnerVolumeSpecName "horizon-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:41:10 crc kubenswrapper[4631]: I1128 13:41:10.763841 4631 reconciler_common.go:293] "Volume detached for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/111c5349-528d-4856-a3e4-1277e3b10889-horizon-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 28 13:41:10 crc kubenswrapper[4631]: I1128 13:41:10.763879 4631 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/111c5349-528d-4856-a3e4-1277e3b10889-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 13:41:10 crc kubenswrapper[4631]: I1128 13:41:10.763889 4631 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/111c5349-528d-4856-a3e4-1277e3b10889-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Nov 28 13:41:10 crc kubenswrapper[4631]: I1128 13:41:10.763899 4631 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/111c5349-528d-4856-a3e4-1277e3b10889-logs\") on node \"crc\" DevicePath \"\"" Nov 28 13:41:10 crc kubenswrapper[4631]: I1128 13:41:10.763909 4631 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/111c5349-528d-4856-a3e4-1277e3b10889-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 13:41:10 crc kubenswrapper[4631]: I1128 13:41:10.763918 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k8tns\" (UniqueName: \"kubernetes.io/projected/111c5349-528d-4856-a3e4-1277e3b10889-kube-api-access-k8tns\") on node \"crc\" DevicePath \"\"" Nov 28 13:41:10 crc kubenswrapper[4631]: I1128 13:41:10.763932 4631 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/111c5349-528d-4856-a3e4-1277e3b10889-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 13:41:11 crc kubenswrapper[4631]: I1128 13:41:11.027234 4631 generic.go:334] "Generic (PLEG): container finished" podID="111c5349-528d-4856-a3e4-1277e3b10889" containerID="7bec69392b7f0e2c28b3311f3ec81e5bb4480735e00154f9e67c85a33543bc29" exitCode=137 Nov 28 13:41:11 crc kubenswrapper[4631]: I1128 13:41:11.027351 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-64f8fdf97b-jn5mg" Nov 28 13:41:11 crc kubenswrapper[4631]: I1128 13:41:11.027415 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-64f8fdf97b-jn5mg" event={"ID":"111c5349-528d-4856-a3e4-1277e3b10889","Type":"ContainerDied","Data":"7bec69392b7f0e2c28b3311f3ec81e5bb4480735e00154f9e67c85a33543bc29"} Nov 28 13:41:11 crc kubenswrapper[4631]: I1128 13:41:11.027448 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-64f8fdf97b-jn5mg" event={"ID":"111c5349-528d-4856-a3e4-1277e3b10889","Type":"ContainerDied","Data":"23b889f2631cf8acb4cb5fb4d20d550abb6dce9afa18c85b7c3660b3d4a38d3c"} Nov 28 13:41:11 crc kubenswrapper[4631]: I1128 13:41:11.027468 4631 scope.go:117] "RemoveContainer" containerID="3b826a5556ea3b27f6357d4cc74220f4bce293db6dbcb49fa5131013bef0e517" Nov 28 13:41:11 crc kubenswrapper[4631]: I1128 13:41:11.165899 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-64f8fdf97b-jn5mg"] Nov 28 13:41:11 crc kubenswrapper[4631]: I1128 13:41:11.184618 4631 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-64f8fdf97b-jn5mg"] Nov 28 13:41:11 crc kubenswrapper[4631]: I1128 13:41:11.251503 4631 scope.go:117] "RemoveContainer" containerID="7bec69392b7f0e2c28b3311f3ec81e5bb4480735e00154f9e67c85a33543bc29" Nov 28 13:41:11 crc kubenswrapper[4631]: I1128 13:41:11.294775 4631 scope.go:117] "RemoveContainer" containerID="3b826a5556ea3b27f6357d4cc74220f4bce293db6dbcb49fa5131013bef0e517" Nov 28 13:41:11 crc kubenswrapper[4631]: E1128 13:41:11.295382 4631 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3b826a5556ea3b27f6357d4cc74220f4bce293db6dbcb49fa5131013bef0e517\": container with ID starting with 3b826a5556ea3b27f6357d4cc74220f4bce293db6dbcb49fa5131013bef0e517 not found: ID does not exist" containerID="3b826a5556ea3b27f6357d4cc74220f4bce293db6dbcb49fa5131013bef0e517" Nov 28 13:41:11 crc kubenswrapper[4631]: I1128 13:41:11.295440 4631 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3b826a5556ea3b27f6357d4cc74220f4bce293db6dbcb49fa5131013bef0e517"} err="failed to get container status \"3b826a5556ea3b27f6357d4cc74220f4bce293db6dbcb49fa5131013bef0e517\": rpc error: code = NotFound desc = could not find container \"3b826a5556ea3b27f6357d4cc74220f4bce293db6dbcb49fa5131013bef0e517\": container with ID starting with 3b826a5556ea3b27f6357d4cc74220f4bce293db6dbcb49fa5131013bef0e517 not found: ID does not exist" Nov 28 13:41:11 crc kubenswrapper[4631]: I1128 13:41:11.295478 4631 scope.go:117] "RemoveContainer" containerID="7bec69392b7f0e2c28b3311f3ec81e5bb4480735e00154f9e67c85a33543bc29" Nov 28 13:41:11 crc kubenswrapper[4631]: E1128 13:41:11.296428 4631 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7bec69392b7f0e2c28b3311f3ec81e5bb4480735e00154f9e67c85a33543bc29\": container with ID starting with 7bec69392b7f0e2c28b3311f3ec81e5bb4480735e00154f9e67c85a33543bc29 not found: ID does not exist" containerID="7bec69392b7f0e2c28b3311f3ec81e5bb4480735e00154f9e67c85a33543bc29" Nov 28 13:41:11 crc kubenswrapper[4631]: I1128 13:41:11.296478 4631 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7bec69392b7f0e2c28b3311f3ec81e5bb4480735e00154f9e67c85a33543bc29"} err="failed to get container status \"7bec69392b7f0e2c28b3311f3ec81e5bb4480735e00154f9e67c85a33543bc29\": rpc error: code = NotFound desc = could not find container \"7bec69392b7f0e2c28b3311f3ec81e5bb4480735e00154f9e67c85a33543bc29\": container with ID starting with 7bec69392b7f0e2c28b3311f3ec81e5bb4480735e00154f9e67c85a33543bc29 not found: ID does not exist" Nov 28 13:41:11 crc kubenswrapper[4631]: I1128 13:41:11.524011 4631 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="111c5349-528d-4856-a3e4-1277e3b10889" path="/var/lib/kubelet/pods/111c5349-528d-4856-a3e4-1277e3b10889/volumes" Nov 28 13:41:13 crc kubenswrapper[4631]: I1128 13:41:13.352884 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Nov 28 13:41:13 crc kubenswrapper[4631]: I1128 13:41:13.532916 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Nov 28 13:41:15 crc kubenswrapper[4631]: I1128 13:41:15.331079 4631 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Nov 28 13:41:15 crc kubenswrapper[4631]: I1128 13:41:15.331546 4631 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Nov 28 13:41:16 crc kubenswrapper[4631]: I1128 13:41:16.355184 4631 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="fb5b797e-a328-4f2e-a239-333213537f2a" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.194:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 28 13:41:16 crc kubenswrapper[4631]: I1128 13:41:16.355457 4631 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="fb5b797e-a328-4f2e-a239-333213537f2a" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.194:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 28 13:41:18 crc kubenswrapper[4631]: I1128 13:41:18.505476 4631 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 28 13:41:18 crc kubenswrapper[4631]: I1128 13:41:18.506022 4631 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 28 13:41:18 crc kubenswrapper[4631]: I1128 13:41:18.532965 4631 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Nov 28 13:41:18 crc kubenswrapper[4631]: I1128 13:41:18.569435 4631 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Nov 28 13:41:19 crc kubenswrapper[4631]: I1128 13:41:19.153570 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Nov 28 13:41:19 crc kubenswrapper[4631]: E1128 13:41:19.301424 4631 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod111c5349_528d_4856_a3e4_1277e3b10889.slice/crio-3b826a5556ea3b27f6357d4cc74220f4bce293db6dbcb49fa5131013bef0e517.scope\": RecentStats: unable to find data in memory cache]" Nov 28 13:41:19 crc kubenswrapper[4631]: I1128 13:41:19.589583 4631 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="44c10c63-05fe-47e8-acc6-c927c9c3db7e" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.196:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 28 13:41:19 crc kubenswrapper[4631]: I1128 13:41:19.590000 4631 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="44c10c63-05fe-47e8-acc6-c927c9c3db7e" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.196:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 28 13:41:19 crc kubenswrapper[4631]: I1128 13:41:19.635274 4631 patch_prober.go:28] interesting pod/machine-config-daemon-47sxw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 28 13:41:19 crc kubenswrapper[4631]: I1128 13:41:19.635383 4631 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" podUID="8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 28 13:41:19 crc kubenswrapper[4631]: I1128 13:41:19.635457 4631 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" Nov 28 13:41:19 crc kubenswrapper[4631]: I1128 13:41:19.636404 4631 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"ae892f616f5c89a5af75dda3b0799a9e1ba33e9ae4579727d53b425382bdbdc2"} pod="openshift-machine-config-operator/machine-config-daemon-47sxw" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 28 13:41:19 crc kubenswrapper[4631]: I1128 13:41:19.636484 4631 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" podUID="8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1" containerName="machine-config-daemon" containerID="cri-o://ae892f616f5c89a5af75dda3b0799a9e1ba33e9ae4579727d53b425382bdbdc2" gracePeriod=600 Nov 28 13:41:20 crc kubenswrapper[4631]: I1128 13:41:20.123317 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" event={"ID":"8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1","Type":"ContainerDied","Data":"ae892f616f5c89a5af75dda3b0799a9e1ba33e9ae4579727d53b425382bdbdc2"} Nov 28 13:41:20 crc kubenswrapper[4631]: I1128 13:41:20.124205 4631 scope.go:117] "RemoveContainer" containerID="a4b336d735e84003b86dbce21368f3f29fb0ded3e7a12a4562271fa2767a4ae8" Nov 28 13:41:20 crc kubenswrapper[4631]: I1128 13:41:20.124383 4631 generic.go:334] "Generic (PLEG): container finished" podID="8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1" containerID="ae892f616f5c89a5af75dda3b0799a9e1ba33e9ae4579727d53b425382bdbdc2" exitCode=0 Nov 28 13:41:20 crc kubenswrapper[4631]: I1128 13:41:20.124500 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" event={"ID":"8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1","Type":"ContainerStarted","Data":"05e6feabd7664ef6344f7bae4443e46fb2b0ab0814dcbdcb00623d5d82759c9c"} Nov 28 13:41:23 crc kubenswrapper[4631]: I1128 13:41:23.099616 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 28 13:41:23 crc kubenswrapper[4631]: I1128 13:41:23.153837 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e8a60bc-6e73-499a-959a-c0f104014641-combined-ca-bundle\") pod \"5e8a60bc-6e73-499a-959a-c0f104014641\" (UID: \"5e8a60bc-6e73-499a-959a-c0f104014641\") " Nov 28 13:41:23 crc kubenswrapper[4631]: I1128 13:41:23.153895 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f2fhf\" (UniqueName: \"kubernetes.io/projected/5e8a60bc-6e73-499a-959a-c0f104014641-kube-api-access-f2fhf\") pod \"5e8a60bc-6e73-499a-959a-c0f104014641\" (UID: \"5e8a60bc-6e73-499a-959a-c0f104014641\") " Nov 28 13:41:23 crc kubenswrapper[4631]: I1128 13:41:23.154043 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5e8a60bc-6e73-499a-959a-c0f104014641-config-data\") pod \"5e8a60bc-6e73-499a-959a-c0f104014641\" (UID: \"5e8a60bc-6e73-499a-959a-c0f104014641\") " Nov 28 13:41:23 crc kubenswrapper[4631]: I1128 13:41:23.162205 4631 generic.go:334] "Generic (PLEG): container finished" podID="5e8a60bc-6e73-499a-959a-c0f104014641" containerID="2955d8dedee901ef27b5cf6a4ef41a86cc3d2d1ea69ec7f71f29c857428aaadc" exitCode=137 Nov 28 13:41:23 crc kubenswrapper[4631]: I1128 13:41:23.162275 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"5e8a60bc-6e73-499a-959a-c0f104014641","Type":"ContainerDied","Data":"2955d8dedee901ef27b5cf6a4ef41a86cc3d2d1ea69ec7f71f29c857428aaadc"} Nov 28 13:41:23 crc kubenswrapper[4631]: I1128 13:41:23.162402 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"5e8a60bc-6e73-499a-959a-c0f104014641","Type":"ContainerDied","Data":"131605dec7c38ec98c2deaa0be880784fabc50133f77ad682d68346450009164"} Nov 28 13:41:23 crc kubenswrapper[4631]: I1128 13:41:23.162434 4631 scope.go:117] "RemoveContainer" containerID="2955d8dedee901ef27b5cf6a4ef41a86cc3d2d1ea69ec7f71f29c857428aaadc" Nov 28 13:41:23 crc kubenswrapper[4631]: I1128 13:41:23.162690 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 28 13:41:23 crc kubenswrapper[4631]: I1128 13:41:23.162959 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5e8a60bc-6e73-499a-959a-c0f104014641-kube-api-access-f2fhf" (OuterVolumeSpecName: "kube-api-access-f2fhf") pod "5e8a60bc-6e73-499a-959a-c0f104014641" (UID: "5e8a60bc-6e73-499a-959a-c0f104014641"). InnerVolumeSpecName "kube-api-access-f2fhf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:41:23 crc kubenswrapper[4631]: I1128 13:41:23.187171 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5e8a60bc-6e73-499a-959a-c0f104014641-config-data" (OuterVolumeSpecName: "config-data") pod "5e8a60bc-6e73-499a-959a-c0f104014641" (UID: "5e8a60bc-6e73-499a-959a-c0f104014641"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:41:23 crc kubenswrapper[4631]: I1128 13:41:23.203765 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5e8a60bc-6e73-499a-959a-c0f104014641-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5e8a60bc-6e73-499a-959a-c0f104014641" (UID: "5e8a60bc-6e73-499a-959a-c0f104014641"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:41:23 crc kubenswrapper[4631]: I1128 13:41:23.257470 4631 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e8a60bc-6e73-499a-959a-c0f104014641-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 13:41:23 crc kubenswrapper[4631]: I1128 13:41:23.257509 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f2fhf\" (UniqueName: \"kubernetes.io/projected/5e8a60bc-6e73-499a-959a-c0f104014641-kube-api-access-f2fhf\") on node \"crc\" DevicePath \"\"" Nov 28 13:41:23 crc kubenswrapper[4631]: I1128 13:41:23.257527 4631 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5e8a60bc-6e73-499a-959a-c0f104014641-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 13:41:23 crc kubenswrapper[4631]: I1128 13:41:23.261044 4631 scope.go:117] "RemoveContainer" containerID="2955d8dedee901ef27b5cf6a4ef41a86cc3d2d1ea69ec7f71f29c857428aaadc" Nov 28 13:41:23 crc kubenswrapper[4631]: E1128 13:41:23.261676 4631 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2955d8dedee901ef27b5cf6a4ef41a86cc3d2d1ea69ec7f71f29c857428aaadc\": container with ID starting with 2955d8dedee901ef27b5cf6a4ef41a86cc3d2d1ea69ec7f71f29c857428aaadc not found: ID does not exist" containerID="2955d8dedee901ef27b5cf6a4ef41a86cc3d2d1ea69ec7f71f29c857428aaadc" Nov 28 13:41:23 crc kubenswrapper[4631]: I1128 13:41:23.261748 4631 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2955d8dedee901ef27b5cf6a4ef41a86cc3d2d1ea69ec7f71f29c857428aaadc"} err="failed to get container status \"2955d8dedee901ef27b5cf6a4ef41a86cc3d2d1ea69ec7f71f29c857428aaadc\": rpc error: code = NotFound desc = could not find container \"2955d8dedee901ef27b5cf6a4ef41a86cc3d2d1ea69ec7f71f29c857428aaadc\": container with ID starting with 2955d8dedee901ef27b5cf6a4ef41a86cc3d2d1ea69ec7f71f29c857428aaadc not found: ID does not exist" Nov 28 13:41:23 crc kubenswrapper[4631]: I1128 13:41:23.546274 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 28 13:41:23 crc kubenswrapper[4631]: I1128 13:41:23.573601 4631 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 28 13:41:23 crc kubenswrapper[4631]: I1128 13:41:23.590281 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 28 13:41:23 crc kubenswrapper[4631]: E1128 13:41:23.590812 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="111c5349-528d-4856-a3e4-1277e3b10889" containerName="horizon" Nov 28 13:41:23 crc kubenswrapper[4631]: I1128 13:41:23.590827 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="111c5349-528d-4856-a3e4-1277e3b10889" containerName="horizon" Nov 28 13:41:23 crc kubenswrapper[4631]: E1128 13:41:23.590849 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5e8a60bc-6e73-499a-959a-c0f104014641" containerName="nova-cell1-novncproxy-novncproxy" Nov 28 13:41:23 crc kubenswrapper[4631]: I1128 13:41:23.590858 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="5e8a60bc-6e73-499a-959a-c0f104014641" containerName="nova-cell1-novncproxy-novncproxy" Nov 28 13:41:23 crc kubenswrapper[4631]: E1128 13:41:23.590868 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="111c5349-528d-4856-a3e4-1277e3b10889" containerName="horizon-log" Nov 28 13:41:23 crc kubenswrapper[4631]: I1128 13:41:23.590875 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="111c5349-528d-4856-a3e4-1277e3b10889" containerName="horizon-log" Nov 28 13:41:23 crc kubenswrapper[4631]: E1128 13:41:23.590888 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="111c5349-528d-4856-a3e4-1277e3b10889" containerName="horizon" Nov 28 13:41:23 crc kubenswrapper[4631]: I1128 13:41:23.590895 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="111c5349-528d-4856-a3e4-1277e3b10889" containerName="horizon" Nov 28 13:41:23 crc kubenswrapper[4631]: E1128 13:41:23.590910 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="111c5349-528d-4856-a3e4-1277e3b10889" containerName="horizon" Nov 28 13:41:23 crc kubenswrapper[4631]: I1128 13:41:23.590916 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="111c5349-528d-4856-a3e4-1277e3b10889" containerName="horizon" Nov 28 13:41:23 crc kubenswrapper[4631]: I1128 13:41:23.591139 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="5e8a60bc-6e73-499a-959a-c0f104014641" containerName="nova-cell1-novncproxy-novncproxy" Nov 28 13:41:23 crc kubenswrapper[4631]: I1128 13:41:23.591158 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="111c5349-528d-4856-a3e4-1277e3b10889" containerName="horizon" Nov 28 13:41:23 crc kubenswrapper[4631]: I1128 13:41:23.591170 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="111c5349-528d-4856-a3e4-1277e3b10889" containerName="horizon" Nov 28 13:41:23 crc kubenswrapper[4631]: I1128 13:41:23.591179 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="111c5349-528d-4856-a3e4-1277e3b10889" containerName="horizon-log" Nov 28 13:41:23 crc kubenswrapper[4631]: I1128 13:41:23.591190 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="111c5349-528d-4856-a3e4-1277e3b10889" containerName="horizon" Nov 28 13:41:23 crc kubenswrapper[4631]: I1128 13:41:23.591932 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 28 13:41:23 crc kubenswrapper[4631]: I1128 13:41:23.598896 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Nov 28 13:41:23 crc kubenswrapper[4631]: I1128 13:41:23.603130 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Nov 28 13:41:23 crc kubenswrapper[4631]: I1128 13:41:23.603458 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Nov 28 13:41:23 crc kubenswrapper[4631]: I1128 13:41:23.631254 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 28 13:41:23 crc kubenswrapper[4631]: I1128 13:41:23.666984 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9c3a5a08-9590-4249-be26-8224d51da976-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"9c3a5a08-9590-4249-be26-8224d51da976\") " pod="openstack/nova-cell1-novncproxy-0" Nov 28 13:41:23 crc kubenswrapper[4631]: I1128 13:41:23.667123 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w6vcw\" (UniqueName: \"kubernetes.io/projected/9c3a5a08-9590-4249-be26-8224d51da976-kube-api-access-w6vcw\") pod \"nova-cell1-novncproxy-0\" (UID: \"9c3a5a08-9590-4249-be26-8224d51da976\") " pod="openstack/nova-cell1-novncproxy-0" Nov 28 13:41:23 crc kubenswrapper[4631]: I1128 13:41:23.667228 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/9c3a5a08-9590-4249-be26-8224d51da976-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"9c3a5a08-9590-4249-be26-8224d51da976\") " pod="openstack/nova-cell1-novncproxy-0" Nov 28 13:41:23 crc kubenswrapper[4631]: I1128 13:41:23.668623 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/9c3a5a08-9590-4249-be26-8224d51da976-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"9c3a5a08-9590-4249-be26-8224d51da976\") " pod="openstack/nova-cell1-novncproxy-0" Nov 28 13:41:23 crc kubenswrapper[4631]: I1128 13:41:23.668699 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c3a5a08-9590-4249-be26-8224d51da976-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"9c3a5a08-9590-4249-be26-8224d51da976\") " pod="openstack/nova-cell1-novncproxy-0" Nov 28 13:41:23 crc kubenswrapper[4631]: I1128 13:41:23.771043 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9c3a5a08-9590-4249-be26-8224d51da976-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"9c3a5a08-9590-4249-be26-8224d51da976\") " pod="openstack/nova-cell1-novncproxy-0" Nov 28 13:41:23 crc kubenswrapper[4631]: I1128 13:41:23.771113 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w6vcw\" (UniqueName: \"kubernetes.io/projected/9c3a5a08-9590-4249-be26-8224d51da976-kube-api-access-w6vcw\") pod \"nova-cell1-novncproxy-0\" (UID: \"9c3a5a08-9590-4249-be26-8224d51da976\") " pod="openstack/nova-cell1-novncproxy-0" Nov 28 13:41:23 crc kubenswrapper[4631]: I1128 13:41:23.771147 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/9c3a5a08-9590-4249-be26-8224d51da976-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"9c3a5a08-9590-4249-be26-8224d51da976\") " pod="openstack/nova-cell1-novncproxy-0" Nov 28 13:41:23 crc kubenswrapper[4631]: I1128 13:41:23.771215 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/9c3a5a08-9590-4249-be26-8224d51da976-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"9c3a5a08-9590-4249-be26-8224d51da976\") " pod="openstack/nova-cell1-novncproxy-0" Nov 28 13:41:23 crc kubenswrapper[4631]: I1128 13:41:23.771265 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c3a5a08-9590-4249-be26-8224d51da976-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"9c3a5a08-9590-4249-be26-8224d51da976\") " pod="openstack/nova-cell1-novncproxy-0" Nov 28 13:41:23 crc kubenswrapper[4631]: I1128 13:41:23.779489 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/9c3a5a08-9590-4249-be26-8224d51da976-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"9c3a5a08-9590-4249-be26-8224d51da976\") " pod="openstack/nova-cell1-novncproxy-0" Nov 28 13:41:23 crc kubenswrapper[4631]: I1128 13:41:23.779569 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/9c3a5a08-9590-4249-be26-8224d51da976-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"9c3a5a08-9590-4249-be26-8224d51da976\") " pod="openstack/nova-cell1-novncproxy-0" Nov 28 13:41:23 crc kubenswrapper[4631]: I1128 13:41:23.780692 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9c3a5a08-9590-4249-be26-8224d51da976-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"9c3a5a08-9590-4249-be26-8224d51da976\") " pod="openstack/nova-cell1-novncproxy-0" Nov 28 13:41:23 crc kubenswrapper[4631]: I1128 13:41:23.782621 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c3a5a08-9590-4249-be26-8224d51da976-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"9c3a5a08-9590-4249-be26-8224d51da976\") " pod="openstack/nova-cell1-novncproxy-0" Nov 28 13:41:23 crc kubenswrapper[4631]: I1128 13:41:23.799717 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w6vcw\" (UniqueName: \"kubernetes.io/projected/9c3a5a08-9590-4249-be26-8224d51da976-kube-api-access-w6vcw\") pod \"nova-cell1-novncproxy-0\" (UID: \"9c3a5a08-9590-4249-be26-8224d51da976\") " pod="openstack/nova-cell1-novncproxy-0" Nov 28 13:41:23 crc kubenswrapper[4631]: I1128 13:41:23.937785 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 28 13:41:24 crc kubenswrapper[4631]: I1128 13:41:24.468729 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 28 13:41:25 crc kubenswrapper[4631]: I1128 13:41:25.194190 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"9c3a5a08-9590-4249-be26-8224d51da976","Type":"ContainerStarted","Data":"270fc2c722b847c03a4869e47c440270c6047bee05b8a2ba313ba0b33126c27c"} Nov 28 13:41:25 crc kubenswrapper[4631]: I1128 13:41:25.195042 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"9c3a5a08-9590-4249-be26-8224d51da976","Type":"ContainerStarted","Data":"1e76d62d756a700d577cc1d50e5ffef50def47c347944e2e3cd5b06cb18891f1"} Nov 28 13:41:25 crc kubenswrapper[4631]: I1128 13:41:25.230255 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.230231327 podStartE2EDuration="2.230231327s" podCreationTimestamp="2025-11-28 13:41:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 13:41:25.219914427 +0000 UTC m=+1242.027217771" watchObservedRunningTime="2025-11-28 13:41:25.230231327 +0000 UTC m=+1242.037534671" Nov 28 13:41:25 crc kubenswrapper[4631]: I1128 13:41:25.337274 4631 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Nov 28 13:41:25 crc kubenswrapper[4631]: I1128 13:41:25.341759 4631 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Nov 28 13:41:25 crc kubenswrapper[4631]: I1128 13:41:25.347088 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Nov 28 13:41:25 crc kubenswrapper[4631]: I1128 13:41:25.539397 4631 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5e8a60bc-6e73-499a-959a-c0f104014641" path="/var/lib/kubelet/pods/5e8a60bc-6e73-499a-959a-c0f104014641/volumes" Nov 28 13:41:26 crc kubenswrapper[4631]: I1128 13:41:26.220274 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Nov 28 13:41:28 crc kubenswrapper[4631]: I1128 13:41:28.519808 4631 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Nov 28 13:41:28 crc kubenswrapper[4631]: I1128 13:41:28.520890 4631 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Nov 28 13:41:28 crc kubenswrapper[4631]: I1128 13:41:28.521127 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Nov 28 13:41:28 crc kubenswrapper[4631]: I1128 13:41:28.532111 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Nov 28 13:41:28 crc kubenswrapper[4631]: I1128 13:41:28.939356 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Nov 28 13:41:29 crc kubenswrapper[4631]: I1128 13:41:29.255316 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Nov 28 13:41:29 crc kubenswrapper[4631]: I1128 13:41:29.262016 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Nov 28 13:41:29 crc kubenswrapper[4631]: I1128 13:41:29.547334 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-89c5cd4d5-qvl5b"] Nov 28 13:41:29 crc kubenswrapper[4631]: I1128 13:41:29.549272 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-89c5cd4d5-qvl5b" Nov 28 13:41:29 crc kubenswrapper[4631]: I1128 13:41:29.562633 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-89c5cd4d5-qvl5b"] Nov 28 13:41:29 crc kubenswrapper[4631]: I1128 13:41:29.626916 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/931de4b9-12a1-45bf-b69e-b0498c6c770f-dns-swift-storage-0\") pod \"dnsmasq-dns-89c5cd4d5-qvl5b\" (UID: \"931de4b9-12a1-45bf-b69e-b0498c6c770f\") " pod="openstack/dnsmasq-dns-89c5cd4d5-qvl5b" Nov 28 13:41:29 crc kubenswrapper[4631]: I1128 13:41:29.626995 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/931de4b9-12a1-45bf-b69e-b0498c6c770f-ovsdbserver-nb\") pod \"dnsmasq-dns-89c5cd4d5-qvl5b\" (UID: \"931de4b9-12a1-45bf-b69e-b0498c6c770f\") " pod="openstack/dnsmasq-dns-89c5cd4d5-qvl5b" Nov 28 13:41:29 crc kubenswrapper[4631]: I1128 13:41:29.627026 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/931de4b9-12a1-45bf-b69e-b0498c6c770f-dns-svc\") pod \"dnsmasq-dns-89c5cd4d5-qvl5b\" (UID: \"931de4b9-12a1-45bf-b69e-b0498c6c770f\") " pod="openstack/dnsmasq-dns-89c5cd4d5-qvl5b" Nov 28 13:41:29 crc kubenswrapper[4631]: I1128 13:41:29.627117 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/931de4b9-12a1-45bf-b69e-b0498c6c770f-ovsdbserver-sb\") pod \"dnsmasq-dns-89c5cd4d5-qvl5b\" (UID: \"931de4b9-12a1-45bf-b69e-b0498c6c770f\") " pod="openstack/dnsmasq-dns-89c5cd4d5-qvl5b" Nov 28 13:41:29 crc kubenswrapper[4631]: I1128 13:41:29.627150 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z6dbm\" (UniqueName: \"kubernetes.io/projected/931de4b9-12a1-45bf-b69e-b0498c6c770f-kube-api-access-z6dbm\") pod \"dnsmasq-dns-89c5cd4d5-qvl5b\" (UID: \"931de4b9-12a1-45bf-b69e-b0498c6c770f\") " pod="openstack/dnsmasq-dns-89c5cd4d5-qvl5b" Nov 28 13:41:29 crc kubenswrapper[4631]: I1128 13:41:29.627174 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/931de4b9-12a1-45bf-b69e-b0498c6c770f-config\") pod \"dnsmasq-dns-89c5cd4d5-qvl5b\" (UID: \"931de4b9-12a1-45bf-b69e-b0498c6c770f\") " pod="openstack/dnsmasq-dns-89c5cd4d5-qvl5b" Nov 28 13:41:29 crc kubenswrapper[4631]: E1128 13:41:29.687949 4631 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod111c5349_528d_4856_a3e4_1277e3b10889.slice/crio-3b826a5556ea3b27f6357d4cc74220f4bce293db6dbcb49fa5131013bef0e517.scope\": RecentStats: unable to find data in memory cache]" Nov 28 13:41:29 crc kubenswrapper[4631]: I1128 13:41:29.729538 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z6dbm\" (UniqueName: \"kubernetes.io/projected/931de4b9-12a1-45bf-b69e-b0498c6c770f-kube-api-access-z6dbm\") pod \"dnsmasq-dns-89c5cd4d5-qvl5b\" (UID: \"931de4b9-12a1-45bf-b69e-b0498c6c770f\") " pod="openstack/dnsmasq-dns-89c5cd4d5-qvl5b" Nov 28 13:41:29 crc kubenswrapper[4631]: I1128 13:41:29.730183 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/931de4b9-12a1-45bf-b69e-b0498c6c770f-config\") pod \"dnsmasq-dns-89c5cd4d5-qvl5b\" (UID: \"931de4b9-12a1-45bf-b69e-b0498c6c770f\") " pod="openstack/dnsmasq-dns-89c5cd4d5-qvl5b" Nov 28 13:41:29 crc kubenswrapper[4631]: I1128 13:41:29.730366 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/931de4b9-12a1-45bf-b69e-b0498c6c770f-dns-swift-storage-0\") pod \"dnsmasq-dns-89c5cd4d5-qvl5b\" (UID: \"931de4b9-12a1-45bf-b69e-b0498c6c770f\") " pod="openstack/dnsmasq-dns-89c5cd4d5-qvl5b" Nov 28 13:41:29 crc kubenswrapper[4631]: I1128 13:41:29.730485 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/931de4b9-12a1-45bf-b69e-b0498c6c770f-ovsdbserver-nb\") pod \"dnsmasq-dns-89c5cd4d5-qvl5b\" (UID: \"931de4b9-12a1-45bf-b69e-b0498c6c770f\") " pod="openstack/dnsmasq-dns-89c5cd4d5-qvl5b" Nov 28 13:41:29 crc kubenswrapper[4631]: I1128 13:41:29.730572 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/931de4b9-12a1-45bf-b69e-b0498c6c770f-dns-svc\") pod \"dnsmasq-dns-89c5cd4d5-qvl5b\" (UID: \"931de4b9-12a1-45bf-b69e-b0498c6c770f\") " pod="openstack/dnsmasq-dns-89c5cd4d5-qvl5b" Nov 28 13:41:29 crc kubenswrapper[4631]: I1128 13:41:29.730706 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/931de4b9-12a1-45bf-b69e-b0498c6c770f-ovsdbserver-sb\") pod \"dnsmasq-dns-89c5cd4d5-qvl5b\" (UID: \"931de4b9-12a1-45bf-b69e-b0498c6c770f\") " pod="openstack/dnsmasq-dns-89c5cd4d5-qvl5b" Nov 28 13:41:29 crc kubenswrapper[4631]: I1128 13:41:29.731406 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/931de4b9-12a1-45bf-b69e-b0498c6c770f-config\") pod \"dnsmasq-dns-89c5cd4d5-qvl5b\" (UID: \"931de4b9-12a1-45bf-b69e-b0498c6c770f\") " pod="openstack/dnsmasq-dns-89c5cd4d5-qvl5b" Nov 28 13:41:29 crc kubenswrapper[4631]: I1128 13:41:29.731735 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/931de4b9-12a1-45bf-b69e-b0498c6c770f-ovsdbserver-sb\") pod \"dnsmasq-dns-89c5cd4d5-qvl5b\" (UID: \"931de4b9-12a1-45bf-b69e-b0498c6c770f\") " pod="openstack/dnsmasq-dns-89c5cd4d5-qvl5b" Nov 28 13:41:29 crc kubenswrapper[4631]: I1128 13:41:29.732702 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/931de4b9-12a1-45bf-b69e-b0498c6c770f-ovsdbserver-nb\") pod \"dnsmasq-dns-89c5cd4d5-qvl5b\" (UID: \"931de4b9-12a1-45bf-b69e-b0498c6c770f\") " pod="openstack/dnsmasq-dns-89c5cd4d5-qvl5b" Nov 28 13:41:29 crc kubenswrapper[4631]: I1128 13:41:29.733789 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/931de4b9-12a1-45bf-b69e-b0498c6c770f-dns-svc\") pod \"dnsmasq-dns-89c5cd4d5-qvl5b\" (UID: \"931de4b9-12a1-45bf-b69e-b0498c6c770f\") " pod="openstack/dnsmasq-dns-89c5cd4d5-qvl5b" Nov 28 13:41:29 crc kubenswrapper[4631]: I1128 13:41:29.736666 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/931de4b9-12a1-45bf-b69e-b0498c6c770f-dns-swift-storage-0\") pod \"dnsmasq-dns-89c5cd4d5-qvl5b\" (UID: \"931de4b9-12a1-45bf-b69e-b0498c6c770f\") " pod="openstack/dnsmasq-dns-89c5cd4d5-qvl5b" Nov 28 13:41:29 crc kubenswrapper[4631]: I1128 13:41:29.755482 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z6dbm\" (UniqueName: \"kubernetes.io/projected/931de4b9-12a1-45bf-b69e-b0498c6c770f-kube-api-access-z6dbm\") pod \"dnsmasq-dns-89c5cd4d5-qvl5b\" (UID: \"931de4b9-12a1-45bf-b69e-b0498c6c770f\") " pod="openstack/dnsmasq-dns-89c5cd4d5-qvl5b" Nov 28 13:41:29 crc kubenswrapper[4631]: I1128 13:41:29.881673 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-89c5cd4d5-qvl5b" Nov 28 13:41:30 crc kubenswrapper[4631]: I1128 13:41:30.403210 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-89c5cd4d5-qvl5b"] Nov 28 13:41:31 crc kubenswrapper[4631]: I1128 13:41:31.278368 4631 generic.go:334] "Generic (PLEG): container finished" podID="931de4b9-12a1-45bf-b69e-b0498c6c770f" containerID="584175cb653c5b5b9801311291b1034c9cab49768b9f11cbf85a176a8b325e03" exitCode=0 Nov 28 13:41:31 crc kubenswrapper[4631]: I1128 13:41:31.279567 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-89c5cd4d5-qvl5b" event={"ID":"931de4b9-12a1-45bf-b69e-b0498c6c770f","Type":"ContainerDied","Data":"584175cb653c5b5b9801311291b1034c9cab49768b9f11cbf85a176a8b325e03"} Nov 28 13:41:31 crc kubenswrapper[4631]: I1128 13:41:31.279647 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-89c5cd4d5-qvl5b" event={"ID":"931de4b9-12a1-45bf-b69e-b0498c6c770f","Type":"ContainerStarted","Data":"2e70e7cb3fcb57c4c650696f213c0b32fe875a1662c34d60be6bbd8b0a354173"} Nov 28 13:41:32 crc kubenswrapper[4631]: I1128 13:41:32.014723 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 28 13:41:32 crc kubenswrapper[4631]: I1128 13:41:32.015792 4631 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a00de8c7-a886-4851-b7cd-c8784a3c6bd9" containerName="ceilometer-central-agent" containerID="cri-o://3e70f5937aa64c1f6b17c239bba41cef0aa54ddf458ce5b8cafeccbb3e212581" gracePeriod=30 Nov 28 13:41:32 crc kubenswrapper[4631]: I1128 13:41:32.016575 4631 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a00de8c7-a886-4851-b7cd-c8784a3c6bd9" containerName="proxy-httpd" containerID="cri-o://d1a7db8ddb7853f6005959d4900a1ada50c4962ce089b37695da99fd552f8573" gracePeriod=30 Nov 28 13:41:32 crc kubenswrapper[4631]: I1128 13:41:32.016646 4631 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a00de8c7-a886-4851-b7cd-c8784a3c6bd9" containerName="sg-core" containerID="cri-o://11e2e2d4a8aff3ef637fdd4f4485a2cb4c252744fa8b2147478739303369e0e2" gracePeriod=30 Nov 28 13:41:32 crc kubenswrapper[4631]: I1128 13:41:32.016699 4631 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a00de8c7-a886-4851-b7cd-c8784a3c6bd9" containerName="ceilometer-notification-agent" containerID="cri-o://0f6318a7db020712586bb2e56cdd6f8046b9a725a41e45e0a31daac51781c62c" gracePeriod=30 Nov 28 13:41:32 crc kubenswrapper[4631]: I1128 13:41:32.294001 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-89c5cd4d5-qvl5b" event={"ID":"931de4b9-12a1-45bf-b69e-b0498c6c770f","Type":"ContainerStarted","Data":"4db6ded4946f647032503312a75a1009a07c3e44165a5c47a10ae0217d33bec5"} Nov 28 13:41:32 crc kubenswrapper[4631]: I1128 13:41:32.294903 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-89c5cd4d5-qvl5b" Nov 28 13:41:32 crc kubenswrapper[4631]: I1128 13:41:32.298055 4631 generic.go:334] "Generic (PLEG): container finished" podID="a00de8c7-a886-4851-b7cd-c8784a3c6bd9" containerID="d1a7db8ddb7853f6005959d4900a1ada50c4962ce089b37695da99fd552f8573" exitCode=0 Nov 28 13:41:32 crc kubenswrapper[4631]: I1128 13:41:32.298087 4631 generic.go:334] "Generic (PLEG): container finished" podID="a00de8c7-a886-4851-b7cd-c8784a3c6bd9" containerID="11e2e2d4a8aff3ef637fdd4f4485a2cb4c252744fa8b2147478739303369e0e2" exitCode=2 Nov 28 13:41:32 crc kubenswrapper[4631]: I1128 13:41:32.298111 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a00de8c7-a886-4851-b7cd-c8784a3c6bd9","Type":"ContainerDied","Data":"d1a7db8ddb7853f6005959d4900a1ada50c4962ce089b37695da99fd552f8573"} Nov 28 13:41:32 crc kubenswrapper[4631]: I1128 13:41:32.298140 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a00de8c7-a886-4851-b7cd-c8784a3c6bd9","Type":"ContainerDied","Data":"11e2e2d4a8aff3ef637fdd4f4485a2cb4c252744fa8b2147478739303369e0e2"} Nov 28 13:41:32 crc kubenswrapper[4631]: I1128 13:41:32.316773 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-89c5cd4d5-qvl5b" podStartSLOduration=3.316748079 podStartE2EDuration="3.316748079s" podCreationTimestamp="2025-11-28 13:41:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 13:41:32.316523604 +0000 UTC m=+1249.123826968" watchObservedRunningTime="2025-11-28 13:41:32.316748079 +0000 UTC m=+1249.124051423" Nov 28 13:41:32 crc kubenswrapper[4631]: I1128 13:41:32.747722 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 28 13:41:32 crc kubenswrapper[4631]: I1128 13:41:32.748422 4631 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="44c10c63-05fe-47e8-acc6-c927c9c3db7e" containerName="nova-api-log" containerID="cri-o://4103b4cf6b619a45021786723e84b8ad4e2fc43a3f9eab4611393b14c989a39e" gracePeriod=30 Nov 28 13:41:32 crc kubenswrapper[4631]: I1128 13:41:32.748785 4631 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="44c10c63-05fe-47e8-acc6-c927c9c3db7e" containerName="nova-api-api" containerID="cri-o://cf453b65de3fb82eee03c5027c805c5311f5c1efd14c3e31bbfb557b4dcd04a8" gracePeriod=30 Nov 28 13:41:33 crc kubenswrapper[4631]: I1128 13:41:33.310431 4631 generic.go:334] "Generic (PLEG): container finished" podID="44c10c63-05fe-47e8-acc6-c927c9c3db7e" containerID="4103b4cf6b619a45021786723e84b8ad4e2fc43a3f9eab4611393b14c989a39e" exitCode=143 Nov 28 13:41:33 crc kubenswrapper[4631]: I1128 13:41:33.310484 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"44c10c63-05fe-47e8-acc6-c927c9c3db7e","Type":"ContainerDied","Data":"4103b4cf6b619a45021786723e84b8ad4e2fc43a3f9eab4611393b14c989a39e"} Nov 28 13:41:33 crc kubenswrapper[4631]: I1128 13:41:33.316586 4631 generic.go:334] "Generic (PLEG): container finished" podID="a00de8c7-a886-4851-b7cd-c8784a3c6bd9" containerID="3e70f5937aa64c1f6b17c239bba41cef0aa54ddf458ce5b8cafeccbb3e212581" exitCode=0 Nov 28 13:41:33 crc kubenswrapper[4631]: I1128 13:41:33.316659 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a00de8c7-a886-4851-b7cd-c8784a3c6bd9","Type":"ContainerDied","Data":"3e70f5937aa64c1f6b17c239bba41cef0aa54ddf458ce5b8cafeccbb3e212581"} Nov 28 13:41:33 crc kubenswrapper[4631]: I1128 13:41:33.938937 4631 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Nov 28 13:41:33 crc kubenswrapper[4631]: I1128 13:41:33.978253 4631 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Nov 28 13:41:34 crc kubenswrapper[4631]: I1128 13:41:34.360042 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Nov 28 13:41:34 crc kubenswrapper[4631]: I1128 13:41:34.625012 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-24v2l"] Nov 28 13:41:34 crc kubenswrapper[4631]: I1128 13:41:34.626503 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-24v2l" Nov 28 13:41:34 crc kubenswrapper[4631]: I1128 13:41:34.630169 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Nov 28 13:41:34 crc kubenswrapper[4631]: I1128 13:41:34.632893 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Nov 28 13:41:34 crc kubenswrapper[4631]: I1128 13:41:34.641646 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-24v2l"] Nov 28 13:41:34 crc kubenswrapper[4631]: I1128 13:41:34.758592 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0339440e-8081-4f68-9c6f-b62e5be5fc27-scripts\") pod \"nova-cell1-cell-mapping-24v2l\" (UID: \"0339440e-8081-4f68-9c6f-b62e5be5fc27\") " pod="openstack/nova-cell1-cell-mapping-24v2l" Nov 28 13:41:34 crc kubenswrapper[4631]: I1128 13:41:34.758672 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0339440e-8081-4f68-9c6f-b62e5be5fc27-config-data\") pod \"nova-cell1-cell-mapping-24v2l\" (UID: \"0339440e-8081-4f68-9c6f-b62e5be5fc27\") " pod="openstack/nova-cell1-cell-mapping-24v2l" Nov 28 13:41:34 crc kubenswrapper[4631]: I1128 13:41:34.758717 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0339440e-8081-4f68-9c6f-b62e5be5fc27-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-24v2l\" (UID: \"0339440e-8081-4f68-9c6f-b62e5be5fc27\") " pod="openstack/nova-cell1-cell-mapping-24v2l" Nov 28 13:41:34 crc kubenswrapper[4631]: I1128 13:41:34.758751 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t84lv\" (UniqueName: \"kubernetes.io/projected/0339440e-8081-4f68-9c6f-b62e5be5fc27-kube-api-access-t84lv\") pod \"nova-cell1-cell-mapping-24v2l\" (UID: \"0339440e-8081-4f68-9c6f-b62e5be5fc27\") " pod="openstack/nova-cell1-cell-mapping-24v2l" Nov 28 13:41:34 crc kubenswrapper[4631]: I1128 13:41:34.861470 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0339440e-8081-4f68-9c6f-b62e5be5fc27-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-24v2l\" (UID: \"0339440e-8081-4f68-9c6f-b62e5be5fc27\") " pod="openstack/nova-cell1-cell-mapping-24v2l" Nov 28 13:41:34 crc kubenswrapper[4631]: I1128 13:41:34.861516 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t84lv\" (UniqueName: \"kubernetes.io/projected/0339440e-8081-4f68-9c6f-b62e5be5fc27-kube-api-access-t84lv\") pod \"nova-cell1-cell-mapping-24v2l\" (UID: \"0339440e-8081-4f68-9c6f-b62e5be5fc27\") " pod="openstack/nova-cell1-cell-mapping-24v2l" Nov 28 13:41:34 crc kubenswrapper[4631]: I1128 13:41:34.861676 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0339440e-8081-4f68-9c6f-b62e5be5fc27-scripts\") pod \"nova-cell1-cell-mapping-24v2l\" (UID: \"0339440e-8081-4f68-9c6f-b62e5be5fc27\") " pod="openstack/nova-cell1-cell-mapping-24v2l" Nov 28 13:41:34 crc kubenswrapper[4631]: I1128 13:41:34.861704 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0339440e-8081-4f68-9c6f-b62e5be5fc27-config-data\") pod \"nova-cell1-cell-mapping-24v2l\" (UID: \"0339440e-8081-4f68-9c6f-b62e5be5fc27\") " pod="openstack/nova-cell1-cell-mapping-24v2l" Nov 28 13:41:34 crc kubenswrapper[4631]: I1128 13:41:34.872016 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0339440e-8081-4f68-9c6f-b62e5be5fc27-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-24v2l\" (UID: \"0339440e-8081-4f68-9c6f-b62e5be5fc27\") " pod="openstack/nova-cell1-cell-mapping-24v2l" Nov 28 13:41:34 crc kubenswrapper[4631]: I1128 13:41:34.872199 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0339440e-8081-4f68-9c6f-b62e5be5fc27-scripts\") pod \"nova-cell1-cell-mapping-24v2l\" (UID: \"0339440e-8081-4f68-9c6f-b62e5be5fc27\") " pod="openstack/nova-cell1-cell-mapping-24v2l" Nov 28 13:41:34 crc kubenswrapper[4631]: I1128 13:41:34.873570 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0339440e-8081-4f68-9c6f-b62e5be5fc27-config-data\") pod \"nova-cell1-cell-mapping-24v2l\" (UID: \"0339440e-8081-4f68-9c6f-b62e5be5fc27\") " pod="openstack/nova-cell1-cell-mapping-24v2l" Nov 28 13:41:34 crc kubenswrapper[4631]: I1128 13:41:34.897973 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t84lv\" (UniqueName: \"kubernetes.io/projected/0339440e-8081-4f68-9c6f-b62e5be5fc27-kube-api-access-t84lv\") pod \"nova-cell1-cell-mapping-24v2l\" (UID: \"0339440e-8081-4f68-9c6f-b62e5be5fc27\") " pod="openstack/nova-cell1-cell-mapping-24v2l" Nov 28 13:41:34 crc kubenswrapper[4631]: I1128 13:41:34.958230 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-24v2l" Nov 28 13:41:35 crc kubenswrapper[4631]: I1128 13:41:35.686990 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-24v2l"] Nov 28 13:41:36 crc kubenswrapper[4631]: I1128 13:41:36.100218 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 28 13:41:36 crc kubenswrapper[4631]: I1128 13:41:36.195211 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a00de8c7-a886-4851-b7cd-c8784a3c6bd9-combined-ca-bundle\") pod \"a00de8c7-a886-4851-b7cd-c8784a3c6bd9\" (UID: \"a00de8c7-a886-4851-b7cd-c8784a3c6bd9\") " Nov 28 13:41:36 crc kubenswrapper[4631]: I1128 13:41:36.195338 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a00de8c7-a886-4851-b7cd-c8784a3c6bd9-sg-core-conf-yaml\") pod \"a00de8c7-a886-4851-b7cd-c8784a3c6bd9\" (UID: \"a00de8c7-a886-4851-b7cd-c8784a3c6bd9\") " Nov 28 13:41:36 crc kubenswrapper[4631]: I1128 13:41:36.195531 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xnzmq\" (UniqueName: \"kubernetes.io/projected/a00de8c7-a886-4851-b7cd-c8784a3c6bd9-kube-api-access-xnzmq\") pod \"a00de8c7-a886-4851-b7cd-c8784a3c6bd9\" (UID: \"a00de8c7-a886-4851-b7cd-c8784a3c6bd9\") " Nov 28 13:41:36 crc kubenswrapper[4631]: I1128 13:41:36.195570 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a00de8c7-a886-4851-b7cd-c8784a3c6bd9-log-httpd\") pod \"a00de8c7-a886-4851-b7cd-c8784a3c6bd9\" (UID: \"a00de8c7-a886-4851-b7cd-c8784a3c6bd9\") " Nov 28 13:41:36 crc kubenswrapper[4631]: I1128 13:41:36.195599 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a00de8c7-a886-4851-b7cd-c8784a3c6bd9-scripts\") pod \"a00de8c7-a886-4851-b7cd-c8784a3c6bd9\" (UID: \"a00de8c7-a886-4851-b7cd-c8784a3c6bd9\") " Nov 28 13:41:36 crc kubenswrapper[4631]: I1128 13:41:36.195623 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a00de8c7-a886-4851-b7cd-c8784a3c6bd9-config-data\") pod \"a00de8c7-a886-4851-b7cd-c8784a3c6bd9\" (UID: \"a00de8c7-a886-4851-b7cd-c8784a3c6bd9\") " Nov 28 13:41:36 crc kubenswrapper[4631]: I1128 13:41:36.195656 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a00de8c7-a886-4851-b7cd-c8784a3c6bd9-run-httpd\") pod \"a00de8c7-a886-4851-b7cd-c8784a3c6bd9\" (UID: \"a00de8c7-a886-4851-b7cd-c8784a3c6bd9\") " Nov 28 13:41:36 crc kubenswrapper[4631]: I1128 13:41:36.195723 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/a00de8c7-a886-4851-b7cd-c8784a3c6bd9-ceilometer-tls-certs\") pod \"a00de8c7-a886-4851-b7cd-c8784a3c6bd9\" (UID: \"a00de8c7-a886-4851-b7cd-c8784a3c6bd9\") " Nov 28 13:41:36 crc kubenswrapper[4631]: I1128 13:41:36.196403 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a00de8c7-a886-4851-b7cd-c8784a3c6bd9-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "a00de8c7-a886-4851-b7cd-c8784a3c6bd9" (UID: "a00de8c7-a886-4851-b7cd-c8784a3c6bd9"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 13:41:36 crc kubenswrapper[4631]: I1128 13:41:36.220350 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a00de8c7-a886-4851-b7cd-c8784a3c6bd9-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "a00de8c7-a886-4851-b7cd-c8784a3c6bd9" (UID: "a00de8c7-a886-4851-b7cd-c8784a3c6bd9"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 13:41:36 crc kubenswrapper[4631]: I1128 13:41:36.228500 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a00de8c7-a886-4851-b7cd-c8784a3c6bd9-kube-api-access-xnzmq" (OuterVolumeSpecName: "kube-api-access-xnzmq") pod "a00de8c7-a886-4851-b7cd-c8784a3c6bd9" (UID: "a00de8c7-a886-4851-b7cd-c8784a3c6bd9"). InnerVolumeSpecName "kube-api-access-xnzmq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:41:36 crc kubenswrapper[4631]: I1128 13:41:36.258529 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a00de8c7-a886-4851-b7cd-c8784a3c6bd9-scripts" (OuterVolumeSpecName: "scripts") pod "a00de8c7-a886-4851-b7cd-c8784a3c6bd9" (UID: "a00de8c7-a886-4851-b7cd-c8784a3c6bd9"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:41:36 crc kubenswrapper[4631]: I1128 13:41:36.304135 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xnzmq\" (UniqueName: \"kubernetes.io/projected/a00de8c7-a886-4851-b7cd-c8784a3c6bd9-kube-api-access-xnzmq\") on node \"crc\" DevicePath \"\"" Nov 28 13:41:36 crc kubenswrapper[4631]: I1128 13:41:36.304665 4631 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a00de8c7-a886-4851-b7cd-c8784a3c6bd9-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 28 13:41:36 crc kubenswrapper[4631]: I1128 13:41:36.304680 4631 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a00de8c7-a886-4851-b7cd-c8784a3c6bd9-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 13:41:36 crc kubenswrapper[4631]: I1128 13:41:36.304690 4631 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a00de8c7-a886-4851-b7cd-c8784a3c6bd9-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 28 13:41:36 crc kubenswrapper[4631]: I1128 13:41:36.301136 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a00de8c7-a886-4851-b7cd-c8784a3c6bd9-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "a00de8c7-a886-4851-b7cd-c8784a3c6bd9" (UID: "a00de8c7-a886-4851-b7cd-c8784a3c6bd9"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:41:36 crc kubenswrapper[4631]: I1128 13:41:36.406850 4631 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a00de8c7-a886-4851-b7cd-c8784a3c6bd9-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 28 13:41:36 crc kubenswrapper[4631]: I1128 13:41:36.407102 4631 generic.go:334] "Generic (PLEG): container finished" podID="a00de8c7-a886-4851-b7cd-c8784a3c6bd9" containerID="0f6318a7db020712586bb2e56cdd6f8046b9a725a41e45e0a31daac51781c62c" exitCode=0 Nov 28 13:41:36 crc kubenswrapper[4631]: I1128 13:41:36.407138 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a00de8c7-a886-4851-b7cd-c8784a3c6bd9","Type":"ContainerDied","Data":"0f6318a7db020712586bb2e56cdd6f8046b9a725a41e45e0a31daac51781c62c"} Nov 28 13:41:36 crc kubenswrapper[4631]: I1128 13:41:36.408210 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a00de8c7-a886-4851-b7cd-c8784a3c6bd9","Type":"ContainerDied","Data":"83a90711a3c7c51a24a1c070bf63ee69b6f7698a291854e4ceffb37c59d5f4f5"} Nov 28 13:41:36 crc kubenswrapper[4631]: I1128 13:41:36.408361 4631 scope.go:117] "RemoveContainer" containerID="d1a7db8ddb7853f6005959d4900a1ada50c4962ce089b37695da99fd552f8573" Nov 28 13:41:36 crc kubenswrapper[4631]: I1128 13:41:36.407205 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 28 13:41:36 crc kubenswrapper[4631]: I1128 13:41:36.422705 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 28 13:41:36 crc kubenswrapper[4631]: I1128 13:41:36.427491 4631 generic.go:334] "Generic (PLEG): container finished" podID="44c10c63-05fe-47e8-acc6-c927c9c3db7e" containerID="cf453b65de3fb82eee03c5027c805c5311f5c1efd14c3e31bbfb557b4dcd04a8" exitCode=0 Nov 28 13:41:36 crc kubenswrapper[4631]: I1128 13:41:36.427598 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"44c10c63-05fe-47e8-acc6-c927c9c3db7e","Type":"ContainerDied","Data":"cf453b65de3fb82eee03c5027c805c5311f5c1efd14c3e31bbfb557b4dcd04a8"} Nov 28 13:41:36 crc kubenswrapper[4631]: I1128 13:41:36.434157 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-24v2l" event={"ID":"0339440e-8081-4f68-9c6f-b62e5be5fc27","Type":"ContainerStarted","Data":"5bb93c4d68648bdadf3355937b106f393adf48407dd01273d6aa3baebd029d62"} Nov 28 13:41:36 crc kubenswrapper[4631]: I1128 13:41:36.435713 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-24v2l" event={"ID":"0339440e-8081-4f68-9c6f-b62e5be5fc27","Type":"ContainerStarted","Data":"3a3a9d3fb8a28758700544881d4c314af2c0499207773ff730686faef51f5036"} Nov 28 13:41:36 crc kubenswrapper[4631]: I1128 13:41:36.439586 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a00de8c7-a886-4851-b7cd-c8784a3c6bd9-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "a00de8c7-a886-4851-b7cd-c8784a3c6bd9" (UID: "a00de8c7-a886-4851-b7cd-c8784a3c6bd9"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:41:36 crc kubenswrapper[4631]: I1128 13:41:36.493961 4631 scope.go:117] "RemoveContainer" containerID="11e2e2d4a8aff3ef637fdd4f4485a2cb4c252744fa8b2147478739303369e0e2" Nov 28 13:41:36 crc kubenswrapper[4631]: I1128 13:41:36.512405 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/44c10c63-05fe-47e8-acc6-c927c9c3db7e-logs\") pod \"44c10c63-05fe-47e8-acc6-c927c9c3db7e\" (UID: \"44c10c63-05fe-47e8-acc6-c927c9c3db7e\") " Nov 28 13:41:36 crc kubenswrapper[4631]: I1128 13:41:36.512481 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/44c10c63-05fe-47e8-acc6-c927c9c3db7e-combined-ca-bundle\") pod \"44c10c63-05fe-47e8-acc6-c927c9c3db7e\" (UID: \"44c10c63-05fe-47e8-acc6-c927c9c3db7e\") " Nov 28 13:41:36 crc kubenswrapper[4631]: I1128 13:41:36.512537 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/44c10c63-05fe-47e8-acc6-c927c9c3db7e-config-data\") pod \"44c10c63-05fe-47e8-acc6-c927c9c3db7e\" (UID: \"44c10c63-05fe-47e8-acc6-c927c9c3db7e\") " Nov 28 13:41:36 crc kubenswrapper[4631]: I1128 13:41:36.512710 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gb2lq\" (UniqueName: \"kubernetes.io/projected/44c10c63-05fe-47e8-acc6-c927c9c3db7e-kube-api-access-gb2lq\") pod \"44c10c63-05fe-47e8-acc6-c927c9c3db7e\" (UID: \"44c10c63-05fe-47e8-acc6-c927c9c3db7e\") " Nov 28 13:41:36 crc kubenswrapper[4631]: I1128 13:41:36.513197 4631 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/a00de8c7-a886-4851-b7cd-c8784a3c6bd9-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 28 13:41:36 crc kubenswrapper[4631]: I1128 13:41:36.531114 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/44c10c63-05fe-47e8-acc6-c927c9c3db7e-logs" (OuterVolumeSpecName: "logs") pod "44c10c63-05fe-47e8-acc6-c927c9c3db7e" (UID: "44c10c63-05fe-47e8-acc6-c927c9c3db7e"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 13:41:36 crc kubenswrapper[4631]: I1128 13:41:36.533275 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44c10c63-05fe-47e8-acc6-c927c9c3db7e-kube-api-access-gb2lq" (OuterVolumeSpecName: "kube-api-access-gb2lq") pod "44c10c63-05fe-47e8-acc6-c927c9c3db7e" (UID: "44c10c63-05fe-47e8-acc6-c927c9c3db7e"). InnerVolumeSpecName "kube-api-access-gb2lq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:41:36 crc kubenswrapper[4631]: I1128 13:41:36.540012 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-24v2l" podStartSLOduration=2.539985143 podStartE2EDuration="2.539985143s" podCreationTimestamp="2025-11-28 13:41:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 13:41:36.501470278 +0000 UTC m=+1253.308773622" watchObservedRunningTime="2025-11-28 13:41:36.539985143 +0000 UTC m=+1253.347288487" Nov 28 13:41:36 crc kubenswrapper[4631]: I1128 13:41:36.556088 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a00de8c7-a886-4851-b7cd-c8784a3c6bd9-config-data" (OuterVolumeSpecName: "config-data") pod "a00de8c7-a886-4851-b7cd-c8784a3c6bd9" (UID: "a00de8c7-a886-4851-b7cd-c8784a3c6bd9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:41:36 crc kubenswrapper[4631]: I1128 13:41:36.556113 4631 scope.go:117] "RemoveContainer" containerID="0f6318a7db020712586bb2e56cdd6f8046b9a725a41e45e0a31daac51781c62c" Nov 28 13:41:36 crc kubenswrapper[4631]: I1128 13:41:36.556206 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a00de8c7-a886-4851-b7cd-c8784a3c6bd9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a00de8c7-a886-4851-b7cd-c8784a3c6bd9" (UID: "a00de8c7-a886-4851-b7cd-c8784a3c6bd9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:41:36 crc kubenswrapper[4631]: I1128 13:41:36.593621 4631 scope.go:117] "RemoveContainer" containerID="3e70f5937aa64c1f6b17c239bba41cef0aa54ddf458ce5b8cafeccbb3e212581" Nov 28 13:41:36 crc kubenswrapper[4631]: I1128 13:41:36.603996 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/44c10c63-05fe-47e8-acc6-c927c9c3db7e-config-data" (OuterVolumeSpecName: "config-data") pod "44c10c63-05fe-47e8-acc6-c927c9c3db7e" (UID: "44c10c63-05fe-47e8-acc6-c927c9c3db7e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:41:36 crc kubenswrapper[4631]: I1128 13:41:36.604128 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/44c10c63-05fe-47e8-acc6-c927c9c3db7e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "44c10c63-05fe-47e8-acc6-c927c9c3db7e" (UID: "44c10c63-05fe-47e8-acc6-c927c9c3db7e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:41:36 crc kubenswrapper[4631]: I1128 13:41:36.614906 4631 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/44c10c63-05fe-47e8-acc6-c927c9c3db7e-logs\") on node \"crc\" DevicePath \"\"" Nov 28 13:41:36 crc kubenswrapper[4631]: I1128 13:41:36.614953 4631 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/44c10c63-05fe-47e8-acc6-c927c9c3db7e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 13:41:36 crc kubenswrapper[4631]: I1128 13:41:36.614968 4631 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/44c10c63-05fe-47e8-acc6-c927c9c3db7e-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 13:41:36 crc kubenswrapper[4631]: I1128 13:41:36.614981 4631 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a00de8c7-a886-4851-b7cd-c8784a3c6bd9-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 13:41:36 crc kubenswrapper[4631]: I1128 13:41:36.614993 4631 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a00de8c7-a886-4851-b7cd-c8784a3c6bd9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 13:41:36 crc kubenswrapper[4631]: I1128 13:41:36.615038 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gb2lq\" (UniqueName: \"kubernetes.io/projected/44c10c63-05fe-47e8-acc6-c927c9c3db7e-kube-api-access-gb2lq\") on node \"crc\" DevicePath \"\"" Nov 28 13:41:36 crc kubenswrapper[4631]: I1128 13:41:36.646163 4631 scope.go:117] "RemoveContainer" containerID="d1a7db8ddb7853f6005959d4900a1ada50c4962ce089b37695da99fd552f8573" Nov 28 13:41:36 crc kubenswrapper[4631]: E1128 13:41:36.646839 4631 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d1a7db8ddb7853f6005959d4900a1ada50c4962ce089b37695da99fd552f8573\": container with ID starting with d1a7db8ddb7853f6005959d4900a1ada50c4962ce089b37695da99fd552f8573 not found: ID does not exist" containerID="d1a7db8ddb7853f6005959d4900a1ada50c4962ce089b37695da99fd552f8573" Nov 28 13:41:36 crc kubenswrapper[4631]: I1128 13:41:36.646886 4631 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d1a7db8ddb7853f6005959d4900a1ada50c4962ce089b37695da99fd552f8573"} err="failed to get container status \"d1a7db8ddb7853f6005959d4900a1ada50c4962ce089b37695da99fd552f8573\": rpc error: code = NotFound desc = could not find container \"d1a7db8ddb7853f6005959d4900a1ada50c4962ce089b37695da99fd552f8573\": container with ID starting with d1a7db8ddb7853f6005959d4900a1ada50c4962ce089b37695da99fd552f8573 not found: ID does not exist" Nov 28 13:41:36 crc kubenswrapper[4631]: I1128 13:41:36.646915 4631 scope.go:117] "RemoveContainer" containerID="11e2e2d4a8aff3ef637fdd4f4485a2cb4c252744fa8b2147478739303369e0e2" Nov 28 13:41:36 crc kubenswrapper[4631]: E1128 13:41:36.647650 4631 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"11e2e2d4a8aff3ef637fdd4f4485a2cb4c252744fa8b2147478739303369e0e2\": container with ID starting with 11e2e2d4a8aff3ef637fdd4f4485a2cb4c252744fa8b2147478739303369e0e2 not found: ID does not exist" containerID="11e2e2d4a8aff3ef637fdd4f4485a2cb4c252744fa8b2147478739303369e0e2" Nov 28 13:41:36 crc kubenswrapper[4631]: I1128 13:41:36.647679 4631 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"11e2e2d4a8aff3ef637fdd4f4485a2cb4c252744fa8b2147478739303369e0e2"} err="failed to get container status \"11e2e2d4a8aff3ef637fdd4f4485a2cb4c252744fa8b2147478739303369e0e2\": rpc error: code = NotFound desc = could not find container \"11e2e2d4a8aff3ef637fdd4f4485a2cb4c252744fa8b2147478739303369e0e2\": container with ID starting with 11e2e2d4a8aff3ef637fdd4f4485a2cb4c252744fa8b2147478739303369e0e2 not found: ID does not exist" Nov 28 13:41:36 crc kubenswrapper[4631]: I1128 13:41:36.647694 4631 scope.go:117] "RemoveContainer" containerID="0f6318a7db020712586bb2e56cdd6f8046b9a725a41e45e0a31daac51781c62c" Nov 28 13:41:36 crc kubenswrapper[4631]: E1128 13:41:36.653449 4631 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0f6318a7db020712586bb2e56cdd6f8046b9a725a41e45e0a31daac51781c62c\": container with ID starting with 0f6318a7db020712586bb2e56cdd6f8046b9a725a41e45e0a31daac51781c62c not found: ID does not exist" containerID="0f6318a7db020712586bb2e56cdd6f8046b9a725a41e45e0a31daac51781c62c" Nov 28 13:41:36 crc kubenswrapper[4631]: I1128 13:41:36.653480 4631 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0f6318a7db020712586bb2e56cdd6f8046b9a725a41e45e0a31daac51781c62c"} err="failed to get container status \"0f6318a7db020712586bb2e56cdd6f8046b9a725a41e45e0a31daac51781c62c\": rpc error: code = NotFound desc = could not find container \"0f6318a7db020712586bb2e56cdd6f8046b9a725a41e45e0a31daac51781c62c\": container with ID starting with 0f6318a7db020712586bb2e56cdd6f8046b9a725a41e45e0a31daac51781c62c not found: ID does not exist" Nov 28 13:41:36 crc kubenswrapper[4631]: I1128 13:41:36.653495 4631 scope.go:117] "RemoveContainer" containerID="3e70f5937aa64c1f6b17c239bba41cef0aa54ddf458ce5b8cafeccbb3e212581" Nov 28 13:41:36 crc kubenswrapper[4631]: E1128 13:41:36.654742 4631 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3e70f5937aa64c1f6b17c239bba41cef0aa54ddf458ce5b8cafeccbb3e212581\": container with ID starting with 3e70f5937aa64c1f6b17c239bba41cef0aa54ddf458ce5b8cafeccbb3e212581 not found: ID does not exist" containerID="3e70f5937aa64c1f6b17c239bba41cef0aa54ddf458ce5b8cafeccbb3e212581" Nov 28 13:41:36 crc kubenswrapper[4631]: I1128 13:41:36.654763 4631 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3e70f5937aa64c1f6b17c239bba41cef0aa54ddf458ce5b8cafeccbb3e212581"} err="failed to get container status \"3e70f5937aa64c1f6b17c239bba41cef0aa54ddf458ce5b8cafeccbb3e212581\": rpc error: code = NotFound desc = could not find container \"3e70f5937aa64c1f6b17c239bba41cef0aa54ddf458ce5b8cafeccbb3e212581\": container with ID starting with 3e70f5937aa64c1f6b17c239bba41cef0aa54ddf458ce5b8cafeccbb3e212581 not found: ID does not exist" Nov 28 13:41:36 crc kubenswrapper[4631]: I1128 13:41:36.654777 4631 scope.go:117] "RemoveContainer" containerID="cf453b65de3fb82eee03c5027c805c5311f5c1efd14c3e31bbfb557b4dcd04a8" Nov 28 13:41:36 crc kubenswrapper[4631]: I1128 13:41:36.688504 4631 scope.go:117] "RemoveContainer" containerID="4103b4cf6b619a45021786723e84b8ad4e2fc43a3f9eab4611393b14c989a39e" Nov 28 13:41:36 crc kubenswrapper[4631]: I1128 13:41:36.777836 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 28 13:41:36 crc kubenswrapper[4631]: I1128 13:41:36.798409 4631 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 28 13:41:36 crc kubenswrapper[4631]: I1128 13:41:36.808233 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 28 13:41:36 crc kubenswrapper[4631]: E1128 13:41:36.809358 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a00de8c7-a886-4851-b7cd-c8784a3c6bd9" containerName="sg-core" Nov 28 13:41:36 crc kubenswrapper[4631]: I1128 13:41:36.809442 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="a00de8c7-a886-4851-b7cd-c8784a3c6bd9" containerName="sg-core" Nov 28 13:41:36 crc kubenswrapper[4631]: E1128 13:41:36.809525 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="44c10c63-05fe-47e8-acc6-c927c9c3db7e" containerName="nova-api-log" Nov 28 13:41:36 crc kubenswrapper[4631]: I1128 13:41:36.809581 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="44c10c63-05fe-47e8-acc6-c927c9c3db7e" containerName="nova-api-log" Nov 28 13:41:36 crc kubenswrapper[4631]: E1128 13:41:36.809661 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a00de8c7-a886-4851-b7cd-c8784a3c6bd9" containerName="proxy-httpd" Nov 28 13:41:36 crc kubenswrapper[4631]: I1128 13:41:36.809741 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="a00de8c7-a886-4851-b7cd-c8784a3c6bd9" containerName="proxy-httpd" Nov 28 13:41:36 crc kubenswrapper[4631]: E1128 13:41:36.809830 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a00de8c7-a886-4851-b7cd-c8784a3c6bd9" containerName="ceilometer-central-agent" Nov 28 13:41:36 crc kubenswrapper[4631]: I1128 13:41:36.809894 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="a00de8c7-a886-4851-b7cd-c8784a3c6bd9" containerName="ceilometer-central-agent" Nov 28 13:41:36 crc kubenswrapper[4631]: E1128 13:41:36.809980 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="44c10c63-05fe-47e8-acc6-c927c9c3db7e" containerName="nova-api-api" Nov 28 13:41:36 crc kubenswrapper[4631]: I1128 13:41:36.810043 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="44c10c63-05fe-47e8-acc6-c927c9c3db7e" containerName="nova-api-api" Nov 28 13:41:36 crc kubenswrapper[4631]: E1128 13:41:36.810119 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a00de8c7-a886-4851-b7cd-c8784a3c6bd9" containerName="ceilometer-notification-agent" Nov 28 13:41:36 crc kubenswrapper[4631]: I1128 13:41:36.810189 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="a00de8c7-a886-4851-b7cd-c8784a3c6bd9" containerName="ceilometer-notification-agent" Nov 28 13:41:36 crc kubenswrapper[4631]: I1128 13:41:36.810475 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="44c10c63-05fe-47e8-acc6-c927c9c3db7e" containerName="nova-api-api" Nov 28 13:41:36 crc kubenswrapper[4631]: I1128 13:41:36.810561 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="44c10c63-05fe-47e8-acc6-c927c9c3db7e" containerName="nova-api-log" Nov 28 13:41:36 crc kubenswrapper[4631]: I1128 13:41:36.810625 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="a00de8c7-a886-4851-b7cd-c8784a3c6bd9" containerName="ceilometer-central-agent" Nov 28 13:41:36 crc kubenswrapper[4631]: I1128 13:41:36.810687 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="a00de8c7-a886-4851-b7cd-c8784a3c6bd9" containerName="ceilometer-notification-agent" Nov 28 13:41:36 crc kubenswrapper[4631]: I1128 13:41:36.810748 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="a00de8c7-a886-4851-b7cd-c8784a3c6bd9" containerName="sg-core" Nov 28 13:41:36 crc kubenswrapper[4631]: I1128 13:41:36.810799 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="a00de8c7-a886-4851-b7cd-c8784a3c6bd9" containerName="proxy-httpd" Nov 28 13:41:36 crc kubenswrapper[4631]: I1128 13:41:36.813312 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 28 13:41:36 crc kubenswrapper[4631]: I1128 13:41:36.817243 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Nov 28 13:41:36 crc kubenswrapper[4631]: I1128 13:41:36.817518 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 28 13:41:36 crc kubenswrapper[4631]: I1128 13:41:36.817789 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 28 13:41:36 crc kubenswrapper[4631]: I1128 13:41:36.833417 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 28 13:41:36 crc kubenswrapper[4631]: I1128 13:41:36.923548 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4cde0c05-4a6a-410c-b5c3-2f14289fc37a-log-httpd\") pod \"ceilometer-0\" (UID: \"4cde0c05-4a6a-410c-b5c3-2f14289fc37a\") " pod="openstack/ceilometer-0" Nov 28 13:41:36 crc kubenswrapper[4631]: I1128 13:41:36.923623 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6snlh\" (UniqueName: \"kubernetes.io/projected/4cde0c05-4a6a-410c-b5c3-2f14289fc37a-kube-api-access-6snlh\") pod \"ceilometer-0\" (UID: \"4cde0c05-4a6a-410c-b5c3-2f14289fc37a\") " pod="openstack/ceilometer-0" Nov 28 13:41:36 crc kubenswrapper[4631]: I1128 13:41:36.923817 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4cde0c05-4a6a-410c-b5c3-2f14289fc37a-config-data\") pod \"ceilometer-0\" (UID: \"4cde0c05-4a6a-410c-b5c3-2f14289fc37a\") " pod="openstack/ceilometer-0" Nov 28 13:41:36 crc kubenswrapper[4631]: I1128 13:41:36.924059 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/4cde0c05-4a6a-410c-b5c3-2f14289fc37a-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"4cde0c05-4a6a-410c-b5c3-2f14289fc37a\") " pod="openstack/ceilometer-0" Nov 28 13:41:36 crc kubenswrapper[4631]: I1128 13:41:36.924245 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4cde0c05-4a6a-410c-b5c3-2f14289fc37a-run-httpd\") pod \"ceilometer-0\" (UID: \"4cde0c05-4a6a-410c-b5c3-2f14289fc37a\") " pod="openstack/ceilometer-0" Nov 28 13:41:36 crc kubenswrapper[4631]: I1128 13:41:36.924393 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4cde0c05-4a6a-410c-b5c3-2f14289fc37a-scripts\") pod \"ceilometer-0\" (UID: \"4cde0c05-4a6a-410c-b5c3-2f14289fc37a\") " pod="openstack/ceilometer-0" Nov 28 13:41:36 crc kubenswrapper[4631]: I1128 13:41:36.924462 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4cde0c05-4a6a-410c-b5c3-2f14289fc37a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"4cde0c05-4a6a-410c-b5c3-2f14289fc37a\") " pod="openstack/ceilometer-0" Nov 28 13:41:36 crc kubenswrapper[4631]: I1128 13:41:36.924677 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4cde0c05-4a6a-410c-b5c3-2f14289fc37a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"4cde0c05-4a6a-410c-b5c3-2f14289fc37a\") " pod="openstack/ceilometer-0" Nov 28 13:41:37 crc kubenswrapper[4631]: I1128 13:41:37.027555 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6snlh\" (UniqueName: \"kubernetes.io/projected/4cde0c05-4a6a-410c-b5c3-2f14289fc37a-kube-api-access-6snlh\") pod \"ceilometer-0\" (UID: \"4cde0c05-4a6a-410c-b5c3-2f14289fc37a\") " pod="openstack/ceilometer-0" Nov 28 13:41:37 crc kubenswrapper[4631]: I1128 13:41:37.027635 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4cde0c05-4a6a-410c-b5c3-2f14289fc37a-config-data\") pod \"ceilometer-0\" (UID: \"4cde0c05-4a6a-410c-b5c3-2f14289fc37a\") " pod="openstack/ceilometer-0" Nov 28 13:41:37 crc kubenswrapper[4631]: I1128 13:41:37.027685 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/4cde0c05-4a6a-410c-b5c3-2f14289fc37a-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"4cde0c05-4a6a-410c-b5c3-2f14289fc37a\") " pod="openstack/ceilometer-0" Nov 28 13:41:37 crc kubenswrapper[4631]: I1128 13:41:37.027782 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4cde0c05-4a6a-410c-b5c3-2f14289fc37a-run-httpd\") pod \"ceilometer-0\" (UID: \"4cde0c05-4a6a-410c-b5c3-2f14289fc37a\") " pod="openstack/ceilometer-0" Nov 28 13:41:37 crc kubenswrapper[4631]: I1128 13:41:37.027819 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4cde0c05-4a6a-410c-b5c3-2f14289fc37a-scripts\") pod \"ceilometer-0\" (UID: \"4cde0c05-4a6a-410c-b5c3-2f14289fc37a\") " pod="openstack/ceilometer-0" Nov 28 13:41:37 crc kubenswrapper[4631]: I1128 13:41:37.027845 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4cde0c05-4a6a-410c-b5c3-2f14289fc37a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"4cde0c05-4a6a-410c-b5c3-2f14289fc37a\") " pod="openstack/ceilometer-0" Nov 28 13:41:37 crc kubenswrapper[4631]: I1128 13:41:37.027892 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4cde0c05-4a6a-410c-b5c3-2f14289fc37a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"4cde0c05-4a6a-410c-b5c3-2f14289fc37a\") " pod="openstack/ceilometer-0" Nov 28 13:41:37 crc kubenswrapper[4631]: I1128 13:41:37.027976 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4cde0c05-4a6a-410c-b5c3-2f14289fc37a-log-httpd\") pod \"ceilometer-0\" (UID: \"4cde0c05-4a6a-410c-b5c3-2f14289fc37a\") " pod="openstack/ceilometer-0" Nov 28 13:41:37 crc kubenswrapper[4631]: I1128 13:41:37.028548 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4cde0c05-4a6a-410c-b5c3-2f14289fc37a-run-httpd\") pod \"ceilometer-0\" (UID: \"4cde0c05-4a6a-410c-b5c3-2f14289fc37a\") " pod="openstack/ceilometer-0" Nov 28 13:41:37 crc kubenswrapper[4631]: I1128 13:41:37.028699 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4cde0c05-4a6a-410c-b5c3-2f14289fc37a-log-httpd\") pod \"ceilometer-0\" (UID: \"4cde0c05-4a6a-410c-b5c3-2f14289fc37a\") " pod="openstack/ceilometer-0" Nov 28 13:41:37 crc kubenswrapper[4631]: I1128 13:41:37.033869 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/4cde0c05-4a6a-410c-b5c3-2f14289fc37a-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"4cde0c05-4a6a-410c-b5c3-2f14289fc37a\") " pod="openstack/ceilometer-0" Nov 28 13:41:37 crc kubenswrapper[4631]: I1128 13:41:37.034665 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4cde0c05-4a6a-410c-b5c3-2f14289fc37a-scripts\") pod \"ceilometer-0\" (UID: \"4cde0c05-4a6a-410c-b5c3-2f14289fc37a\") " pod="openstack/ceilometer-0" Nov 28 13:41:37 crc kubenswrapper[4631]: I1128 13:41:37.034885 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4cde0c05-4a6a-410c-b5c3-2f14289fc37a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"4cde0c05-4a6a-410c-b5c3-2f14289fc37a\") " pod="openstack/ceilometer-0" Nov 28 13:41:37 crc kubenswrapper[4631]: I1128 13:41:37.040793 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4cde0c05-4a6a-410c-b5c3-2f14289fc37a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"4cde0c05-4a6a-410c-b5c3-2f14289fc37a\") " pod="openstack/ceilometer-0" Nov 28 13:41:37 crc kubenswrapper[4631]: I1128 13:41:37.041490 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4cde0c05-4a6a-410c-b5c3-2f14289fc37a-config-data\") pod \"ceilometer-0\" (UID: \"4cde0c05-4a6a-410c-b5c3-2f14289fc37a\") " pod="openstack/ceilometer-0" Nov 28 13:41:37 crc kubenswrapper[4631]: I1128 13:41:37.046832 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6snlh\" (UniqueName: \"kubernetes.io/projected/4cde0c05-4a6a-410c-b5c3-2f14289fc37a-kube-api-access-6snlh\") pod \"ceilometer-0\" (UID: \"4cde0c05-4a6a-410c-b5c3-2f14289fc37a\") " pod="openstack/ceilometer-0" Nov 28 13:41:37 crc kubenswrapper[4631]: I1128 13:41:37.130698 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 28 13:41:37 crc kubenswrapper[4631]: I1128 13:41:37.448115 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"44c10c63-05fe-47e8-acc6-c927c9c3db7e","Type":"ContainerDied","Data":"8c94b35aaaf92a53e543b23309fc94a5193c4352d5814a7f17d376531709b814"} Nov 28 13:41:37 crc kubenswrapper[4631]: I1128 13:41:37.448139 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 28 13:41:37 crc kubenswrapper[4631]: I1128 13:41:37.500882 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 28 13:41:37 crc kubenswrapper[4631]: I1128 13:41:37.509359 4631 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Nov 28 13:41:37 crc kubenswrapper[4631]: I1128 13:41:37.529229 4631 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44c10c63-05fe-47e8-acc6-c927c9c3db7e" path="/var/lib/kubelet/pods/44c10c63-05fe-47e8-acc6-c927c9c3db7e/volumes" Nov 28 13:41:37 crc kubenswrapper[4631]: I1128 13:41:37.530218 4631 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a00de8c7-a886-4851-b7cd-c8784a3c6bd9" path="/var/lib/kubelet/pods/a00de8c7-a886-4851-b7cd-c8784a3c6bd9/volumes" Nov 28 13:41:37 crc kubenswrapper[4631]: I1128 13:41:37.534532 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Nov 28 13:41:37 crc kubenswrapper[4631]: I1128 13:41:37.536409 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 28 13:41:37 crc kubenswrapper[4631]: I1128 13:41:37.541675 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Nov 28 13:41:37 crc kubenswrapper[4631]: I1128 13:41:37.541956 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Nov 28 13:41:37 crc kubenswrapper[4631]: I1128 13:41:37.542191 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Nov 28 13:41:37 crc kubenswrapper[4631]: I1128 13:41:37.556135 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 28 13:41:37 crc kubenswrapper[4631]: I1128 13:41:37.640209 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 28 13:41:37 crc kubenswrapper[4631]: I1128 13:41:37.644570 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/122d85b6-1e49-4493-87be-34104a78a04a-public-tls-certs\") pod \"nova-api-0\" (UID: \"122d85b6-1e49-4493-87be-34104a78a04a\") " pod="openstack/nova-api-0" Nov 28 13:41:37 crc kubenswrapper[4631]: I1128 13:41:37.644702 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/122d85b6-1e49-4493-87be-34104a78a04a-internal-tls-certs\") pod \"nova-api-0\" (UID: \"122d85b6-1e49-4493-87be-34104a78a04a\") " pod="openstack/nova-api-0" Nov 28 13:41:37 crc kubenswrapper[4631]: I1128 13:41:37.644763 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/122d85b6-1e49-4493-87be-34104a78a04a-logs\") pod \"nova-api-0\" (UID: \"122d85b6-1e49-4493-87be-34104a78a04a\") " pod="openstack/nova-api-0" Nov 28 13:41:37 crc kubenswrapper[4631]: I1128 13:41:37.644812 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/122d85b6-1e49-4493-87be-34104a78a04a-config-data\") pod \"nova-api-0\" (UID: \"122d85b6-1e49-4493-87be-34104a78a04a\") " pod="openstack/nova-api-0" Nov 28 13:41:37 crc kubenswrapper[4631]: I1128 13:41:37.644840 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2pkdd\" (UniqueName: \"kubernetes.io/projected/122d85b6-1e49-4493-87be-34104a78a04a-kube-api-access-2pkdd\") pod \"nova-api-0\" (UID: \"122d85b6-1e49-4493-87be-34104a78a04a\") " pod="openstack/nova-api-0" Nov 28 13:41:37 crc kubenswrapper[4631]: I1128 13:41:37.644879 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/122d85b6-1e49-4493-87be-34104a78a04a-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"122d85b6-1e49-4493-87be-34104a78a04a\") " pod="openstack/nova-api-0" Nov 28 13:41:37 crc kubenswrapper[4631]: I1128 13:41:37.746775 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/122d85b6-1e49-4493-87be-34104a78a04a-logs\") pod \"nova-api-0\" (UID: \"122d85b6-1e49-4493-87be-34104a78a04a\") " pod="openstack/nova-api-0" Nov 28 13:41:37 crc kubenswrapper[4631]: I1128 13:41:37.747965 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/122d85b6-1e49-4493-87be-34104a78a04a-config-data\") pod \"nova-api-0\" (UID: \"122d85b6-1e49-4493-87be-34104a78a04a\") " pod="openstack/nova-api-0" Nov 28 13:41:37 crc kubenswrapper[4631]: I1128 13:41:37.748114 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2pkdd\" (UniqueName: \"kubernetes.io/projected/122d85b6-1e49-4493-87be-34104a78a04a-kube-api-access-2pkdd\") pod \"nova-api-0\" (UID: \"122d85b6-1e49-4493-87be-34104a78a04a\") " pod="openstack/nova-api-0" Nov 28 13:41:37 crc kubenswrapper[4631]: I1128 13:41:37.748240 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/122d85b6-1e49-4493-87be-34104a78a04a-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"122d85b6-1e49-4493-87be-34104a78a04a\") " pod="openstack/nova-api-0" Nov 28 13:41:37 crc kubenswrapper[4631]: I1128 13:41:37.748393 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/122d85b6-1e49-4493-87be-34104a78a04a-public-tls-certs\") pod \"nova-api-0\" (UID: \"122d85b6-1e49-4493-87be-34104a78a04a\") " pod="openstack/nova-api-0" Nov 28 13:41:37 crc kubenswrapper[4631]: I1128 13:41:37.748552 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/122d85b6-1e49-4493-87be-34104a78a04a-internal-tls-certs\") pod \"nova-api-0\" (UID: \"122d85b6-1e49-4493-87be-34104a78a04a\") " pod="openstack/nova-api-0" Nov 28 13:41:37 crc kubenswrapper[4631]: I1128 13:41:37.747336 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/122d85b6-1e49-4493-87be-34104a78a04a-logs\") pod \"nova-api-0\" (UID: \"122d85b6-1e49-4493-87be-34104a78a04a\") " pod="openstack/nova-api-0" Nov 28 13:41:37 crc kubenswrapper[4631]: I1128 13:41:37.754386 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/122d85b6-1e49-4493-87be-34104a78a04a-public-tls-certs\") pod \"nova-api-0\" (UID: \"122d85b6-1e49-4493-87be-34104a78a04a\") " pod="openstack/nova-api-0" Nov 28 13:41:37 crc kubenswrapper[4631]: I1128 13:41:37.755335 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/122d85b6-1e49-4493-87be-34104a78a04a-internal-tls-certs\") pod \"nova-api-0\" (UID: \"122d85b6-1e49-4493-87be-34104a78a04a\") " pod="openstack/nova-api-0" Nov 28 13:41:37 crc kubenswrapper[4631]: I1128 13:41:37.756165 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/122d85b6-1e49-4493-87be-34104a78a04a-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"122d85b6-1e49-4493-87be-34104a78a04a\") " pod="openstack/nova-api-0" Nov 28 13:41:37 crc kubenswrapper[4631]: I1128 13:41:37.757848 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/122d85b6-1e49-4493-87be-34104a78a04a-config-data\") pod \"nova-api-0\" (UID: \"122d85b6-1e49-4493-87be-34104a78a04a\") " pod="openstack/nova-api-0" Nov 28 13:41:37 crc kubenswrapper[4631]: I1128 13:41:37.767614 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2pkdd\" (UniqueName: \"kubernetes.io/projected/122d85b6-1e49-4493-87be-34104a78a04a-kube-api-access-2pkdd\") pod \"nova-api-0\" (UID: \"122d85b6-1e49-4493-87be-34104a78a04a\") " pod="openstack/nova-api-0" Nov 28 13:41:37 crc kubenswrapper[4631]: I1128 13:41:37.857829 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 28 13:41:38 crc kubenswrapper[4631]: I1128 13:41:38.399607 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 28 13:41:38 crc kubenswrapper[4631]: I1128 13:41:38.466926 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"122d85b6-1e49-4493-87be-34104a78a04a","Type":"ContainerStarted","Data":"b24df505c8b56b8c21762b4c7825f6f3537467020acbab505e024281509d1f2b"} Nov 28 13:41:38 crc kubenswrapper[4631]: I1128 13:41:38.468999 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4cde0c05-4a6a-410c-b5c3-2f14289fc37a","Type":"ContainerStarted","Data":"b3be159d8c277bbe61326902f15f758ac890ee1429d1a9780f2326c46de75391"} Nov 28 13:41:39 crc kubenswrapper[4631]: I1128 13:41:39.484264 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4cde0c05-4a6a-410c-b5c3-2f14289fc37a","Type":"ContainerStarted","Data":"8f0660e5646e89a7496e85ffbe6eaa0eb619ac164918dcf6cfbe6a929b862f13"} Nov 28 13:41:39 crc kubenswrapper[4631]: I1128 13:41:39.485244 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4cde0c05-4a6a-410c-b5c3-2f14289fc37a","Type":"ContainerStarted","Data":"2dc23a60990a33ec863e87132f1e2ba4c95fba7ad7e6167d9a23ea006b856160"} Nov 28 13:41:39 crc kubenswrapper[4631]: I1128 13:41:39.488156 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"122d85b6-1e49-4493-87be-34104a78a04a","Type":"ContainerStarted","Data":"2e12690f233ccdc38fedfc24a0d67e900ed57e8bbe390c331bca481340fdb398"} Nov 28 13:41:39 crc kubenswrapper[4631]: I1128 13:41:39.488191 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"122d85b6-1e49-4493-87be-34104a78a04a","Type":"ContainerStarted","Data":"734b5c7339b19dfeb770c6c153a659a1c25016ba4f54511557ef97d7267a8dd1"} Nov 28 13:41:39 crc kubenswrapper[4631]: I1128 13:41:39.537988 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.537954573 podStartE2EDuration="2.537954573s" podCreationTimestamp="2025-11-28 13:41:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 13:41:39.518774637 +0000 UTC m=+1256.326078001" watchObservedRunningTime="2025-11-28 13:41:39.537954573 +0000 UTC m=+1256.345257917" Nov 28 13:41:39 crc kubenswrapper[4631]: I1128 13:41:39.883227 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-89c5cd4d5-qvl5b" Nov 28 13:41:40 crc kubenswrapper[4631]: I1128 13:41:40.021082 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-757b4f8459-j7gzn"] Nov 28 13:41:40 crc kubenswrapper[4631]: I1128 13:41:40.021408 4631 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-757b4f8459-j7gzn" podUID="e1644d33-bda9-437e-b10c-5c48042cb1f2" containerName="dnsmasq-dns" containerID="cri-o://33c781584549893b9d761b9f653c86dbdd728267f0f6a18b982048c3cdf700a6" gracePeriod=10 Nov 28 13:41:40 crc kubenswrapper[4631]: E1128 13:41:40.070597 4631 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod111c5349_528d_4856_a3e4_1277e3b10889.slice/crio-3b826a5556ea3b27f6357d4cc74220f4bce293db6dbcb49fa5131013bef0e517.scope\": RecentStats: unable to find data in memory cache]" Nov 28 13:41:40 crc kubenswrapper[4631]: I1128 13:41:40.528989 4631 generic.go:334] "Generic (PLEG): container finished" podID="e1644d33-bda9-437e-b10c-5c48042cb1f2" containerID="33c781584549893b9d761b9f653c86dbdd728267f0f6a18b982048c3cdf700a6" exitCode=0 Nov 28 13:41:40 crc kubenswrapper[4631]: I1128 13:41:40.529561 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-757b4f8459-j7gzn" event={"ID":"e1644d33-bda9-437e-b10c-5c48042cb1f2","Type":"ContainerDied","Data":"33c781584549893b9d761b9f653c86dbdd728267f0f6a18b982048c3cdf700a6"} Nov 28 13:41:40 crc kubenswrapper[4631]: I1128 13:41:40.546173 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4cde0c05-4a6a-410c-b5c3-2f14289fc37a","Type":"ContainerStarted","Data":"9882a72cfb32800dbb90ce24e0548ddb4d14f7f799a41a891b0dd57f70ba4b8a"} Nov 28 13:41:40 crc kubenswrapper[4631]: I1128 13:41:40.666973 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-757b4f8459-j7gzn" Nov 28 13:41:40 crc kubenswrapper[4631]: I1128 13:41:40.753457 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zz4pn\" (UniqueName: \"kubernetes.io/projected/e1644d33-bda9-437e-b10c-5c48042cb1f2-kube-api-access-zz4pn\") pod \"e1644d33-bda9-437e-b10c-5c48042cb1f2\" (UID: \"e1644d33-bda9-437e-b10c-5c48042cb1f2\") " Nov 28 13:41:40 crc kubenswrapper[4631]: I1128 13:41:40.753552 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e1644d33-bda9-437e-b10c-5c48042cb1f2-dns-svc\") pod \"e1644d33-bda9-437e-b10c-5c48042cb1f2\" (UID: \"e1644d33-bda9-437e-b10c-5c48042cb1f2\") " Nov 28 13:41:40 crc kubenswrapper[4631]: I1128 13:41:40.753602 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e1644d33-bda9-437e-b10c-5c48042cb1f2-config\") pod \"e1644d33-bda9-437e-b10c-5c48042cb1f2\" (UID: \"e1644d33-bda9-437e-b10c-5c48042cb1f2\") " Nov 28 13:41:40 crc kubenswrapper[4631]: I1128 13:41:40.753642 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e1644d33-bda9-437e-b10c-5c48042cb1f2-dns-swift-storage-0\") pod \"e1644d33-bda9-437e-b10c-5c48042cb1f2\" (UID: \"e1644d33-bda9-437e-b10c-5c48042cb1f2\") " Nov 28 13:41:40 crc kubenswrapper[4631]: I1128 13:41:40.753701 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e1644d33-bda9-437e-b10c-5c48042cb1f2-ovsdbserver-nb\") pod \"e1644d33-bda9-437e-b10c-5c48042cb1f2\" (UID: \"e1644d33-bda9-437e-b10c-5c48042cb1f2\") " Nov 28 13:41:40 crc kubenswrapper[4631]: I1128 13:41:40.753754 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e1644d33-bda9-437e-b10c-5c48042cb1f2-ovsdbserver-sb\") pod \"e1644d33-bda9-437e-b10c-5c48042cb1f2\" (UID: \"e1644d33-bda9-437e-b10c-5c48042cb1f2\") " Nov 28 13:41:40 crc kubenswrapper[4631]: I1128 13:41:40.785958 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e1644d33-bda9-437e-b10c-5c48042cb1f2-kube-api-access-zz4pn" (OuterVolumeSpecName: "kube-api-access-zz4pn") pod "e1644d33-bda9-437e-b10c-5c48042cb1f2" (UID: "e1644d33-bda9-437e-b10c-5c48042cb1f2"). InnerVolumeSpecName "kube-api-access-zz4pn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:41:40 crc kubenswrapper[4631]: I1128 13:41:40.856923 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e1644d33-bda9-437e-b10c-5c48042cb1f2-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "e1644d33-bda9-437e-b10c-5c48042cb1f2" (UID: "e1644d33-bda9-437e-b10c-5c48042cb1f2"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 13:41:40 crc kubenswrapper[4631]: I1128 13:41:40.866184 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e1644d33-bda9-437e-b10c-5c48042cb1f2-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "e1644d33-bda9-437e-b10c-5c48042cb1f2" (UID: "e1644d33-bda9-437e-b10c-5c48042cb1f2"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 13:41:40 crc kubenswrapper[4631]: I1128 13:41:40.870690 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zz4pn\" (UniqueName: \"kubernetes.io/projected/e1644d33-bda9-437e-b10c-5c48042cb1f2-kube-api-access-zz4pn\") on node \"crc\" DevicePath \"\"" Nov 28 13:41:40 crc kubenswrapper[4631]: I1128 13:41:40.870733 4631 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e1644d33-bda9-437e-b10c-5c48042cb1f2-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 28 13:41:40 crc kubenswrapper[4631]: I1128 13:41:40.870753 4631 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e1644d33-bda9-437e-b10c-5c48042cb1f2-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 28 13:41:40 crc kubenswrapper[4631]: I1128 13:41:40.899226 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e1644d33-bda9-437e-b10c-5c48042cb1f2-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "e1644d33-bda9-437e-b10c-5c48042cb1f2" (UID: "e1644d33-bda9-437e-b10c-5c48042cb1f2"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 13:41:40 crc kubenswrapper[4631]: I1128 13:41:40.939892 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e1644d33-bda9-437e-b10c-5c48042cb1f2-config" (OuterVolumeSpecName: "config") pod "e1644d33-bda9-437e-b10c-5c48042cb1f2" (UID: "e1644d33-bda9-437e-b10c-5c48042cb1f2"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 13:41:40 crc kubenswrapper[4631]: I1128 13:41:40.940706 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e1644d33-bda9-437e-b10c-5c48042cb1f2-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "e1644d33-bda9-437e-b10c-5c48042cb1f2" (UID: "e1644d33-bda9-437e-b10c-5c48042cb1f2"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 13:41:40 crc kubenswrapper[4631]: I1128 13:41:40.973272 4631 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e1644d33-bda9-437e-b10c-5c48042cb1f2-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 28 13:41:40 crc kubenswrapper[4631]: I1128 13:41:40.973320 4631 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e1644d33-bda9-437e-b10c-5c48042cb1f2-config\") on node \"crc\" DevicePath \"\"" Nov 28 13:41:40 crc kubenswrapper[4631]: I1128 13:41:40.973333 4631 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e1644d33-bda9-437e-b10c-5c48042cb1f2-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 28 13:41:41 crc kubenswrapper[4631]: I1128 13:41:41.557591 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-757b4f8459-j7gzn" event={"ID":"e1644d33-bda9-437e-b10c-5c48042cb1f2","Type":"ContainerDied","Data":"800facb9428255a99f2c1e48e7b635193d89440e1be749b4236bd11ac22bf4fc"} Nov 28 13:41:41 crc kubenswrapper[4631]: I1128 13:41:41.557750 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-757b4f8459-j7gzn" Nov 28 13:41:41 crc kubenswrapper[4631]: I1128 13:41:41.558809 4631 scope.go:117] "RemoveContainer" containerID="33c781584549893b9d761b9f653c86dbdd728267f0f6a18b982048c3cdf700a6" Nov 28 13:41:41 crc kubenswrapper[4631]: I1128 13:41:41.585500 4631 scope.go:117] "RemoveContainer" containerID="cb0fed8758d1f40a3a884391b7df6bf2761bdff294e9368089d1c7134ae186fe" Nov 28 13:41:41 crc kubenswrapper[4631]: I1128 13:41:41.587563 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-757b4f8459-j7gzn"] Nov 28 13:41:41 crc kubenswrapper[4631]: I1128 13:41:41.599333 4631 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-757b4f8459-j7gzn"] Nov 28 13:41:42 crc kubenswrapper[4631]: I1128 13:41:42.589713 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4cde0c05-4a6a-410c-b5c3-2f14289fc37a","Type":"ContainerStarted","Data":"ab83bfb449d3f94b880f0430a8532d440b92b3179fd594b71bbf290d756a4da6"} Nov 28 13:41:42 crc kubenswrapper[4631]: I1128 13:41:42.590278 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 28 13:41:42 crc kubenswrapper[4631]: I1128 13:41:42.632601 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.487147256 podStartE2EDuration="6.632575251s" podCreationTimestamp="2025-11-28 13:41:36 +0000 UTC" firstStartedPulling="2025-11-28 13:41:37.644013978 +0000 UTC m=+1254.451317322" lastFinishedPulling="2025-11-28 13:41:41.789441973 +0000 UTC m=+1258.596745317" observedRunningTime="2025-11-28 13:41:42.622762803 +0000 UTC m=+1259.430066157" watchObservedRunningTime="2025-11-28 13:41:42.632575251 +0000 UTC m=+1259.439878595" Nov 28 13:41:43 crc kubenswrapper[4631]: I1128 13:41:43.528265 4631 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e1644d33-bda9-437e-b10c-5c48042cb1f2" path="/var/lib/kubelet/pods/e1644d33-bda9-437e-b10c-5c48042cb1f2/volumes" Nov 28 13:41:44 crc kubenswrapper[4631]: I1128 13:41:44.620927 4631 generic.go:334] "Generic (PLEG): container finished" podID="0339440e-8081-4f68-9c6f-b62e5be5fc27" containerID="5bb93c4d68648bdadf3355937b106f393adf48407dd01273d6aa3baebd029d62" exitCode=0 Nov 28 13:41:44 crc kubenswrapper[4631]: I1128 13:41:44.621003 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-24v2l" event={"ID":"0339440e-8081-4f68-9c6f-b62e5be5fc27","Type":"ContainerDied","Data":"5bb93c4d68648bdadf3355937b106f393adf48407dd01273d6aa3baebd029d62"} Nov 28 13:41:45 crc kubenswrapper[4631]: I1128 13:41:45.433365 4631 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-757b4f8459-j7gzn" podUID="e1644d33-bda9-437e-b10c-5c48042cb1f2" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.190:5353: i/o timeout" Nov 28 13:41:46 crc kubenswrapper[4631]: I1128 13:41:46.024361 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-24v2l" Nov 28 13:41:46 crc kubenswrapper[4631]: I1128 13:41:46.089625 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0339440e-8081-4f68-9c6f-b62e5be5fc27-config-data\") pod \"0339440e-8081-4f68-9c6f-b62e5be5fc27\" (UID: \"0339440e-8081-4f68-9c6f-b62e5be5fc27\") " Nov 28 13:41:46 crc kubenswrapper[4631]: I1128 13:41:46.090646 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0339440e-8081-4f68-9c6f-b62e5be5fc27-combined-ca-bundle\") pod \"0339440e-8081-4f68-9c6f-b62e5be5fc27\" (UID: \"0339440e-8081-4f68-9c6f-b62e5be5fc27\") " Nov 28 13:41:46 crc kubenswrapper[4631]: I1128 13:41:46.090706 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t84lv\" (UniqueName: \"kubernetes.io/projected/0339440e-8081-4f68-9c6f-b62e5be5fc27-kube-api-access-t84lv\") pod \"0339440e-8081-4f68-9c6f-b62e5be5fc27\" (UID: \"0339440e-8081-4f68-9c6f-b62e5be5fc27\") " Nov 28 13:41:46 crc kubenswrapper[4631]: I1128 13:41:46.091016 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0339440e-8081-4f68-9c6f-b62e5be5fc27-scripts\") pod \"0339440e-8081-4f68-9c6f-b62e5be5fc27\" (UID: \"0339440e-8081-4f68-9c6f-b62e5be5fc27\") " Nov 28 13:41:46 crc kubenswrapper[4631]: I1128 13:41:46.103545 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0339440e-8081-4f68-9c6f-b62e5be5fc27-scripts" (OuterVolumeSpecName: "scripts") pod "0339440e-8081-4f68-9c6f-b62e5be5fc27" (UID: "0339440e-8081-4f68-9c6f-b62e5be5fc27"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:41:46 crc kubenswrapper[4631]: I1128 13:41:46.103573 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0339440e-8081-4f68-9c6f-b62e5be5fc27-kube-api-access-t84lv" (OuterVolumeSpecName: "kube-api-access-t84lv") pod "0339440e-8081-4f68-9c6f-b62e5be5fc27" (UID: "0339440e-8081-4f68-9c6f-b62e5be5fc27"). InnerVolumeSpecName "kube-api-access-t84lv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:41:46 crc kubenswrapper[4631]: I1128 13:41:46.128259 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0339440e-8081-4f68-9c6f-b62e5be5fc27-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0339440e-8081-4f68-9c6f-b62e5be5fc27" (UID: "0339440e-8081-4f68-9c6f-b62e5be5fc27"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:41:46 crc kubenswrapper[4631]: I1128 13:41:46.131050 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0339440e-8081-4f68-9c6f-b62e5be5fc27-config-data" (OuterVolumeSpecName: "config-data") pod "0339440e-8081-4f68-9c6f-b62e5be5fc27" (UID: "0339440e-8081-4f68-9c6f-b62e5be5fc27"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:41:46 crc kubenswrapper[4631]: I1128 13:41:46.194515 4631 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0339440e-8081-4f68-9c6f-b62e5be5fc27-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 13:41:46 crc kubenswrapper[4631]: I1128 13:41:46.194548 4631 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0339440e-8081-4f68-9c6f-b62e5be5fc27-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 13:41:46 crc kubenswrapper[4631]: I1128 13:41:46.194562 4631 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0339440e-8081-4f68-9c6f-b62e5be5fc27-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 13:41:46 crc kubenswrapper[4631]: I1128 13:41:46.194573 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t84lv\" (UniqueName: \"kubernetes.io/projected/0339440e-8081-4f68-9c6f-b62e5be5fc27-kube-api-access-t84lv\") on node \"crc\" DevicePath \"\"" Nov 28 13:41:46 crc kubenswrapper[4631]: I1128 13:41:46.655948 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-24v2l" event={"ID":"0339440e-8081-4f68-9c6f-b62e5be5fc27","Type":"ContainerDied","Data":"3a3a9d3fb8a28758700544881d4c314af2c0499207773ff730686faef51f5036"} Nov 28 13:41:46 crc kubenswrapper[4631]: I1128 13:41:46.657152 4631 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3a3a9d3fb8a28758700544881d4c314af2c0499207773ff730686faef51f5036" Nov 28 13:41:46 crc kubenswrapper[4631]: I1128 13:41:46.656140 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-24v2l" Nov 28 13:41:46 crc kubenswrapper[4631]: I1128 13:41:46.862763 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Nov 28 13:41:46 crc kubenswrapper[4631]: I1128 13:41:46.863193 4631 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="a52eb82e-ac7b-4a48-9021-4c450da3ecae" containerName="nova-scheduler-scheduler" containerID="cri-o://9806c18b8805bf90bd761a4c5a6a35fa5f6a85fabc73ecead82d8bd9898ac626" gracePeriod=30 Nov 28 13:41:46 crc kubenswrapper[4631]: I1128 13:41:46.885269 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 28 13:41:46 crc kubenswrapper[4631]: I1128 13:41:46.885678 4631 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="122d85b6-1e49-4493-87be-34104a78a04a" containerName="nova-api-log" containerID="cri-o://734b5c7339b19dfeb770c6c153a659a1c25016ba4f54511557ef97d7267a8dd1" gracePeriod=30 Nov 28 13:41:46 crc kubenswrapper[4631]: I1128 13:41:46.885833 4631 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="122d85b6-1e49-4493-87be-34104a78a04a" containerName="nova-api-api" containerID="cri-o://2e12690f233ccdc38fedfc24a0d67e900ed57e8bbe390c331bca481340fdb398" gracePeriod=30 Nov 28 13:41:46 crc kubenswrapper[4631]: I1128 13:41:46.917657 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 28 13:41:46 crc kubenswrapper[4631]: I1128 13:41:46.917968 4631 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="fb5b797e-a328-4f2e-a239-333213537f2a" containerName="nova-metadata-log" containerID="cri-o://3932276d04ef4e1355b3d36c0f83cfb37b4724d2af10ecd90370a7c5ae212aa6" gracePeriod=30 Nov 28 13:41:46 crc kubenswrapper[4631]: I1128 13:41:46.918599 4631 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="fb5b797e-a328-4f2e-a239-333213537f2a" containerName="nova-metadata-metadata" containerID="cri-o://6b05c3bfd6fcbcd36095c9145b56050d4042f642dc03e2d34a74e87996ecbe6b" gracePeriod=30 Nov 28 13:41:47 crc kubenswrapper[4631]: I1128 13:41:47.646532 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 28 13:41:47 crc kubenswrapper[4631]: I1128 13:41:47.694243 4631 generic.go:334] "Generic (PLEG): container finished" podID="fb5b797e-a328-4f2e-a239-333213537f2a" containerID="3932276d04ef4e1355b3d36c0f83cfb37b4724d2af10ecd90370a7c5ae212aa6" exitCode=143 Nov 28 13:41:47 crc kubenswrapper[4631]: I1128 13:41:47.694451 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"fb5b797e-a328-4f2e-a239-333213537f2a","Type":"ContainerDied","Data":"3932276d04ef4e1355b3d36c0f83cfb37b4724d2af10ecd90370a7c5ae212aa6"} Nov 28 13:41:47 crc kubenswrapper[4631]: I1128 13:41:47.713449 4631 generic.go:334] "Generic (PLEG): container finished" podID="122d85b6-1e49-4493-87be-34104a78a04a" containerID="2e12690f233ccdc38fedfc24a0d67e900ed57e8bbe390c331bca481340fdb398" exitCode=0 Nov 28 13:41:47 crc kubenswrapper[4631]: I1128 13:41:47.713529 4631 generic.go:334] "Generic (PLEG): container finished" podID="122d85b6-1e49-4493-87be-34104a78a04a" containerID="734b5c7339b19dfeb770c6c153a659a1c25016ba4f54511557ef97d7267a8dd1" exitCode=143 Nov 28 13:41:47 crc kubenswrapper[4631]: I1128 13:41:47.713574 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"122d85b6-1e49-4493-87be-34104a78a04a","Type":"ContainerDied","Data":"2e12690f233ccdc38fedfc24a0d67e900ed57e8bbe390c331bca481340fdb398"} Nov 28 13:41:47 crc kubenswrapper[4631]: I1128 13:41:47.713610 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"122d85b6-1e49-4493-87be-34104a78a04a","Type":"ContainerDied","Data":"734b5c7339b19dfeb770c6c153a659a1c25016ba4f54511557ef97d7267a8dd1"} Nov 28 13:41:47 crc kubenswrapper[4631]: I1128 13:41:47.713625 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"122d85b6-1e49-4493-87be-34104a78a04a","Type":"ContainerDied","Data":"b24df505c8b56b8c21762b4c7825f6f3537467020acbab505e024281509d1f2b"} Nov 28 13:41:47 crc kubenswrapper[4631]: I1128 13:41:47.713644 4631 scope.go:117] "RemoveContainer" containerID="2e12690f233ccdc38fedfc24a0d67e900ed57e8bbe390c331bca481340fdb398" Nov 28 13:41:47 crc kubenswrapper[4631]: I1128 13:41:47.713884 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 28 13:41:47 crc kubenswrapper[4631]: I1128 13:41:47.740416 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/122d85b6-1e49-4493-87be-34104a78a04a-combined-ca-bundle\") pod \"122d85b6-1e49-4493-87be-34104a78a04a\" (UID: \"122d85b6-1e49-4493-87be-34104a78a04a\") " Nov 28 13:41:47 crc kubenswrapper[4631]: I1128 13:41:47.740501 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/122d85b6-1e49-4493-87be-34104a78a04a-config-data\") pod \"122d85b6-1e49-4493-87be-34104a78a04a\" (UID: \"122d85b6-1e49-4493-87be-34104a78a04a\") " Nov 28 13:41:47 crc kubenswrapper[4631]: I1128 13:41:47.740680 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/122d85b6-1e49-4493-87be-34104a78a04a-logs\") pod \"122d85b6-1e49-4493-87be-34104a78a04a\" (UID: \"122d85b6-1e49-4493-87be-34104a78a04a\") " Nov 28 13:41:47 crc kubenswrapper[4631]: I1128 13:41:47.740710 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/122d85b6-1e49-4493-87be-34104a78a04a-public-tls-certs\") pod \"122d85b6-1e49-4493-87be-34104a78a04a\" (UID: \"122d85b6-1e49-4493-87be-34104a78a04a\") " Nov 28 13:41:47 crc kubenswrapper[4631]: I1128 13:41:47.740743 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/122d85b6-1e49-4493-87be-34104a78a04a-internal-tls-certs\") pod \"122d85b6-1e49-4493-87be-34104a78a04a\" (UID: \"122d85b6-1e49-4493-87be-34104a78a04a\") " Nov 28 13:41:47 crc kubenswrapper[4631]: I1128 13:41:47.740800 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2pkdd\" (UniqueName: \"kubernetes.io/projected/122d85b6-1e49-4493-87be-34104a78a04a-kube-api-access-2pkdd\") pod \"122d85b6-1e49-4493-87be-34104a78a04a\" (UID: \"122d85b6-1e49-4493-87be-34104a78a04a\") " Nov 28 13:41:47 crc kubenswrapper[4631]: I1128 13:41:47.742246 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/122d85b6-1e49-4493-87be-34104a78a04a-logs" (OuterVolumeSpecName: "logs") pod "122d85b6-1e49-4493-87be-34104a78a04a" (UID: "122d85b6-1e49-4493-87be-34104a78a04a"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 13:41:47 crc kubenswrapper[4631]: I1128 13:41:47.782588 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/122d85b6-1e49-4493-87be-34104a78a04a-kube-api-access-2pkdd" (OuterVolumeSpecName: "kube-api-access-2pkdd") pod "122d85b6-1e49-4493-87be-34104a78a04a" (UID: "122d85b6-1e49-4493-87be-34104a78a04a"). InnerVolumeSpecName "kube-api-access-2pkdd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:41:47 crc kubenswrapper[4631]: I1128 13:41:47.819672 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/122d85b6-1e49-4493-87be-34104a78a04a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "122d85b6-1e49-4493-87be-34104a78a04a" (UID: "122d85b6-1e49-4493-87be-34104a78a04a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:41:47 crc kubenswrapper[4631]: I1128 13:41:47.821918 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/122d85b6-1e49-4493-87be-34104a78a04a-config-data" (OuterVolumeSpecName: "config-data") pod "122d85b6-1e49-4493-87be-34104a78a04a" (UID: "122d85b6-1e49-4493-87be-34104a78a04a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:41:47 crc kubenswrapper[4631]: I1128 13:41:47.843815 4631 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/122d85b6-1e49-4493-87be-34104a78a04a-logs\") on node \"crc\" DevicePath \"\"" Nov 28 13:41:47 crc kubenswrapper[4631]: I1128 13:41:47.843869 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2pkdd\" (UniqueName: \"kubernetes.io/projected/122d85b6-1e49-4493-87be-34104a78a04a-kube-api-access-2pkdd\") on node \"crc\" DevicePath \"\"" Nov 28 13:41:47 crc kubenswrapper[4631]: I1128 13:41:47.843880 4631 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/122d85b6-1e49-4493-87be-34104a78a04a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 13:41:47 crc kubenswrapper[4631]: I1128 13:41:47.843889 4631 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/122d85b6-1e49-4493-87be-34104a78a04a-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 13:41:47 crc kubenswrapper[4631]: I1128 13:41:47.851490 4631 scope.go:117] "RemoveContainer" containerID="734b5c7339b19dfeb770c6c153a659a1c25016ba4f54511557ef97d7267a8dd1" Nov 28 13:41:47 crc kubenswrapper[4631]: I1128 13:41:47.895769 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/122d85b6-1e49-4493-87be-34104a78a04a-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "122d85b6-1e49-4493-87be-34104a78a04a" (UID: "122d85b6-1e49-4493-87be-34104a78a04a"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:41:47 crc kubenswrapper[4631]: I1128 13:41:47.928456 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/122d85b6-1e49-4493-87be-34104a78a04a-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "122d85b6-1e49-4493-87be-34104a78a04a" (UID: "122d85b6-1e49-4493-87be-34104a78a04a"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:41:47 crc kubenswrapper[4631]: I1128 13:41:47.929754 4631 scope.go:117] "RemoveContainer" containerID="2e12690f233ccdc38fedfc24a0d67e900ed57e8bbe390c331bca481340fdb398" Nov 28 13:41:47 crc kubenswrapper[4631]: E1128 13:41:47.942941 4631 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2e12690f233ccdc38fedfc24a0d67e900ed57e8bbe390c331bca481340fdb398\": container with ID starting with 2e12690f233ccdc38fedfc24a0d67e900ed57e8bbe390c331bca481340fdb398 not found: ID does not exist" containerID="2e12690f233ccdc38fedfc24a0d67e900ed57e8bbe390c331bca481340fdb398" Nov 28 13:41:47 crc kubenswrapper[4631]: I1128 13:41:47.943022 4631 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2e12690f233ccdc38fedfc24a0d67e900ed57e8bbe390c331bca481340fdb398"} err="failed to get container status \"2e12690f233ccdc38fedfc24a0d67e900ed57e8bbe390c331bca481340fdb398\": rpc error: code = NotFound desc = could not find container \"2e12690f233ccdc38fedfc24a0d67e900ed57e8bbe390c331bca481340fdb398\": container with ID starting with 2e12690f233ccdc38fedfc24a0d67e900ed57e8bbe390c331bca481340fdb398 not found: ID does not exist" Nov 28 13:41:47 crc kubenswrapper[4631]: I1128 13:41:47.943070 4631 scope.go:117] "RemoveContainer" containerID="734b5c7339b19dfeb770c6c153a659a1c25016ba4f54511557ef97d7267a8dd1" Nov 28 13:41:47 crc kubenswrapper[4631]: I1128 13:41:47.947795 4631 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/122d85b6-1e49-4493-87be-34104a78a04a-public-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 28 13:41:47 crc kubenswrapper[4631]: I1128 13:41:47.947847 4631 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/122d85b6-1e49-4493-87be-34104a78a04a-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 28 13:41:47 crc kubenswrapper[4631]: E1128 13:41:47.951777 4631 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"734b5c7339b19dfeb770c6c153a659a1c25016ba4f54511557ef97d7267a8dd1\": container with ID starting with 734b5c7339b19dfeb770c6c153a659a1c25016ba4f54511557ef97d7267a8dd1 not found: ID does not exist" containerID="734b5c7339b19dfeb770c6c153a659a1c25016ba4f54511557ef97d7267a8dd1" Nov 28 13:41:47 crc kubenswrapper[4631]: I1128 13:41:47.951823 4631 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"734b5c7339b19dfeb770c6c153a659a1c25016ba4f54511557ef97d7267a8dd1"} err="failed to get container status \"734b5c7339b19dfeb770c6c153a659a1c25016ba4f54511557ef97d7267a8dd1\": rpc error: code = NotFound desc = could not find container \"734b5c7339b19dfeb770c6c153a659a1c25016ba4f54511557ef97d7267a8dd1\": container with ID starting with 734b5c7339b19dfeb770c6c153a659a1c25016ba4f54511557ef97d7267a8dd1 not found: ID does not exist" Nov 28 13:41:47 crc kubenswrapper[4631]: I1128 13:41:47.951863 4631 scope.go:117] "RemoveContainer" containerID="2e12690f233ccdc38fedfc24a0d67e900ed57e8bbe390c331bca481340fdb398" Nov 28 13:41:47 crc kubenswrapper[4631]: I1128 13:41:47.955657 4631 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2e12690f233ccdc38fedfc24a0d67e900ed57e8bbe390c331bca481340fdb398"} err="failed to get container status \"2e12690f233ccdc38fedfc24a0d67e900ed57e8bbe390c331bca481340fdb398\": rpc error: code = NotFound desc = could not find container \"2e12690f233ccdc38fedfc24a0d67e900ed57e8bbe390c331bca481340fdb398\": container with ID starting with 2e12690f233ccdc38fedfc24a0d67e900ed57e8bbe390c331bca481340fdb398 not found: ID does not exist" Nov 28 13:41:47 crc kubenswrapper[4631]: I1128 13:41:47.955693 4631 scope.go:117] "RemoveContainer" containerID="734b5c7339b19dfeb770c6c153a659a1c25016ba4f54511557ef97d7267a8dd1" Nov 28 13:41:47 crc kubenswrapper[4631]: I1128 13:41:47.959580 4631 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"734b5c7339b19dfeb770c6c153a659a1c25016ba4f54511557ef97d7267a8dd1"} err="failed to get container status \"734b5c7339b19dfeb770c6c153a659a1c25016ba4f54511557ef97d7267a8dd1\": rpc error: code = NotFound desc = could not find container \"734b5c7339b19dfeb770c6c153a659a1c25016ba4f54511557ef97d7267a8dd1\": container with ID starting with 734b5c7339b19dfeb770c6c153a659a1c25016ba4f54511557ef97d7267a8dd1 not found: ID does not exist" Nov 28 13:41:48 crc kubenswrapper[4631]: I1128 13:41:48.080325 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 28 13:41:48 crc kubenswrapper[4631]: I1128 13:41:48.097721 4631 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Nov 28 13:41:48 crc kubenswrapper[4631]: I1128 13:41:48.123948 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Nov 28 13:41:48 crc kubenswrapper[4631]: E1128 13:41:48.124734 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e1644d33-bda9-437e-b10c-5c48042cb1f2" containerName="init" Nov 28 13:41:48 crc kubenswrapper[4631]: I1128 13:41:48.124757 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="e1644d33-bda9-437e-b10c-5c48042cb1f2" containerName="init" Nov 28 13:41:48 crc kubenswrapper[4631]: E1128 13:41:48.124784 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="122d85b6-1e49-4493-87be-34104a78a04a" containerName="nova-api-api" Nov 28 13:41:48 crc kubenswrapper[4631]: I1128 13:41:48.124791 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="122d85b6-1e49-4493-87be-34104a78a04a" containerName="nova-api-api" Nov 28 13:41:48 crc kubenswrapper[4631]: E1128 13:41:48.124815 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e1644d33-bda9-437e-b10c-5c48042cb1f2" containerName="dnsmasq-dns" Nov 28 13:41:48 crc kubenswrapper[4631]: I1128 13:41:48.124822 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="e1644d33-bda9-437e-b10c-5c48042cb1f2" containerName="dnsmasq-dns" Nov 28 13:41:48 crc kubenswrapper[4631]: E1128 13:41:48.124902 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="122d85b6-1e49-4493-87be-34104a78a04a" containerName="nova-api-log" Nov 28 13:41:48 crc kubenswrapper[4631]: I1128 13:41:48.124911 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="122d85b6-1e49-4493-87be-34104a78a04a" containerName="nova-api-log" Nov 28 13:41:48 crc kubenswrapper[4631]: E1128 13:41:48.124928 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0339440e-8081-4f68-9c6f-b62e5be5fc27" containerName="nova-manage" Nov 28 13:41:48 crc kubenswrapper[4631]: I1128 13:41:48.124934 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="0339440e-8081-4f68-9c6f-b62e5be5fc27" containerName="nova-manage" Nov 28 13:41:48 crc kubenswrapper[4631]: I1128 13:41:48.125191 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="0339440e-8081-4f68-9c6f-b62e5be5fc27" containerName="nova-manage" Nov 28 13:41:48 crc kubenswrapper[4631]: I1128 13:41:48.125210 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="e1644d33-bda9-437e-b10c-5c48042cb1f2" containerName="dnsmasq-dns" Nov 28 13:41:48 crc kubenswrapper[4631]: I1128 13:41:48.125231 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="122d85b6-1e49-4493-87be-34104a78a04a" containerName="nova-api-log" Nov 28 13:41:48 crc kubenswrapper[4631]: I1128 13:41:48.125252 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="122d85b6-1e49-4493-87be-34104a78a04a" containerName="nova-api-api" Nov 28 13:41:48 crc kubenswrapper[4631]: I1128 13:41:48.126931 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 28 13:41:48 crc kubenswrapper[4631]: I1128 13:41:48.139072 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Nov 28 13:41:48 crc kubenswrapper[4631]: I1128 13:41:48.140123 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Nov 28 13:41:48 crc kubenswrapper[4631]: I1128 13:41:48.140753 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Nov 28 13:41:48 crc kubenswrapper[4631]: I1128 13:41:48.232221 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 28 13:41:48 crc kubenswrapper[4631]: I1128 13:41:48.294177 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9db3c70b-08ec-4b40-bdb9-94cdbc66fa1d-logs\") pod \"nova-api-0\" (UID: \"9db3c70b-08ec-4b40-bdb9-94cdbc66fa1d\") " pod="openstack/nova-api-0" Nov 28 13:41:48 crc kubenswrapper[4631]: I1128 13:41:48.294452 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9db3c70b-08ec-4b40-bdb9-94cdbc66fa1d-public-tls-certs\") pod \"nova-api-0\" (UID: \"9db3c70b-08ec-4b40-bdb9-94cdbc66fa1d\") " pod="openstack/nova-api-0" Nov 28 13:41:48 crc kubenswrapper[4631]: I1128 13:41:48.294601 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9db3c70b-08ec-4b40-bdb9-94cdbc66fa1d-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"9db3c70b-08ec-4b40-bdb9-94cdbc66fa1d\") " pod="openstack/nova-api-0" Nov 28 13:41:48 crc kubenswrapper[4631]: I1128 13:41:48.294706 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9db3c70b-08ec-4b40-bdb9-94cdbc66fa1d-internal-tls-certs\") pod \"nova-api-0\" (UID: \"9db3c70b-08ec-4b40-bdb9-94cdbc66fa1d\") " pod="openstack/nova-api-0" Nov 28 13:41:48 crc kubenswrapper[4631]: I1128 13:41:48.294997 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6qcs8\" (UniqueName: \"kubernetes.io/projected/9db3c70b-08ec-4b40-bdb9-94cdbc66fa1d-kube-api-access-6qcs8\") pod \"nova-api-0\" (UID: \"9db3c70b-08ec-4b40-bdb9-94cdbc66fa1d\") " pod="openstack/nova-api-0" Nov 28 13:41:48 crc kubenswrapper[4631]: I1128 13:41:48.295412 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9db3c70b-08ec-4b40-bdb9-94cdbc66fa1d-config-data\") pod \"nova-api-0\" (UID: \"9db3c70b-08ec-4b40-bdb9-94cdbc66fa1d\") " pod="openstack/nova-api-0" Nov 28 13:41:48 crc kubenswrapper[4631]: I1128 13:41:48.398610 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6qcs8\" (UniqueName: \"kubernetes.io/projected/9db3c70b-08ec-4b40-bdb9-94cdbc66fa1d-kube-api-access-6qcs8\") pod \"nova-api-0\" (UID: \"9db3c70b-08ec-4b40-bdb9-94cdbc66fa1d\") " pod="openstack/nova-api-0" Nov 28 13:41:48 crc kubenswrapper[4631]: I1128 13:41:48.398710 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9db3c70b-08ec-4b40-bdb9-94cdbc66fa1d-config-data\") pod \"nova-api-0\" (UID: \"9db3c70b-08ec-4b40-bdb9-94cdbc66fa1d\") " pod="openstack/nova-api-0" Nov 28 13:41:48 crc kubenswrapper[4631]: I1128 13:41:48.398759 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9db3c70b-08ec-4b40-bdb9-94cdbc66fa1d-logs\") pod \"nova-api-0\" (UID: \"9db3c70b-08ec-4b40-bdb9-94cdbc66fa1d\") " pod="openstack/nova-api-0" Nov 28 13:41:48 crc kubenswrapper[4631]: I1128 13:41:48.398784 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9db3c70b-08ec-4b40-bdb9-94cdbc66fa1d-public-tls-certs\") pod \"nova-api-0\" (UID: \"9db3c70b-08ec-4b40-bdb9-94cdbc66fa1d\") " pod="openstack/nova-api-0" Nov 28 13:41:48 crc kubenswrapper[4631]: I1128 13:41:48.398854 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9db3c70b-08ec-4b40-bdb9-94cdbc66fa1d-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"9db3c70b-08ec-4b40-bdb9-94cdbc66fa1d\") " pod="openstack/nova-api-0" Nov 28 13:41:48 crc kubenswrapper[4631]: I1128 13:41:48.398888 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9db3c70b-08ec-4b40-bdb9-94cdbc66fa1d-internal-tls-certs\") pod \"nova-api-0\" (UID: \"9db3c70b-08ec-4b40-bdb9-94cdbc66fa1d\") " pod="openstack/nova-api-0" Nov 28 13:41:48 crc kubenswrapper[4631]: I1128 13:41:48.399369 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9db3c70b-08ec-4b40-bdb9-94cdbc66fa1d-logs\") pod \"nova-api-0\" (UID: \"9db3c70b-08ec-4b40-bdb9-94cdbc66fa1d\") " pod="openstack/nova-api-0" Nov 28 13:41:48 crc kubenswrapper[4631]: I1128 13:41:48.407038 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9db3c70b-08ec-4b40-bdb9-94cdbc66fa1d-config-data\") pod \"nova-api-0\" (UID: \"9db3c70b-08ec-4b40-bdb9-94cdbc66fa1d\") " pod="openstack/nova-api-0" Nov 28 13:41:48 crc kubenswrapper[4631]: I1128 13:41:48.415897 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9db3c70b-08ec-4b40-bdb9-94cdbc66fa1d-internal-tls-certs\") pod \"nova-api-0\" (UID: \"9db3c70b-08ec-4b40-bdb9-94cdbc66fa1d\") " pod="openstack/nova-api-0" Nov 28 13:41:48 crc kubenswrapper[4631]: I1128 13:41:48.419485 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9db3c70b-08ec-4b40-bdb9-94cdbc66fa1d-public-tls-certs\") pod \"nova-api-0\" (UID: \"9db3c70b-08ec-4b40-bdb9-94cdbc66fa1d\") " pod="openstack/nova-api-0" Nov 28 13:41:48 crc kubenswrapper[4631]: I1128 13:41:48.422597 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9db3c70b-08ec-4b40-bdb9-94cdbc66fa1d-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"9db3c70b-08ec-4b40-bdb9-94cdbc66fa1d\") " pod="openstack/nova-api-0" Nov 28 13:41:48 crc kubenswrapper[4631]: I1128 13:41:48.427161 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6qcs8\" (UniqueName: \"kubernetes.io/projected/9db3c70b-08ec-4b40-bdb9-94cdbc66fa1d-kube-api-access-6qcs8\") pod \"nova-api-0\" (UID: \"9db3c70b-08ec-4b40-bdb9-94cdbc66fa1d\") " pod="openstack/nova-api-0" Nov 28 13:41:48 crc kubenswrapper[4631]: I1128 13:41:48.533193 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 28 13:41:48 crc kubenswrapper[4631]: E1128 13:41:48.534942 4631 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 9806c18b8805bf90bd761a4c5a6a35fa5f6a85fabc73ecead82d8bd9898ac626 is running failed: container process not found" containerID="9806c18b8805bf90bd761a4c5a6a35fa5f6a85fabc73ecead82d8bd9898ac626" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Nov 28 13:41:48 crc kubenswrapper[4631]: E1128 13:41:48.535205 4631 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 9806c18b8805bf90bd761a4c5a6a35fa5f6a85fabc73ecead82d8bd9898ac626 is running failed: container process not found" containerID="9806c18b8805bf90bd761a4c5a6a35fa5f6a85fabc73ecead82d8bd9898ac626" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Nov 28 13:41:48 crc kubenswrapper[4631]: E1128 13:41:48.535387 4631 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 9806c18b8805bf90bd761a4c5a6a35fa5f6a85fabc73ecead82d8bd9898ac626 is running failed: container process not found" containerID="9806c18b8805bf90bd761a4c5a6a35fa5f6a85fabc73ecead82d8bd9898ac626" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Nov 28 13:41:48 crc kubenswrapper[4631]: E1128 13:41:48.535411 4631 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 9806c18b8805bf90bd761a4c5a6a35fa5f6a85fabc73ecead82d8bd9898ac626 is running failed: container process not found" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="a52eb82e-ac7b-4a48-9021-4c450da3ecae" containerName="nova-scheduler-scheduler" Nov 28 13:41:48 crc kubenswrapper[4631]: I1128 13:41:48.719199 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 28 13:41:48 crc kubenswrapper[4631]: I1128 13:41:48.764034 4631 generic.go:334] "Generic (PLEG): container finished" podID="a52eb82e-ac7b-4a48-9021-4c450da3ecae" containerID="9806c18b8805bf90bd761a4c5a6a35fa5f6a85fabc73ecead82d8bd9898ac626" exitCode=0 Nov 28 13:41:48 crc kubenswrapper[4631]: I1128 13:41:48.764591 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"a52eb82e-ac7b-4a48-9021-4c450da3ecae","Type":"ContainerDied","Data":"9806c18b8805bf90bd761a4c5a6a35fa5f6a85fabc73ecead82d8bd9898ac626"} Nov 28 13:41:48 crc kubenswrapper[4631]: I1128 13:41:48.764628 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"a52eb82e-ac7b-4a48-9021-4c450da3ecae","Type":"ContainerDied","Data":"398d71554c3839952b9b273209cc0a7a898e31e280d5458196d0361360d1bdf9"} Nov 28 13:41:48 crc kubenswrapper[4631]: I1128 13:41:48.764648 4631 scope.go:117] "RemoveContainer" containerID="9806c18b8805bf90bd761a4c5a6a35fa5f6a85fabc73ecead82d8bd9898ac626" Nov 28 13:41:48 crc kubenswrapper[4631]: I1128 13:41:48.764901 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 28 13:41:48 crc kubenswrapper[4631]: I1128 13:41:48.807073 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a52eb82e-ac7b-4a48-9021-4c450da3ecae-config-data\") pod \"a52eb82e-ac7b-4a48-9021-4c450da3ecae\" (UID: \"a52eb82e-ac7b-4a48-9021-4c450da3ecae\") " Nov 28 13:41:48 crc kubenswrapper[4631]: I1128 13:41:48.807467 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a52eb82e-ac7b-4a48-9021-4c450da3ecae-combined-ca-bundle\") pod \"a52eb82e-ac7b-4a48-9021-4c450da3ecae\" (UID: \"a52eb82e-ac7b-4a48-9021-4c450da3ecae\") " Nov 28 13:41:48 crc kubenswrapper[4631]: I1128 13:41:48.807565 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8gjg5\" (UniqueName: \"kubernetes.io/projected/a52eb82e-ac7b-4a48-9021-4c450da3ecae-kube-api-access-8gjg5\") pod \"a52eb82e-ac7b-4a48-9021-4c450da3ecae\" (UID: \"a52eb82e-ac7b-4a48-9021-4c450da3ecae\") " Nov 28 13:41:48 crc kubenswrapper[4631]: I1128 13:41:48.809538 4631 scope.go:117] "RemoveContainer" containerID="9806c18b8805bf90bd761a4c5a6a35fa5f6a85fabc73ecead82d8bd9898ac626" Nov 28 13:41:48 crc kubenswrapper[4631]: E1128 13:41:48.813610 4631 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9806c18b8805bf90bd761a4c5a6a35fa5f6a85fabc73ecead82d8bd9898ac626\": container with ID starting with 9806c18b8805bf90bd761a4c5a6a35fa5f6a85fabc73ecead82d8bd9898ac626 not found: ID does not exist" containerID="9806c18b8805bf90bd761a4c5a6a35fa5f6a85fabc73ecead82d8bd9898ac626" Nov 28 13:41:48 crc kubenswrapper[4631]: I1128 13:41:48.813670 4631 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9806c18b8805bf90bd761a4c5a6a35fa5f6a85fabc73ecead82d8bd9898ac626"} err="failed to get container status \"9806c18b8805bf90bd761a4c5a6a35fa5f6a85fabc73ecead82d8bd9898ac626\": rpc error: code = NotFound desc = could not find container \"9806c18b8805bf90bd761a4c5a6a35fa5f6a85fabc73ecead82d8bd9898ac626\": container with ID starting with 9806c18b8805bf90bd761a4c5a6a35fa5f6a85fabc73ecead82d8bd9898ac626 not found: ID does not exist" Nov 28 13:41:48 crc kubenswrapper[4631]: I1128 13:41:48.814276 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a52eb82e-ac7b-4a48-9021-4c450da3ecae-kube-api-access-8gjg5" (OuterVolumeSpecName: "kube-api-access-8gjg5") pod "a52eb82e-ac7b-4a48-9021-4c450da3ecae" (UID: "a52eb82e-ac7b-4a48-9021-4c450da3ecae"). InnerVolumeSpecName "kube-api-access-8gjg5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:41:48 crc kubenswrapper[4631]: I1128 13:41:48.851971 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a52eb82e-ac7b-4a48-9021-4c450da3ecae-config-data" (OuterVolumeSpecName: "config-data") pod "a52eb82e-ac7b-4a48-9021-4c450da3ecae" (UID: "a52eb82e-ac7b-4a48-9021-4c450da3ecae"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:41:48 crc kubenswrapper[4631]: I1128 13:41:48.852444 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a52eb82e-ac7b-4a48-9021-4c450da3ecae-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a52eb82e-ac7b-4a48-9021-4c450da3ecae" (UID: "a52eb82e-ac7b-4a48-9021-4c450da3ecae"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:41:48 crc kubenswrapper[4631]: I1128 13:41:48.910465 4631 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a52eb82e-ac7b-4a48-9021-4c450da3ecae-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 13:41:48 crc kubenswrapper[4631]: I1128 13:41:48.910508 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8gjg5\" (UniqueName: \"kubernetes.io/projected/a52eb82e-ac7b-4a48-9021-4c450da3ecae-kube-api-access-8gjg5\") on node \"crc\" DevicePath \"\"" Nov 28 13:41:48 crc kubenswrapper[4631]: I1128 13:41:48.910524 4631 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a52eb82e-ac7b-4a48-9021-4c450da3ecae-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 13:41:49 crc kubenswrapper[4631]: I1128 13:41:49.110649 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Nov 28 13:41:49 crc kubenswrapper[4631]: I1128 13:41:49.138393 4631 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Nov 28 13:41:49 crc kubenswrapper[4631]: I1128 13:41:49.156308 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Nov 28 13:41:49 crc kubenswrapper[4631]: E1128 13:41:49.156996 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a52eb82e-ac7b-4a48-9021-4c450da3ecae" containerName="nova-scheduler-scheduler" Nov 28 13:41:49 crc kubenswrapper[4631]: I1128 13:41:49.157025 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="a52eb82e-ac7b-4a48-9021-4c450da3ecae" containerName="nova-scheduler-scheduler" Nov 28 13:41:49 crc kubenswrapper[4631]: I1128 13:41:49.157305 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="a52eb82e-ac7b-4a48-9021-4c450da3ecae" containerName="nova-scheduler-scheduler" Nov 28 13:41:49 crc kubenswrapper[4631]: I1128 13:41:49.158373 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 28 13:41:49 crc kubenswrapper[4631]: I1128 13:41:49.167423 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Nov 28 13:41:49 crc kubenswrapper[4631]: I1128 13:41:49.194485 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 28 13:41:49 crc kubenswrapper[4631]: I1128 13:41:49.206406 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 28 13:41:49 crc kubenswrapper[4631]: I1128 13:41:49.217113 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pnfzj\" (UniqueName: \"kubernetes.io/projected/83514cf7-b36a-4691-b5f9-559a1594044a-kube-api-access-pnfzj\") pod \"nova-scheduler-0\" (UID: \"83514cf7-b36a-4691-b5f9-559a1594044a\") " pod="openstack/nova-scheduler-0" Nov 28 13:41:49 crc kubenswrapper[4631]: I1128 13:41:49.217777 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/83514cf7-b36a-4691-b5f9-559a1594044a-config-data\") pod \"nova-scheduler-0\" (UID: \"83514cf7-b36a-4691-b5f9-559a1594044a\") " pod="openstack/nova-scheduler-0" Nov 28 13:41:49 crc kubenswrapper[4631]: I1128 13:41:49.217842 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/83514cf7-b36a-4691-b5f9-559a1594044a-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"83514cf7-b36a-4691-b5f9-559a1594044a\") " pod="openstack/nova-scheduler-0" Nov 28 13:41:49 crc kubenswrapper[4631]: I1128 13:41:49.320220 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/83514cf7-b36a-4691-b5f9-559a1594044a-config-data\") pod \"nova-scheduler-0\" (UID: \"83514cf7-b36a-4691-b5f9-559a1594044a\") " pod="openstack/nova-scheduler-0" Nov 28 13:41:49 crc kubenswrapper[4631]: I1128 13:41:49.320370 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/83514cf7-b36a-4691-b5f9-559a1594044a-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"83514cf7-b36a-4691-b5f9-559a1594044a\") " pod="openstack/nova-scheduler-0" Nov 28 13:41:49 crc kubenswrapper[4631]: I1128 13:41:49.320764 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pnfzj\" (UniqueName: \"kubernetes.io/projected/83514cf7-b36a-4691-b5f9-559a1594044a-kube-api-access-pnfzj\") pod \"nova-scheduler-0\" (UID: \"83514cf7-b36a-4691-b5f9-559a1594044a\") " pod="openstack/nova-scheduler-0" Nov 28 13:41:49 crc kubenswrapper[4631]: I1128 13:41:49.329926 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/83514cf7-b36a-4691-b5f9-559a1594044a-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"83514cf7-b36a-4691-b5f9-559a1594044a\") " pod="openstack/nova-scheduler-0" Nov 28 13:41:49 crc kubenswrapper[4631]: I1128 13:41:49.330132 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/83514cf7-b36a-4691-b5f9-559a1594044a-config-data\") pod \"nova-scheduler-0\" (UID: \"83514cf7-b36a-4691-b5f9-559a1594044a\") " pod="openstack/nova-scheduler-0" Nov 28 13:41:49 crc kubenswrapper[4631]: I1128 13:41:49.352864 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pnfzj\" (UniqueName: \"kubernetes.io/projected/83514cf7-b36a-4691-b5f9-559a1594044a-kube-api-access-pnfzj\") pod \"nova-scheduler-0\" (UID: \"83514cf7-b36a-4691-b5f9-559a1594044a\") " pod="openstack/nova-scheduler-0" Nov 28 13:41:49 crc kubenswrapper[4631]: I1128 13:41:49.512308 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 28 13:41:49 crc kubenswrapper[4631]: I1128 13:41:49.530413 4631 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="122d85b6-1e49-4493-87be-34104a78a04a" path="/var/lib/kubelet/pods/122d85b6-1e49-4493-87be-34104a78a04a/volumes" Nov 28 13:41:49 crc kubenswrapper[4631]: I1128 13:41:49.531087 4631 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a52eb82e-ac7b-4a48-9021-4c450da3ecae" path="/var/lib/kubelet/pods/a52eb82e-ac7b-4a48-9021-4c450da3ecae/volumes" Nov 28 13:41:49 crc kubenswrapper[4631]: I1128 13:41:49.808809 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"9db3c70b-08ec-4b40-bdb9-94cdbc66fa1d","Type":"ContainerStarted","Data":"f3b39c1949e4a352a1de56d1b4af46ad2a700065878fd50b409de7844cc0102b"} Nov 28 13:41:49 crc kubenswrapper[4631]: I1128 13:41:49.810021 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"9db3c70b-08ec-4b40-bdb9-94cdbc66fa1d","Type":"ContainerStarted","Data":"26df8fa046c522f41d08b0d38fac17a35847b35a6da73b50dd5fb07946b32c67"} Nov 28 13:41:50 crc kubenswrapper[4631]: I1128 13:41:50.031157 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 28 13:41:50 crc kubenswrapper[4631]: W1128 13:41:50.034802 4631 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod83514cf7_b36a_4691_b5f9_559a1594044a.slice/crio-1fdfd33730c10e1258b1ec5d0ee201e7132989bfe785307839ab663819f21c02 WatchSource:0}: Error finding container 1fdfd33730c10e1258b1ec5d0ee201e7132989bfe785307839ab663819f21c02: Status 404 returned error can't find the container with id 1fdfd33730c10e1258b1ec5d0ee201e7132989bfe785307839ab663819f21c02 Nov 28 13:41:50 crc kubenswrapper[4631]: I1128 13:41:50.330194 4631 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="fb5b797e-a328-4f2e-a239-333213537f2a" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.194:8775/\": dial tcp 10.217.0.194:8775: connect: connection refused" Nov 28 13:41:50 crc kubenswrapper[4631]: I1128 13:41:50.330214 4631 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="fb5b797e-a328-4f2e-a239-333213537f2a" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.194:8775/\": dial tcp 10.217.0.194:8775: connect: connection refused" Nov 28 13:41:50 crc kubenswrapper[4631]: E1128 13:41:50.439578 4631 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfb5b797e_a328_4f2e_a239_333213537f2a.slice/crio-conmon-6b05c3bfd6fcbcd36095c9145b56050d4042f642dc03e2d34a74e87996ecbe6b.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfb5b797e_a328_4f2e_a239_333213537f2a.slice/crio-6b05c3bfd6fcbcd36095c9145b56050d4042f642dc03e2d34a74e87996ecbe6b.scope\": RecentStats: unable to find data in memory cache]" Nov 28 13:41:50 crc kubenswrapper[4631]: I1128 13:41:50.574490 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 28 13:41:50 crc kubenswrapper[4631]: I1128 13:41:50.665469 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v5wjs\" (UniqueName: \"kubernetes.io/projected/fb5b797e-a328-4f2e-a239-333213537f2a-kube-api-access-v5wjs\") pod \"fb5b797e-a328-4f2e-a239-333213537f2a\" (UID: \"fb5b797e-a328-4f2e-a239-333213537f2a\") " Nov 28 13:41:50 crc kubenswrapper[4631]: I1128 13:41:50.665571 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fb5b797e-a328-4f2e-a239-333213537f2a-config-data\") pod \"fb5b797e-a328-4f2e-a239-333213537f2a\" (UID: \"fb5b797e-a328-4f2e-a239-333213537f2a\") " Nov 28 13:41:50 crc kubenswrapper[4631]: I1128 13:41:50.665671 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb5b797e-a328-4f2e-a239-333213537f2a-combined-ca-bundle\") pod \"fb5b797e-a328-4f2e-a239-333213537f2a\" (UID: \"fb5b797e-a328-4f2e-a239-333213537f2a\") " Nov 28 13:41:50 crc kubenswrapper[4631]: I1128 13:41:50.665718 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fb5b797e-a328-4f2e-a239-333213537f2a-logs\") pod \"fb5b797e-a328-4f2e-a239-333213537f2a\" (UID: \"fb5b797e-a328-4f2e-a239-333213537f2a\") " Nov 28 13:41:50 crc kubenswrapper[4631]: I1128 13:41:50.665875 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/fb5b797e-a328-4f2e-a239-333213537f2a-nova-metadata-tls-certs\") pod \"fb5b797e-a328-4f2e-a239-333213537f2a\" (UID: \"fb5b797e-a328-4f2e-a239-333213537f2a\") " Nov 28 13:41:50 crc kubenswrapper[4631]: I1128 13:41:50.667559 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fb5b797e-a328-4f2e-a239-333213537f2a-logs" (OuterVolumeSpecName: "logs") pod "fb5b797e-a328-4f2e-a239-333213537f2a" (UID: "fb5b797e-a328-4f2e-a239-333213537f2a"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 13:41:50 crc kubenswrapper[4631]: I1128 13:41:50.684229 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fb5b797e-a328-4f2e-a239-333213537f2a-kube-api-access-v5wjs" (OuterVolumeSpecName: "kube-api-access-v5wjs") pod "fb5b797e-a328-4f2e-a239-333213537f2a" (UID: "fb5b797e-a328-4f2e-a239-333213537f2a"). InnerVolumeSpecName "kube-api-access-v5wjs". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:41:50 crc kubenswrapper[4631]: I1128 13:41:50.709645 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fb5b797e-a328-4f2e-a239-333213537f2a-config-data" (OuterVolumeSpecName: "config-data") pod "fb5b797e-a328-4f2e-a239-333213537f2a" (UID: "fb5b797e-a328-4f2e-a239-333213537f2a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:41:50 crc kubenswrapper[4631]: I1128 13:41:50.713722 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fb5b797e-a328-4f2e-a239-333213537f2a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "fb5b797e-a328-4f2e-a239-333213537f2a" (UID: "fb5b797e-a328-4f2e-a239-333213537f2a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:41:50 crc kubenswrapper[4631]: I1128 13:41:50.759839 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fb5b797e-a328-4f2e-a239-333213537f2a-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "fb5b797e-a328-4f2e-a239-333213537f2a" (UID: "fb5b797e-a328-4f2e-a239-333213537f2a"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:41:50 crc kubenswrapper[4631]: I1128 13:41:50.769525 4631 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb5b797e-a328-4f2e-a239-333213537f2a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 13:41:50 crc kubenswrapper[4631]: I1128 13:41:50.769563 4631 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fb5b797e-a328-4f2e-a239-333213537f2a-logs\") on node \"crc\" DevicePath \"\"" Nov 28 13:41:50 crc kubenswrapper[4631]: I1128 13:41:50.769575 4631 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/fb5b797e-a328-4f2e-a239-333213537f2a-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 28 13:41:50 crc kubenswrapper[4631]: I1128 13:41:50.769588 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v5wjs\" (UniqueName: \"kubernetes.io/projected/fb5b797e-a328-4f2e-a239-333213537f2a-kube-api-access-v5wjs\") on node \"crc\" DevicePath \"\"" Nov 28 13:41:50 crc kubenswrapper[4631]: I1128 13:41:50.769599 4631 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fb5b797e-a328-4f2e-a239-333213537f2a-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 13:41:50 crc kubenswrapper[4631]: I1128 13:41:50.830938 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"9db3c70b-08ec-4b40-bdb9-94cdbc66fa1d","Type":"ContainerStarted","Data":"d2bcabed09bb6423f637c867e28983eba66ed1a67acd4bc16388cadab7331188"} Nov 28 13:41:50 crc kubenswrapper[4631]: I1128 13:41:50.870646 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"83514cf7-b36a-4691-b5f9-559a1594044a","Type":"ContainerStarted","Data":"26ce81da55711d7534dee9d81faea4f1de6586885248cc23d3d89d3bdb37d3fc"} Nov 28 13:41:50 crc kubenswrapper[4631]: I1128 13:41:50.870705 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"83514cf7-b36a-4691-b5f9-559a1594044a","Type":"ContainerStarted","Data":"1fdfd33730c10e1258b1ec5d0ee201e7132989bfe785307839ab663819f21c02"} Nov 28 13:41:50 crc kubenswrapper[4631]: I1128 13:41:50.883020 4631 generic.go:334] "Generic (PLEG): container finished" podID="fb5b797e-a328-4f2e-a239-333213537f2a" containerID="6b05c3bfd6fcbcd36095c9145b56050d4042f642dc03e2d34a74e87996ecbe6b" exitCode=0 Nov 28 13:41:50 crc kubenswrapper[4631]: I1128 13:41:50.883087 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"fb5b797e-a328-4f2e-a239-333213537f2a","Type":"ContainerDied","Data":"6b05c3bfd6fcbcd36095c9145b56050d4042f642dc03e2d34a74e87996ecbe6b"} Nov 28 13:41:50 crc kubenswrapper[4631]: I1128 13:41:50.883123 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"fb5b797e-a328-4f2e-a239-333213537f2a","Type":"ContainerDied","Data":"c8ba24d666dddd639a797bd7984d0503b1da79c33affc620b326fe5ae1efcdad"} Nov 28 13:41:50 crc kubenswrapper[4631]: I1128 13:41:50.883169 4631 scope.go:117] "RemoveContainer" containerID="6b05c3bfd6fcbcd36095c9145b56050d4042f642dc03e2d34a74e87996ecbe6b" Nov 28 13:41:50 crc kubenswrapper[4631]: I1128 13:41:50.883510 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 28 13:41:50 crc kubenswrapper[4631]: I1128 13:41:50.897831 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.897808085 podStartE2EDuration="2.897808085s" podCreationTimestamp="2025-11-28 13:41:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 13:41:50.859178156 +0000 UTC m=+1267.666481510" watchObservedRunningTime="2025-11-28 13:41:50.897808085 +0000 UTC m=+1267.705111429" Nov 28 13:41:50 crc kubenswrapper[4631]: I1128 13:41:50.925917 4631 scope.go:117] "RemoveContainer" containerID="3932276d04ef4e1355b3d36c0f83cfb37b4724d2af10ecd90370a7c5ae212aa6" Nov 28 13:41:50 crc kubenswrapper[4631]: I1128 13:41:50.926945 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=1.9269134110000001 podStartE2EDuration="1.926913411s" podCreationTimestamp="2025-11-28 13:41:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 13:41:50.894420226 +0000 UTC m=+1267.701723590" watchObservedRunningTime="2025-11-28 13:41:50.926913411 +0000 UTC m=+1267.734216755" Nov 28 13:41:50 crc kubenswrapper[4631]: I1128 13:41:50.963042 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 28 13:41:50 crc kubenswrapper[4631]: I1128 13:41:50.970537 4631 scope.go:117] "RemoveContainer" containerID="6b05c3bfd6fcbcd36095c9145b56050d4042f642dc03e2d34a74e87996ecbe6b" Nov 28 13:41:50 crc kubenswrapper[4631]: E1128 13:41:50.972800 4631 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6b05c3bfd6fcbcd36095c9145b56050d4042f642dc03e2d34a74e87996ecbe6b\": container with ID starting with 6b05c3bfd6fcbcd36095c9145b56050d4042f642dc03e2d34a74e87996ecbe6b not found: ID does not exist" containerID="6b05c3bfd6fcbcd36095c9145b56050d4042f642dc03e2d34a74e87996ecbe6b" Nov 28 13:41:50 crc kubenswrapper[4631]: I1128 13:41:50.972843 4631 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6b05c3bfd6fcbcd36095c9145b56050d4042f642dc03e2d34a74e87996ecbe6b"} err="failed to get container status \"6b05c3bfd6fcbcd36095c9145b56050d4042f642dc03e2d34a74e87996ecbe6b\": rpc error: code = NotFound desc = could not find container \"6b05c3bfd6fcbcd36095c9145b56050d4042f642dc03e2d34a74e87996ecbe6b\": container with ID starting with 6b05c3bfd6fcbcd36095c9145b56050d4042f642dc03e2d34a74e87996ecbe6b not found: ID does not exist" Nov 28 13:41:50 crc kubenswrapper[4631]: I1128 13:41:50.972875 4631 scope.go:117] "RemoveContainer" containerID="3932276d04ef4e1355b3d36c0f83cfb37b4724d2af10ecd90370a7c5ae212aa6" Nov 28 13:41:50 crc kubenswrapper[4631]: E1128 13:41:50.974579 4631 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3932276d04ef4e1355b3d36c0f83cfb37b4724d2af10ecd90370a7c5ae212aa6\": container with ID starting with 3932276d04ef4e1355b3d36c0f83cfb37b4724d2af10ecd90370a7c5ae212aa6 not found: ID does not exist" containerID="3932276d04ef4e1355b3d36c0f83cfb37b4724d2af10ecd90370a7c5ae212aa6" Nov 28 13:41:50 crc kubenswrapper[4631]: I1128 13:41:50.974635 4631 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3932276d04ef4e1355b3d36c0f83cfb37b4724d2af10ecd90370a7c5ae212aa6"} err="failed to get container status \"3932276d04ef4e1355b3d36c0f83cfb37b4724d2af10ecd90370a7c5ae212aa6\": rpc error: code = NotFound desc = could not find container \"3932276d04ef4e1355b3d36c0f83cfb37b4724d2af10ecd90370a7c5ae212aa6\": container with ID starting with 3932276d04ef4e1355b3d36c0f83cfb37b4724d2af10ecd90370a7c5ae212aa6 not found: ID does not exist" Nov 28 13:41:50 crc kubenswrapper[4631]: I1128 13:41:50.986557 4631 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Nov 28 13:41:50 crc kubenswrapper[4631]: I1128 13:41:50.998539 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Nov 28 13:41:50 crc kubenswrapper[4631]: E1128 13:41:50.999164 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fb5b797e-a328-4f2e-a239-333213537f2a" containerName="nova-metadata-log" Nov 28 13:41:50 crc kubenswrapper[4631]: I1128 13:41:50.999181 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="fb5b797e-a328-4f2e-a239-333213537f2a" containerName="nova-metadata-log" Nov 28 13:41:50 crc kubenswrapper[4631]: E1128 13:41:50.999212 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fb5b797e-a328-4f2e-a239-333213537f2a" containerName="nova-metadata-metadata" Nov 28 13:41:50 crc kubenswrapper[4631]: I1128 13:41:50.999220 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="fb5b797e-a328-4f2e-a239-333213537f2a" containerName="nova-metadata-metadata" Nov 28 13:41:51 crc kubenswrapper[4631]: I1128 13:41:50.999987 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="fb5b797e-a328-4f2e-a239-333213537f2a" containerName="nova-metadata-metadata" Nov 28 13:41:51 crc kubenswrapper[4631]: I1128 13:41:51.000041 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="fb5b797e-a328-4f2e-a239-333213537f2a" containerName="nova-metadata-log" Nov 28 13:41:51 crc kubenswrapper[4631]: I1128 13:41:51.001490 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 28 13:41:51 crc kubenswrapper[4631]: I1128 13:41:51.006861 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Nov 28 13:41:51 crc kubenswrapper[4631]: I1128 13:41:51.007143 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Nov 28 13:41:51 crc kubenswrapper[4631]: I1128 13:41:51.013806 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 28 13:41:51 crc kubenswrapper[4631]: I1128 13:41:51.084522 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/52191ac4-b50d-4c9a-8a42-b3afe3cffe7f-logs\") pod \"nova-metadata-0\" (UID: \"52191ac4-b50d-4c9a-8a42-b3afe3cffe7f\") " pod="openstack/nova-metadata-0" Nov 28 13:41:51 crc kubenswrapper[4631]: I1128 13:41:51.085120 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/52191ac4-b50d-4c9a-8a42-b3afe3cffe7f-config-data\") pod \"nova-metadata-0\" (UID: \"52191ac4-b50d-4c9a-8a42-b3afe3cffe7f\") " pod="openstack/nova-metadata-0" Nov 28 13:41:51 crc kubenswrapper[4631]: I1128 13:41:51.085159 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/52191ac4-b50d-4c9a-8a42-b3afe3cffe7f-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"52191ac4-b50d-4c9a-8a42-b3afe3cffe7f\") " pod="openstack/nova-metadata-0" Nov 28 13:41:51 crc kubenswrapper[4631]: I1128 13:41:51.085214 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8kxrk\" (UniqueName: \"kubernetes.io/projected/52191ac4-b50d-4c9a-8a42-b3afe3cffe7f-kube-api-access-8kxrk\") pod \"nova-metadata-0\" (UID: \"52191ac4-b50d-4c9a-8a42-b3afe3cffe7f\") " pod="openstack/nova-metadata-0" Nov 28 13:41:51 crc kubenswrapper[4631]: I1128 13:41:51.085300 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/52191ac4-b50d-4c9a-8a42-b3afe3cffe7f-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"52191ac4-b50d-4c9a-8a42-b3afe3cffe7f\") " pod="openstack/nova-metadata-0" Nov 28 13:41:51 crc kubenswrapper[4631]: I1128 13:41:51.187501 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/52191ac4-b50d-4c9a-8a42-b3afe3cffe7f-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"52191ac4-b50d-4c9a-8a42-b3afe3cffe7f\") " pod="openstack/nova-metadata-0" Nov 28 13:41:51 crc kubenswrapper[4631]: I1128 13:41:51.187604 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/52191ac4-b50d-4c9a-8a42-b3afe3cffe7f-logs\") pod \"nova-metadata-0\" (UID: \"52191ac4-b50d-4c9a-8a42-b3afe3cffe7f\") " pod="openstack/nova-metadata-0" Nov 28 13:41:51 crc kubenswrapper[4631]: I1128 13:41:51.187638 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/52191ac4-b50d-4c9a-8a42-b3afe3cffe7f-config-data\") pod \"nova-metadata-0\" (UID: \"52191ac4-b50d-4c9a-8a42-b3afe3cffe7f\") " pod="openstack/nova-metadata-0" Nov 28 13:41:51 crc kubenswrapper[4631]: I1128 13:41:51.187670 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/52191ac4-b50d-4c9a-8a42-b3afe3cffe7f-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"52191ac4-b50d-4c9a-8a42-b3afe3cffe7f\") " pod="openstack/nova-metadata-0" Nov 28 13:41:51 crc kubenswrapper[4631]: I1128 13:41:51.187721 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8kxrk\" (UniqueName: \"kubernetes.io/projected/52191ac4-b50d-4c9a-8a42-b3afe3cffe7f-kube-api-access-8kxrk\") pod \"nova-metadata-0\" (UID: \"52191ac4-b50d-4c9a-8a42-b3afe3cffe7f\") " pod="openstack/nova-metadata-0" Nov 28 13:41:51 crc kubenswrapper[4631]: I1128 13:41:51.188673 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/52191ac4-b50d-4c9a-8a42-b3afe3cffe7f-logs\") pod \"nova-metadata-0\" (UID: \"52191ac4-b50d-4c9a-8a42-b3afe3cffe7f\") " pod="openstack/nova-metadata-0" Nov 28 13:41:51 crc kubenswrapper[4631]: I1128 13:41:51.194047 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/52191ac4-b50d-4c9a-8a42-b3afe3cffe7f-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"52191ac4-b50d-4c9a-8a42-b3afe3cffe7f\") " pod="openstack/nova-metadata-0" Nov 28 13:41:51 crc kubenswrapper[4631]: I1128 13:41:51.199773 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/52191ac4-b50d-4c9a-8a42-b3afe3cffe7f-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"52191ac4-b50d-4c9a-8a42-b3afe3cffe7f\") " pod="openstack/nova-metadata-0" Nov 28 13:41:51 crc kubenswrapper[4631]: I1128 13:41:51.207825 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8kxrk\" (UniqueName: \"kubernetes.io/projected/52191ac4-b50d-4c9a-8a42-b3afe3cffe7f-kube-api-access-8kxrk\") pod \"nova-metadata-0\" (UID: \"52191ac4-b50d-4c9a-8a42-b3afe3cffe7f\") " pod="openstack/nova-metadata-0" Nov 28 13:41:51 crc kubenswrapper[4631]: I1128 13:41:51.211657 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/52191ac4-b50d-4c9a-8a42-b3afe3cffe7f-config-data\") pod \"nova-metadata-0\" (UID: \"52191ac4-b50d-4c9a-8a42-b3afe3cffe7f\") " pod="openstack/nova-metadata-0" Nov 28 13:41:51 crc kubenswrapper[4631]: I1128 13:41:51.386785 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 28 13:41:51 crc kubenswrapper[4631]: I1128 13:41:51.554170 4631 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fb5b797e-a328-4f2e-a239-333213537f2a" path="/var/lib/kubelet/pods/fb5b797e-a328-4f2e-a239-333213537f2a/volumes" Nov 28 13:41:51 crc kubenswrapper[4631]: I1128 13:41:51.893048 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 28 13:41:51 crc kubenswrapper[4631]: W1128 13:41:51.895510 4631 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod52191ac4_b50d_4c9a_8a42_b3afe3cffe7f.slice/crio-d989265701a9d46861892507eff66a2ca791f8f6fb4870a9ea8e67a545b6ed6a WatchSource:0}: Error finding container d989265701a9d46861892507eff66a2ca791f8f6fb4870a9ea8e67a545b6ed6a: Status 404 returned error can't find the container with id d989265701a9d46861892507eff66a2ca791f8f6fb4870a9ea8e67a545b6ed6a Nov 28 13:41:52 crc kubenswrapper[4631]: I1128 13:41:52.911322 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"52191ac4-b50d-4c9a-8a42-b3afe3cffe7f","Type":"ContainerStarted","Data":"c13b8512b425c9f494c2b7da8a1445977a089749ff9b738f8385397569b08f93"} Nov 28 13:41:52 crc kubenswrapper[4631]: I1128 13:41:52.911788 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"52191ac4-b50d-4c9a-8a42-b3afe3cffe7f","Type":"ContainerStarted","Data":"cc2257c2902dcc350175971ac9e4b28018e228500dc01a500eedd5ec26f0afa4"} Nov 28 13:41:52 crc kubenswrapper[4631]: I1128 13:41:52.911806 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"52191ac4-b50d-4c9a-8a42-b3afe3cffe7f","Type":"ContainerStarted","Data":"d989265701a9d46861892507eff66a2ca791f8f6fb4870a9ea8e67a545b6ed6a"} Nov 28 13:41:52 crc kubenswrapper[4631]: I1128 13:41:52.930911 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.930880215 podStartE2EDuration="2.930880215s" podCreationTimestamp="2025-11-28 13:41:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 13:41:52.928199293 +0000 UTC m=+1269.735502657" watchObservedRunningTime="2025-11-28 13:41:52.930880215 +0000 UTC m=+1269.738183559" Nov 28 13:41:54 crc kubenswrapper[4631]: I1128 13:41:54.513008 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Nov 28 13:41:56 crc kubenswrapper[4631]: I1128 13:41:56.388045 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 28 13:41:56 crc kubenswrapper[4631]: I1128 13:41:56.388879 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 28 13:41:58 crc kubenswrapper[4631]: I1128 13:41:58.534856 4631 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 28 13:41:58 crc kubenswrapper[4631]: I1128 13:41:58.534939 4631 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 28 13:41:59 crc kubenswrapper[4631]: I1128 13:41:59.531697 4631 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Nov 28 13:41:59 crc kubenswrapper[4631]: I1128 13:41:59.552403 4631 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Nov 28 13:41:59 crc kubenswrapper[4631]: I1128 13:41:59.552510 4631 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="9db3c70b-08ec-4b40-bdb9-94cdbc66fa1d" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.202:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 28 13:41:59 crc kubenswrapper[4631]: I1128 13:41:59.552528 4631 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="9db3c70b-08ec-4b40-bdb9-94cdbc66fa1d" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.202:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 28 13:42:00 crc kubenswrapper[4631]: I1128 13:42:00.022654 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Nov 28 13:42:01 crc kubenswrapper[4631]: I1128 13:42:01.389336 4631 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Nov 28 13:42:01 crc kubenswrapper[4631]: I1128 13:42:01.389764 4631 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Nov 28 13:42:02 crc kubenswrapper[4631]: I1128 13:42:02.402500 4631 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="52191ac4-b50d-4c9a-8a42-b3afe3cffe7f" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.204:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 28 13:42:02 crc kubenswrapper[4631]: I1128 13:42:02.402533 4631 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="52191ac4-b50d-4c9a-8a42-b3afe3cffe7f" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.204:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 28 13:42:07 crc kubenswrapper[4631]: I1128 13:42:07.138551 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Nov 28 13:42:08 crc kubenswrapper[4631]: I1128 13:42:08.546139 4631 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Nov 28 13:42:08 crc kubenswrapper[4631]: I1128 13:42:08.547097 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Nov 28 13:42:08 crc kubenswrapper[4631]: I1128 13:42:08.549462 4631 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Nov 28 13:42:08 crc kubenswrapper[4631]: I1128 13:42:08.555895 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Nov 28 13:42:09 crc kubenswrapper[4631]: I1128 13:42:09.081987 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Nov 28 13:42:09 crc kubenswrapper[4631]: I1128 13:42:09.089393 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Nov 28 13:42:11 crc kubenswrapper[4631]: I1128 13:42:11.394360 4631 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Nov 28 13:42:11 crc kubenswrapper[4631]: I1128 13:42:11.396016 4631 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Nov 28 13:42:11 crc kubenswrapper[4631]: I1128 13:42:11.403913 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Nov 28 13:42:12 crc kubenswrapper[4631]: I1128 13:42:12.124774 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Nov 28 13:42:20 crc kubenswrapper[4631]: I1128 13:42:20.347646 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 28 13:42:21 crc kubenswrapper[4631]: I1128 13:42:21.651794 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 28 13:42:26 crc kubenswrapper[4631]: I1128 13:42:26.146884 4631 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="3a91e9fa-34b7-4688-9171-56a62043759d" containerName="rabbitmq" containerID="cri-o://798f0920b82c4321f979cbb346e8d3258fa7abb13c5c825bae8d9f3b06fd453e" gracePeriod=604795 Nov 28 13:42:28 crc kubenswrapper[4631]: I1128 13:42:28.023154 4631 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="ed4141f2-5eff-4252-8dc7-ad60a1580189" containerName="rabbitmq" containerID="cri-o://bc7825d2e428610e7d264dba3f92af76fc65d8e94ba32e808f0899e4d096a085" gracePeriod=604794 Nov 28 13:42:32 crc kubenswrapper[4631]: I1128 13:42:32.397702 4631 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="3a91e9fa-34b7-4688-9171-56a62043759d" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.99:5671: connect: connection refused" Nov 28 13:42:32 crc kubenswrapper[4631]: I1128 13:42:32.852000 4631 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="ed4141f2-5eff-4252-8dc7-ad60a1580189" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.100:5671: connect: connection refused" Nov 28 13:42:34 crc kubenswrapper[4631]: I1128 13:42:34.369863 4631 generic.go:334] "Generic (PLEG): container finished" podID="ed4141f2-5eff-4252-8dc7-ad60a1580189" containerID="bc7825d2e428610e7d264dba3f92af76fc65d8e94ba32e808f0899e4d096a085" exitCode=0 Nov 28 13:42:34 crc kubenswrapper[4631]: I1128 13:42:34.370013 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"ed4141f2-5eff-4252-8dc7-ad60a1580189","Type":"ContainerDied","Data":"bc7825d2e428610e7d264dba3f92af76fc65d8e94ba32e808f0899e4d096a085"} Nov 28 13:42:34 crc kubenswrapper[4631]: I1128 13:42:34.374949 4631 generic.go:334] "Generic (PLEG): container finished" podID="3a91e9fa-34b7-4688-9171-56a62043759d" containerID="798f0920b82c4321f979cbb346e8d3258fa7abb13c5c825bae8d9f3b06fd453e" exitCode=0 Nov 28 13:42:34 crc kubenswrapper[4631]: I1128 13:42:34.375012 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"3a91e9fa-34b7-4688-9171-56a62043759d","Type":"ContainerDied","Data":"798f0920b82c4321f979cbb346e8d3258fa7abb13c5c825bae8d9f3b06fd453e"} Nov 28 13:42:35 crc kubenswrapper[4631]: I1128 13:42:34.693537 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 28 13:42:35 crc kubenswrapper[4631]: I1128 13:42:34.705150 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 28 13:42:35 crc kubenswrapper[4631]: I1128 13:42:34.890198 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/3a91e9fa-34b7-4688-9171-56a62043759d-rabbitmq-tls\") pod \"3a91e9fa-34b7-4688-9171-56a62043759d\" (UID: \"3a91e9fa-34b7-4688-9171-56a62043759d\") " Nov 28 13:42:35 crc kubenswrapper[4631]: I1128 13:42:34.890254 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/3a91e9fa-34b7-4688-9171-56a62043759d-rabbitmq-confd\") pod \"3a91e9fa-34b7-4688-9171-56a62043759d\" (UID: \"3a91e9fa-34b7-4688-9171-56a62043759d\") " Nov 28 13:42:35 crc kubenswrapper[4631]: I1128 13:42:34.890280 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/ed4141f2-5eff-4252-8dc7-ad60a1580189-rabbitmq-plugins\") pod \"ed4141f2-5eff-4252-8dc7-ad60a1580189\" (UID: \"ed4141f2-5eff-4252-8dc7-ad60a1580189\") " Nov 28 13:42:35 crc kubenswrapper[4631]: I1128 13:42:34.890332 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"ed4141f2-5eff-4252-8dc7-ad60a1580189\" (UID: \"ed4141f2-5eff-4252-8dc7-ad60a1580189\") " Nov 28 13:42:35 crc kubenswrapper[4631]: I1128 13:42:34.890364 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/3a91e9fa-34b7-4688-9171-56a62043759d-rabbitmq-erlang-cookie\") pod \"3a91e9fa-34b7-4688-9171-56a62043759d\" (UID: \"3a91e9fa-34b7-4688-9171-56a62043759d\") " Nov 28 13:42:35 crc kubenswrapper[4631]: I1128 13:42:34.890393 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4x8mv\" (UniqueName: \"kubernetes.io/projected/3a91e9fa-34b7-4688-9171-56a62043759d-kube-api-access-4x8mv\") pod \"3a91e9fa-34b7-4688-9171-56a62043759d\" (UID: \"3a91e9fa-34b7-4688-9171-56a62043759d\") " Nov 28 13:42:35 crc kubenswrapper[4631]: I1128 13:42:34.890429 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4vt5k\" (UniqueName: \"kubernetes.io/projected/ed4141f2-5eff-4252-8dc7-ad60a1580189-kube-api-access-4vt5k\") pod \"ed4141f2-5eff-4252-8dc7-ad60a1580189\" (UID: \"ed4141f2-5eff-4252-8dc7-ad60a1580189\") " Nov 28 13:42:35 crc kubenswrapper[4631]: I1128 13:42:34.890460 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/ed4141f2-5eff-4252-8dc7-ad60a1580189-rabbitmq-confd\") pod \"ed4141f2-5eff-4252-8dc7-ad60a1580189\" (UID: \"ed4141f2-5eff-4252-8dc7-ad60a1580189\") " Nov 28 13:42:35 crc kubenswrapper[4631]: I1128 13:42:34.890501 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/ed4141f2-5eff-4252-8dc7-ad60a1580189-rabbitmq-erlang-cookie\") pod \"ed4141f2-5eff-4252-8dc7-ad60a1580189\" (UID: \"ed4141f2-5eff-4252-8dc7-ad60a1580189\") " Nov 28 13:42:35 crc kubenswrapper[4631]: I1128 13:42:34.890534 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3a91e9fa-34b7-4688-9171-56a62043759d-config-data\") pod \"3a91e9fa-34b7-4688-9171-56a62043759d\" (UID: \"3a91e9fa-34b7-4688-9171-56a62043759d\") " Nov 28 13:42:35 crc kubenswrapper[4631]: I1128 13:42:34.890567 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/3a91e9fa-34b7-4688-9171-56a62043759d-plugins-conf\") pod \"3a91e9fa-34b7-4688-9171-56a62043759d\" (UID: \"3a91e9fa-34b7-4688-9171-56a62043759d\") " Nov 28 13:42:35 crc kubenswrapper[4631]: I1128 13:42:34.890591 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/3a91e9fa-34b7-4688-9171-56a62043759d-server-conf\") pod \"3a91e9fa-34b7-4688-9171-56a62043759d\" (UID: \"3a91e9fa-34b7-4688-9171-56a62043759d\") " Nov 28 13:42:35 crc kubenswrapper[4631]: I1128 13:42:34.890614 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/ed4141f2-5eff-4252-8dc7-ad60a1580189-plugins-conf\") pod \"ed4141f2-5eff-4252-8dc7-ad60a1580189\" (UID: \"ed4141f2-5eff-4252-8dc7-ad60a1580189\") " Nov 28 13:42:35 crc kubenswrapper[4631]: I1128 13:42:34.890672 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"3a91e9fa-34b7-4688-9171-56a62043759d\" (UID: \"3a91e9fa-34b7-4688-9171-56a62043759d\") " Nov 28 13:42:35 crc kubenswrapper[4631]: I1128 13:42:34.891740 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/3a91e9fa-34b7-4688-9171-56a62043759d-pod-info\") pod \"3a91e9fa-34b7-4688-9171-56a62043759d\" (UID: \"3a91e9fa-34b7-4688-9171-56a62043759d\") " Nov 28 13:42:35 crc kubenswrapper[4631]: I1128 13:42:34.891810 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/ed4141f2-5eff-4252-8dc7-ad60a1580189-server-conf\") pod \"ed4141f2-5eff-4252-8dc7-ad60a1580189\" (UID: \"ed4141f2-5eff-4252-8dc7-ad60a1580189\") " Nov 28 13:42:35 crc kubenswrapper[4631]: I1128 13:42:34.891857 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/ed4141f2-5eff-4252-8dc7-ad60a1580189-pod-info\") pod \"ed4141f2-5eff-4252-8dc7-ad60a1580189\" (UID: \"ed4141f2-5eff-4252-8dc7-ad60a1580189\") " Nov 28 13:42:35 crc kubenswrapper[4631]: I1128 13:42:34.891894 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/3a91e9fa-34b7-4688-9171-56a62043759d-erlang-cookie-secret\") pod \"3a91e9fa-34b7-4688-9171-56a62043759d\" (UID: \"3a91e9fa-34b7-4688-9171-56a62043759d\") " Nov 28 13:42:35 crc kubenswrapper[4631]: I1128 13:42:34.891946 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/ed4141f2-5eff-4252-8dc7-ad60a1580189-erlang-cookie-secret\") pod \"ed4141f2-5eff-4252-8dc7-ad60a1580189\" (UID: \"ed4141f2-5eff-4252-8dc7-ad60a1580189\") " Nov 28 13:42:35 crc kubenswrapper[4631]: I1128 13:42:34.891991 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/3a91e9fa-34b7-4688-9171-56a62043759d-rabbitmq-plugins\") pod \"3a91e9fa-34b7-4688-9171-56a62043759d\" (UID: \"3a91e9fa-34b7-4688-9171-56a62043759d\") " Nov 28 13:42:35 crc kubenswrapper[4631]: I1128 13:42:34.892056 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/ed4141f2-5eff-4252-8dc7-ad60a1580189-rabbitmq-tls\") pod \"ed4141f2-5eff-4252-8dc7-ad60a1580189\" (UID: \"ed4141f2-5eff-4252-8dc7-ad60a1580189\") " Nov 28 13:42:35 crc kubenswrapper[4631]: I1128 13:42:34.892088 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ed4141f2-5eff-4252-8dc7-ad60a1580189-config-data\") pod \"ed4141f2-5eff-4252-8dc7-ad60a1580189\" (UID: \"ed4141f2-5eff-4252-8dc7-ad60a1580189\") " Nov 28 13:42:35 crc kubenswrapper[4631]: I1128 13:42:34.897701 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ed4141f2-5eff-4252-8dc7-ad60a1580189-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "ed4141f2-5eff-4252-8dc7-ad60a1580189" (UID: "ed4141f2-5eff-4252-8dc7-ad60a1580189"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 13:42:35 crc kubenswrapper[4631]: I1128 13:42:34.918245 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3a91e9fa-34b7-4688-9171-56a62043759d-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "3a91e9fa-34b7-4688-9171-56a62043759d" (UID: "3a91e9fa-34b7-4688-9171-56a62043759d"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 13:42:35 crc kubenswrapper[4631]: I1128 13:42:34.918326 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3a91e9fa-34b7-4688-9171-56a62043759d-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "3a91e9fa-34b7-4688-9171-56a62043759d" (UID: "3a91e9fa-34b7-4688-9171-56a62043759d"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 13:42:35 crc kubenswrapper[4631]: I1128 13:42:34.918412 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ed4141f2-5eff-4252-8dc7-ad60a1580189-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "ed4141f2-5eff-4252-8dc7-ad60a1580189" (UID: "ed4141f2-5eff-4252-8dc7-ad60a1580189"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 13:42:35 crc kubenswrapper[4631]: I1128 13:42:34.922781 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3a91e9fa-34b7-4688-9171-56a62043759d-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "3a91e9fa-34b7-4688-9171-56a62043759d" (UID: "3a91e9fa-34b7-4688-9171-56a62043759d"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 13:42:35 crc kubenswrapper[4631]: I1128 13:42:34.934255 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/ed4141f2-5eff-4252-8dc7-ad60a1580189-pod-info" (OuterVolumeSpecName: "pod-info") pod "ed4141f2-5eff-4252-8dc7-ad60a1580189" (UID: "ed4141f2-5eff-4252-8dc7-ad60a1580189"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Nov 28 13:42:35 crc kubenswrapper[4631]: I1128 13:42:34.934640 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage10-crc" (OuterVolumeSpecName: "persistence") pod "ed4141f2-5eff-4252-8dc7-ad60a1580189" (UID: "ed4141f2-5eff-4252-8dc7-ad60a1580189"). InnerVolumeSpecName "local-storage10-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 28 13:42:35 crc kubenswrapper[4631]: I1128 13:42:34.947757 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ed4141f2-5eff-4252-8dc7-ad60a1580189-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "ed4141f2-5eff-4252-8dc7-ad60a1580189" (UID: "ed4141f2-5eff-4252-8dc7-ad60a1580189"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 13:42:35 crc kubenswrapper[4631]: I1128 13:42:34.948346 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3a91e9fa-34b7-4688-9171-56a62043759d-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "3a91e9fa-34b7-4688-9171-56a62043759d" (UID: "3a91e9fa-34b7-4688-9171-56a62043759d"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:42:35 crc kubenswrapper[4631]: I1128 13:42:34.948631 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ed4141f2-5eff-4252-8dc7-ad60a1580189-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "ed4141f2-5eff-4252-8dc7-ad60a1580189" (UID: "ed4141f2-5eff-4252-8dc7-ad60a1580189"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:42:35 crc kubenswrapper[4631]: I1128 13:42:34.948833 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ed4141f2-5eff-4252-8dc7-ad60a1580189-kube-api-access-4vt5k" (OuterVolumeSpecName: "kube-api-access-4vt5k") pod "ed4141f2-5eff-4252-8dc7-ad60a1580189" (UID: "ed4141f2-5eff-4252-8dc7-ad60a1580189"). InnerVolumeSpecName "kube-api-access-4vt5k". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:42:35 crc kubenswrapper[4631]: I1128 13:42:34.948899 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3a91e9fa-34b7-4688-9171-56a62043759d-kube-api-access-4x8mv" (OuterVolumeSpecName: "kube-api-access-4x8mv") pod "3a91e9fa-34b7-4688-9171-56a62043759d" (UID: "3a91e9fa-34b7-4688-9171-56a62043759d"). InnerVolumeSpecName "kube-api-access-4x8mv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:42:35 crc kubenswrapper[4631]: I1128 13:42:34.951214 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3a91e9fa-34b7-4688-9171-56a62043759d-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "3a91e9fa-34b7-4688-9171-56a62043759d" (UID: "3a91e9fa-34b7-4688-9171-56a62043759d"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:42:35 crc kubenswrapper[4631]: I1128 13:42:34.980496 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ed4141f2-5eff-4252-8dc7-ad60a1580189-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "ed4141f2-5eff-4252-8dc7-ad60a1580189" (UID: "ed4141f2-5eff-4252-8dc7-ad60a1580189"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:42:35 crc kubenswrapper[4631]: I1128 13:42:34.982600 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/3a91e9fa-34b7-4688-9171-56a62043759d-pod-info" (OuterVolumeSpecName: "pod-info") pod "3a91e9fa-34b7-4688-9171-56a62043759d" (UID: "3a91e9fa-34b7-4688-9171-56a62043759d"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Nov 28 13:42:35 crc kubenswrapper[4631]: I1128 13:42:35.000015 4631 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/ed4141f2-5eff-4252-8dc7-ad60a1580189-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Nov 28 13:42:35 crc kubenswrapper[4631]: I1128 13:42:35.000051 4631 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/3a91e9fa-34b7-4688-9171-56a62043759d-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Nov 28 13:42:35 crc kubenswrapper[4631]: I1128 13:42:35.000061 4631 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/ed4141f2-5eff-4252-8dc7-ad60a1580189-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Nov 28 13:42:35 crc kubenswrapper[4631]: I1128 13:42:35.000096 4631 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" " Nov 28 13:42:35 crc kubenswrapper[4631]: I1128 13:42:35.000109 4631 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/3a91e9fa-34b7-4688-9171-56a62043759d-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Nov 28 13:42:35 crc kubenswrapper[4631]: I1128 13:42:35.000122 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4x8mv\" (UniqueName: \"kubernetes.io/projected/3a91e9fa-34b7-4688-9171-56a62043759d-kube-api-access-4x8mv\") on node \"crc\" DevicePath \"\"" Nov 28 13:42:35 crc kubenswrapper[4631]: I1128 13:42:35.000135 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4vt5k\" (UniqueName: \"kubernetes.io/projected/ed4141f2-5eff-4252-8dc7-ad60a1580189-kube-api-access-4vt5k\") on node \"crc\" DevicePath \"\"" Nov 28 13:42:35 crc kubenswrapper[4631]: I1128 13:42:35.000144 4631 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/ed4141f2-5eff-4252-8dc7-ad60a1580189-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Nov 28 13:42:35 crc kubenswrapper[4631]: I1128 13:42:35.000153 4631 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/3a91e9fa-34b7-4688-9171-56a62043759d-plugins-conf\") on node \"crc\" DevicePath \"\"" Nov 28 13:42:35 crc kubenswrapper[4631]: I1128 13:42:35.000179 4631 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/ed4141f2-5eff-4252-8dc7-ad60a1580189-plugins-conf\") on node \"crc\" DevicePath \"\"" Nov 28 13:42:35 crc kubenswrapper[4631]: I1128 13:42:35.000188 4631 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/3a91e9fa-34b7-4688-9171-56a62043759d-pod-info\") on node \"crc\" DevicePath \"\"" Nov 28 13:42:35 crc kubenswrapper[4631]: I1128 13:42:35.000196 4631 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/ed4141f2-5eff-4252-8dc7-ad60a1580189-pod-info\") on node \"crc\" DevicePath \"\"" Nov 28 13:42:35 crc kubenswrapper[4631]: I1128 13:42:35.000204 4631 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/3a91e9fa-34b7-4688-9171-56a62043759d-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Nov 28 13:42:35 crc kubenswrapper[4631]: I1128 13:42:35.000239 4631 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/ed4141f2-5eff-4252-8dc7-ad60a1580189-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Nov 28 13:42:35 crc kubenswrapper[4631]: I1128 13:42:35.000248 4631 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/3a91e9fa-34b7-4688-9171-56a62043759d-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Nov 28 13:42:35 crc kubenswrapper[4631]: I1128 13:42:35.006532 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage01-crc" (OuterVolumeSpecName: "persistence") pod "3a91e9fa-34b7-4688-9171-56a62043759d" (UID: "3a91e9fa-34b7-4688-9171-56a62043759d"). InnerVolumeSpecName "local-storage01-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 28 13:42:35 crc kubenswrapper[4631]: I1128 13:42:35.025127 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ed4141f2-5eff-4252-8dc7-ad60a1580189-config-data" (OuterVolumeSpecName: "config-data") pod "ed4141f2-5eff-4252-8dc7-ad60a1580189" (UID: "ed4141f2-5eff-4252-8dc7-ad60a1580189"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 13:42:35 crc kubenswrapper[4631]: I1128 13:42:35.079270 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3a91e9fa-34b7-4688-9171-56a62043759d-config-data" (OuterVolumeSpecName: "config-data") pod "3a91e9fa-34b7-4688-9171-56a62043759d" (UID: "3a91e9fa-34b7-4688-9171-56a62043759d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 13:42:35 crc kubenswrapper[4631]: I1128 13:42:35.096705 4631 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage10-crc" (UniqueName: "kubernetes.io/local-volume/local-storage10-crc") on node "crc" Nov 28 13:42:35 crc kubenswrapper[4631]: I1128 13:42:35.101510 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ed4141f2-5eff-4252-8dc7-ad60a1580189-server-conf" (OuterVolumeSpecName: "server-conf") pod "ed4141f2-5eff-4252-8dc7-ad60a1580189" (UID: "ed4141f2-5eff-4252-8dc7-ad60a1580189"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 13:42:35 crc kubenswrapper[4631]: I1128 13:42:35.102370 4631 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/ed4141f2-5eff-4252-8dc7-ad60a1580189-server-conf\") on node \"crc\" DevicePath \"\"" Nov 28 13:42:35 crc kubenswrapper[4631]: I1128 13:42:35.102403 4631 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ed4141f2-5eff-4252-8dc7-ad60a1580189-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 13:42:35 crc kubenswrapper[4631]: I1128 13:42:35.102415 4631 reconciler_common.go:293] "Volume detached for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" DevicePath \"\"" Nov 28 13:42:35 crc kubenswrapper[4631]: I1128 13:42:35.102428 4631 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3a91e9fa-34b7-4688-9171-56a62043759d-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 13:42:35 crc kubenswrapper[4631]: I1128 13:42:35.102472 4631 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" " Nov 28 13:42:35 crc kubenswrapper[4631]: I1128 13:42:35.109708 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3a91e9fa-34b7-4688-9171-56a62043759d-server-conf" (OuterVolumeSpecName: "server-conf") pod "3a91e9fa-34b7-4688-9171-56a62043759d" (UID: "3a91e9fa-34b7-4688-9171-56a62043759d"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 13:42:35 crc kubenswrapper[4631]: I1128 13:42:35.155798 4631 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage01-crc" (UniqueName: "kubernetes.io/local-volume/local-storage01-crc") on node "crc" Nov 28 13:42:35 crc kubenswrapper[4631]: I1128 13:42:35.199649 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ed4141f2-5eff-4252-8dc7-ad60a1580189-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "ed4141f2-5eff-4252-8dc7-ad60a1580189" (UID: "ed4141f2-5eff-4252-8dc7-ad60a1580189"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:42:35 crc kubenswrapper[4631]: I1128 13:42:35.205086 4631 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/3a91e9fa-34b7-4688-9171-56a62043759d-server-conf\") on node \"crc\" DevicePath \"\"" Nov 28 13:42:35 crc kubenswrapper[4631]: I1128 13:42:35.205120 4631 reconciler_common.go:293] "Volume detached for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" DevicePath \"\"" Nov 28 13:42:35 crc kubenswrapper[4631]: I1128 13:42:35.205130 4631 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/ed4141f2-5eff-4252-8dc7-ad60a1580189-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Nov 28 13:42:35 crc kubenswrapper[4631]: I1128 13:42:35.214731 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3a91e9fa-34b7-4688-9171-56a62043759d-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "3a91e9fa-34b7-4688-9171-56a62043759d" (UID: "3a91e9fa-34b7-4688-9171-56a62043759d"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:42:35 crc kubenswrapper[4631]: I1128 13:42:35.307934 4631 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/3a91e9fa-34b7-4688-9171-56a62043759d-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Nov 28 13:42:35 crc kubenswrapper[4631]: I1128 13:42:35.387508 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"ed4141f2-5eff-4252-8dc7-ad60a1580189","Type":"ContainerDied","Data":"f3f6ed0a79454bb6e8514a61efc6444fdc3d68ebeae1172304b680407d122dff"} Nov 28 13:42:35 crc kubenswrapper[4631]: I1128 13:42:35.387566 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 28 13:42:35 crc kubenswrapper[4631]: I1128 13:42:35.387567 4631 scope.go:117] "RemoveContainer" containerID="bc7825d2e428610e7d264dba3f92af76fc65d8e94ba32e808f0899e4d096a085" Nov 28 13:42:35 crc kubenswrapper[4631]: I1128 13:42:35.392154 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"3a91e9fa-34b7-4688-9171-56a62043759d","Type":"ContainerDied","Data":"eca7fbb416ef9f4ae40f525f94aa899b89d54b720c33ba9321840fe268dd2633"} Nov 28 13:42:35 crc kubenswrapper[4631]: I1128 13:42:35.392246 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 28 13:42:35 crc kubenswrapper[4631]: I1128 13:42:35.422483 4631 scope.go:117] "RemoveContainer" containerID="41d8c218651f1d165fa68a2452a35246fa25140ebacda04483d76b1fc68b2e42" Nov 28 13:42:35 crc kubenswrapper[4631]: I1128 13:42:35.438390 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 28 13:42:35 crc kubenswrapper[4631]: I1128 13:42:35.448955 4631 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 28 13:42:35 crc kubenswrapper[4631]: I1128 13:42:35.495622 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 28 13:42:35 crc kubenswrapper[4631]: I1128 13:42:35.516231 4631 scope.go:117] "RemoveContainer" containerID="798f0920b82c4321f979cbb346e8d3258fa7abb13c5c825bae8d9f3b06fd453e" Nov 28 13:42:35 crc kubenswrapper[4631]: I1128 13:42:35.548421 4631 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ed4141f2-5eff-4252-8dc7-ad60a1580189" path="/var/lib/kubelet/pods/ed4141f2-5eff-4252-8dc7-ad60a1580189/volumes" Nov 28 13:42:35 crc kubenswrapper[4631]: I1128 13:42:35.549565 4631 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 28 13:42:35 crc kubenswrapper[4631]: I1128 13:42:35.562424 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Nov 28 13:42:35 crc kubenswrapper[4631]: E1128 13:42:35.563053 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ed4141f2-5eff-4252-8dc7-ad60a1580189" containerName="rabbitmq" Nov 28 13:42:35 crc kubenswrapper[4631]: I1128 13:42:35.563068 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="ed4141f2-5eff-4252-8dc7-ad60a1580189" containerName="rabbitmq" Nov 28 13:42:35 crc kubenswrapper[4631]: E1128 13:42:35.563084 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3a91e9fa-34b7-4688-9171-56a62043759d" containerName="rabbitmq" Nov 28 13:42:35 crc kubenswrapper[4631]: I1128 13:42:35.563090 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a91e9fa-34b7-4688-9171-56a62043759d" containerName="rabbitmq" Nov 28 13:42:35 crc kubenswrapper[4631]: E1128 13:42:35.563132 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ed4141f2-5eff-4252-8dc7-ad60a1580189" containerName="setup-container" Nov 28 13:42:35 crc kubenswrapper[4631]: I1128 13:42:35.563140 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="ed4141f2-5eff-4252-8dc7-ad60a1580189" containerName="setup-container" Nov 28 13:42:35 crc kubenswrapper[4631]: E1128 13:42:35.563149 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3a91e9fa-34b7-4688-9171-56a62043759d" containerName="setup-container" Nov 28 13:42:35 crc kubenswrapper[4631]: I1128 13:42:35.563157 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a91e9fa-34b7-4688-9171-56a62043759d" containerName="setup-container" Nov 28 13:42:35 crc kubenswrapper[4631]: I1128 13:42:35.563401 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="ed4141f2-5eff-4252-8dc7-ad60a1580189" containerName="rabbitmq" Nov 28 13:42:35 crc kubenswrapper[4631]: I1128 13:42:35.563413 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="3a91e9fa-34b7-4688-9171-56a62043759d" containerName="rabbitmq" Nov 28 13:42:35 crc kubenswrapper[4631]: I1128 13:42:35.564773 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 28 13:42:35 crc kubenswrapper[4631]: I1128 13:42:35.572218 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 28 13:42:35 crc kubenswrapper[4631]: I1128 13:42:35.573013 4631 scope.go:117] "RemoveContainer" containerID="5f2c613e618512327c6478adcd7de0b71456862e29e5333eaab2b49e071f2777" Nov 28 13:42:35 crc kubenswrapper[4631]: I1128 13:42:35.573199 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Nov 28 13:42:35 crc kubenswrapper[4631]: I1128 13:42:35.573236 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Nov 28 13:42:35 crc kubenswrapper[4631]: I1128 13:42:35.573362 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Nov 28 13:42:35 crc kubenswrapper[4631]: I1128 13:42:35.573398 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Nov 28 13:42:35 crc kubenswrapper[4631]: I1128 13:42:35.573477 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Nov 28 13:42:35 crc kubenswrapper[4631]: I1128 13:42:35.573549 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Nov 28 13:42:35 crc kubenswrapper[4631]: I1128 13:42:35.574158 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 28 13:42:35 crc kubenswrapper[4631]: I1128 13:42:35.574781 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-p7rc4" Nov 28 13:42:35 crc kubenswrapper[4631]: I1128 13:42:35.582737 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Nov 28 13:42:35 crc kubenswrapper[4631]: I1128 13:42:35.582926 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Nov 28 13:42:35 crc kubenswrapper[4631]: I1128 13:42:35.582988 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-vf4xz" Nov 28 13:42:35 crc kubenswrapper[4631]: I1128 13:42:35.583024 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Nov 28 13:42:35 crc kubenswrapper[4631]: I1128 13:42:35.582941 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Nov 28 13:42:35 crc kubenswrapper[4631]: I1128 13:42:35.583098 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Nov 28 13:42:35 crc kubenswrapper[4631]: I1128 13:42:35.583632 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Nov 28 13:42:35 crc kubenswrapper[4631]: I1128 13:42:35.588729 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 28 13:42:35 crc kubenswrapper[4631]: I1128 13:42:35.598700 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 28 13:42:35 crc kubenswrapper[4631]: I1128 13:42:35.629361 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qmhk6\" (UniqueName: \"kubernetes.io/projected/fc5706c2-8b7d-439a-95b3-981907360c92-kube-api-access-qmhk6\") pod \"rabbitmq-server-0\" (UID: \"fc5706c2-8b7d-439a-95b3-981907360c92\") " pod="openstack/rabbitmq-server-0" Nov 28 13:42:35 crc kubenswrapper[4631]: I1128 13:42:35.629673 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/fc5706c2-8b7d-439a-95b3-981907360c92-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"fc5706c2-8b7d-439a-95b3-981907360c92\") " pod="openstack/rabbitmq-server-0" Nov 28 13:42:35 crc kubenswrapper[4631]: I1128 13:42:35.629786 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/fc5706c2-8b7d-439a-95b3-981907360c92-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"fc5706c2-8b7d-439a-95b3-981907360c92\") " pod="openstack/rabbitmq-server-0" Nov 28 13:42:35 crc kubenswrapper[4631]: I1128 13:42:35.629818 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-server-0\" (UID: \"fc5706c2-8b7d-439a-95b3-981907360c92\") " pod="openstack/rabbitmq-server-0" Nov 28 13:42:35 crc kubenswrapper[4631]: I1128 13:42:35.629844 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/fc5706c2-8b7d-439a-95b3-981907360c92-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"fc5706c2-8b7d-439a-95b3-981907360c92\") " pod="openstack/rabbitmq-server-0" Nov 28 13:42:35 crc kubenswrapper[4631]: I1128 13:42:35.629864 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/fc5706c2-8b7d-439a-95b3-981907360c92-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"fc5706c2-8b7d-439a-95b3-981907360c92\") " pod="openstack/rabbitmq-server-0" Nov 28 13:42:35 crc kubenswrapper[4631]: I1128 13:42:35.629940 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/fc5706c2-8b7d-439a-95b3-981907360c92-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"fc5706c2-8b7d-439a-95b3-981907360c92\") " pod="openstack/rabbitmq-server-0" Nov 28 13:42:35 crc kubenswrapper[4631]: I1128 13:42:35.629977 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/fc5706c2-8b7d-439a-95b3-981907360c92-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"fc5706c2-8b7d-439a-95b3-981907360c92\") " pod="openstack/rabbitmq-server-0" Nov 28 13:42:35 crc kubenswrapper[4631]: I1128 13:42:35.630027 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/fc5706c2-8b7d-439a-95b3-981907360c92-pod-info\") pod \"rabbitmq-server-0\" (UID: \"fc5706c2-8b7d-439a-95b3-981907360c92\") " pod="openstack/rabbitmq-server-0" Nov 28 13:42:35 crc kubenswrapper[4631]: I1128 13:42:35.630084 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/fc5706c2-8b7d-439a-95b3-981907360c92-config-data\") pod \"rabbitmq-server-0\" (UID: \"fc5706c2-8b7d-439a-95b3-981907360c92\") " pod="openstack/rabbitmq-server-0" Nov 28 13:42:35 crc kubenswrapper[4631]: I1128 13:42:35.630103 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/fc5706c2-8b7d-439a-95b3-981907360c92-server-conf\") pod \"rabbitmq-server-0\" (UID: \"fc5706c2-8b7d-439a-95b3-981907360c92\") " pod="openstack/rabbitmq-server-0" Nov 28 13:42:35 crc kubenswrapper[4631]: I1128 13:42:35.732779 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/014519bc-68f0-4132-9a46-6b0e5c538c7e-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"014519bc-68f0-4132-9a46-6b0e5c538c7e\") " pod="openstack/rabbitmq-cell1-server-0" Nov 28 13:42:35 crc kubenswrapper[4631]: I1128 13:42:35.733236 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/014519bc-68f0-4132-9a46-6b0e5c538c7e-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"014519bc-68f0-4132-9a46-6b0e5c538c7e\") " pod="openstack/rabbitmq-cell1-server-0" Nov 28 13:42:35 crc kubenswrapper[4631]: I1128 13:42:35.733271 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/fc5706c2-8b7d-439a-95b3-981907360c92-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"fc5706c2-8b7d-439a-95b3-981907360c92\") " pod="openstack/rabbitmq-server-0" Nov 28 13:42:35 crc kubenswrapper[4631]: I1128 13:42:35.733309 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/014519bc-68f0-4132-9a46-6b0e5c538c7e-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"014519bc-68f0-4132-9a46-6b0e5c538c7e\") " pod="openstack/rabbitmq-cell1-server-0" Nov 28 13:42:35 crc kubenswrapper[4631]: I1128 13:42:35.733348 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/fc5706c2-8b7d-439a-95b3-981907360c92-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"fc5706c2-8b7d-439a-95b3-981907360c92\") " pod="openstack/rabbitmq-server-0" Nov 28 13:42:35 crc kubenswrapper[4631]: I1128 13:42:35.733385 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/fc5706c2-8b7d-439a-95b3-981907360c92-pod-info\") pod \"rabbitmq-server-0\" (UID: \"fc5706c2-8b7d-439a-95b3-981907360c92\") " pod="openstack/rabbitmq-server-0" Nov 28 13:42:35 crc kubenswrapper[4631]: I1128 13:42:35.733425 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/014519bc-68f0-4132-9a46-6b0e5c538c7e-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"014519bc-68f0-4132-9a46-6b0e5c538c7e\") " pod="openstack/rabbitmq-cell1-server-0" Nov 28 13:42:35 crc kubenswrapper[4631]: I1128 13:42:35.733450 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/fc5706c2-8b7d-439a-95b3-981907360c92-config-data\") pod \"rabbitmq-server-0\" (UID: \"fc5706c2-8b7d-439a-95b3-981907360c92\") " pod="openstack/rabbitmq-server-0" Nov 28 13:42:35 crc kubenswrapper[4631]: I1128 13:42:35.733473 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/fc5706c2-8b7d-439a-95b3-981907360c92-server-conf\") pod \"rabbitmq-server-0\" (UID: \"fc5706c2-8b7d-439a-95b3-981907360c92\") " pod="openstack/rabbitmq-server-0" Nov 28 13:42:35 crc kubenswrapper[4631]: I1128 13:42:35.733494 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qmhk6\" (UniqueName: \"kubernetes.io/projected/fc5706c2-8b7d-439a-95b3-981907360c92-kube-api-access-qmhk6\") pod \"rabbitmq-server-0\" (UID: \"fc5706c2-8b7d-439a-95b3-981907360c92\") " pod="openstack/rabbitmq-server-0" Nov 28 13:42:35 crc kubenswrapper[4631]: I1128 13:42:35.733525 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/fc5706c2-8b7d-439a-95b3-981907360c92-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"fc5706c2-8b7d-439a-95b3-981907360c92\") " pod="openstack/rabbitmq-server-0" Nov 28 13:42:35 crc kubenswrapper[4631]: I1128 13:42:35.733566 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/014519bc-68f0-4132-9a46-6b0e5c538c7e-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"014519bc-68f0-4132-9a46-6b0e5c538c7e\") " pod="openstack/rabbitmq-cell1-server-0" Nov 28 13:42:35 crc kubenswrapper[4631]: I1128 13:42:35.733591 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/014519bc-68f0-4132-9a46-6b0e5c538c7e-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"014519bc-68f0-4132-9a46-6b0e5c538c7e\") " pod="openstack/rabbitmq-cell1-server-0" Nov 28 13:42:35 crc kubenswrapper[4631]: I1128 13:42:35.733620 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/014519bc-68f0-4132-9a46-6b0e5c538c7e-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"014519bc-68f0-4132-9a46-6b0e5c538c7e\") " pod="openstack/rabbitmq-cell1-server-0" Nov 28 13:42:35 crc kubenswrapper[4631]: I1128 13:42:35.733657 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/fc5706c2-8b7d-439a-95b3-981907360c92-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"fc5706c2-8b7d-439a-95b3-981907360c92\") " pod="openstack/rabbitmq-server-0" Nov 28 13:42:35 crc kubenswrapper[4631]: I1128 13:42:35.733681 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/014519bc-68f0-4132-9a46-6b0e5c538c7e-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"014519bc-68f0-4132-9a46-6b0e5c538c7e\") " pod="openstack/rabbitmq-cell1-server-0" Nov 28 13:42:35 crc kubenswrapper[4631]: I1128 13:42:35.733701 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-server-0\" (UID: \"fc5706c2-8b7d-439a-95b3-981907360c92\") " pod="openstack/rabbitmq-server-0" Nov 28 13:42:35 crc kubenswrapper[4631]: I1128 13:42:35.733722 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/fc5706c2-8b7d-439a-95b3-981907360c92-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"fc5706c2-8b7d-439a-95b3-981907360c92\") " pod="openstack/rabbitmq-server-0" Nov 28 13:42:35 crc kubenswrapper[4631]: I1128 13:42:35.733740 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tvxxq\" (UniqueName: \"kubernetes.io/projected/014519bc-68f0-4132-9a46-6b0e5c538c7e-kube-api-access-tvxxq\") pod \"rabbitmq-cell1-server-0\" (UID: \"014519bc-68f0-4132-9a46-6b0e5c538c7e\") " pod="openstack/rabbitmq-cell1-server-0" Nov 28 13:42:35 crc kubenswrapper[4631]: I1128 13:42:35.733808 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/fc5706c2-8b7d-439a-95b3-981907360c92-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"fc5706c2-8b7d-439a-95b3-981907360c92\") " pod="openstack/rabbitmq-server-0" Nov 28 13:42:35 crc kubenswrapper[4631]: I1128 13:42:35.733837 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/014519bc-68f0-4132-9a46-6b0e5c538c7e-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"014519bc-68f0-4132-9a46-6b0e5c538c7e\") " pod="openstack/rabbitmq-cell1-server-0" Nov 28 13:42:35 crc kubenswrapper[4631]: I1128 13:42:35.733860 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"014519bc-68f0-4132-9a46-6b0e5c538c7e\") " pod="openstack/rabbitmq-cell1-server-0" Nov 28 13:42:35 crc kubenswrapper[4631]: I1128 13:42:35.734525 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/fc5706c2-8b7d-439a-95b3-981907360c92-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"fc5706c2-8b7d-439a-95b3-981907360c92\") " pod="openstack/rabbitmq-server-0" Nov 28 13:42:35 crc kubenswrapper[4631]: I1128 13:42:35.735091 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/fc5706c2-8b7d-439a-95b3-981907360c92-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"fc5706c2-8b7d-439a-95b3-981907360c92\") " pod="openstack/rabbitmq-server-0" Nov 28 13:42:35 crc kubenswrapper[4631]: I1128 13:42:35.735372 4631 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-server-0\" (UID: \"fc5706c2-8b7d-439a-95b3-981907360c92\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/rabbitmq-server-0" Nov 28 13:42:35 crc kubenswrapper[4631]: I1128 13:42:35.737679 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/fc5706c2-8b7d-439a-95b3-981907360c92-server-conf\") pod \"rabbitmq-server-0\" (UID: \"fc5706c2-8b7d-439a-95b3-981907360c92\") " pod="openstack/rabbitmq-server-0" Nov 28 13:42:35 crc kubenswrapper[4631]: I1128 13:42:35.737803 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/fc5706c2-8b7d-439a-95b3-981907360c92-config-data\") pod \"rabbitmq-server-0\" (UID: \"fc5706c2-8b7d-439a-95b3-981907360c92\") " pod="openstack/rabbitmq-server-0" Nov 28 13:42:35 crc kubenswrapper[4631]: I1128 13:42:35.737812 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/fc5706c2-8b7d-439a-95b3-981907360c92-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"fc5706c2-8b7d-439a-95b3-981907360c92\") " pod="openstack/rabbitmq-server-0" Nov 28 13:42:35 crc kubenswrapper[4631]: I1128 13:42:35.741031 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/fc5706c2-8b7d-439a-95b3-981907360c92-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"fc5706c2-8b7d-439a-95b3-981907360c92\") " pod="openstack/rabbitmq-server-0" Nov 28 13:42:35 crc kubenswrapper[4631]: I1128 13:42:35.744982 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/fc5706c2-8b7d-439a-95b3-981907360c92-pod-info\") pod \"rabbitmq-server-0\" (UID: \"fc5706c2-8b7d-439a-95b3-981907360c92\") " pod="openstack/rabbitmq-server-0" Nov 28 13:42:35 crc kubenswrapper[4631]: I1128 13:42:35.752865 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/fc5706c2-8b7d-439a-95b3-981907360c92-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"fc5706c2-8b7d-439a-95b3-981907360c92\") " pod="openstack/rabbitmq-server-0" Nov 28 13:42:35 crc kubenswrapper[4631]: I1128 13:42:35.756275 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/fc5706c2-8b7d-439a-95b3-981907360c92-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"fc5706c2-8b7d-439a-95b3-981907360c92\") " pod="openstack/rabbitmq-server-0" Nov 28 13:42:35 crc kubenswrapper[4631]: I1128 13:42:35.762453 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qmhk6\" (UniqueName: \"kubernetes.io/projected/fc5706c2-8b7d-439a-95b3-981907360c92-kube-api-access-qmhk6\") pod \"rabbitmq-server-0\" (UID: \"fc5706c2-8b7d-439a-95b3-981907360c92\") " pod="openstack/rabbitmq-server-0" Nov 28 13:42:35 crc kubenswrapper[4631]: I1128 13:42:35.792208 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-server-0\" (UID: \"fc5706c2-8b7d-439a-95b3-981907360c92\") " pod="openstack/rabbitmq-server-0" Nov 28 13:42:35 crc kubenswrapper[4631]: I1128 13:42:35.836243 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/014519bc-68f0-4132-9a46-6b0e5c538c7e-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"014519bc-68f0-4132-9a46-6b0e5c538c7e\") " pod="openstack/rabbitmq-cell1-server-0" Nov 28 13:42:35 crc kubenswrapper[4631]: I1128 13:42:35.836326 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tvxxq\" (UniqueName: \"kubernetes.io/projected/014519bc-68f0-4132-9a46-6b0e5c538c7e-kube-api-access-tvxxq\") pod \"rabbitmq-cell1-server-0\" (UID: \"014519bc-68f0-4132-9a46-6b0e5c538c7e\") " pod="openstack/rabbitmq-cell1-server-0" Nov 28 13:42:35 crc kubenswrapper[4631]: I1128 13:42:35.836365 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/014519bc-68f0-4132-9a46-6b0e5c538c7e-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"014519bc-68f0-4132-9a46-6b0e5c538c7e\") " pod="openstack/rabbitmq-cell1-server-0" Nov 28 13:42:35 crc kubenswrapper[4631]: I1128 13:42:35.836387 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"014519bc-68f0-4132-9a46-6b0e5c538c7e\") " pod="openstack/rabbitmq-cell1-server-0" Nov 28 13:42:35 crc kubenswrapper[4631]: I1128 13:42:35.836405 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/014519bc-68f0-4132-9a46-6b0e5c538c7e-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"014519bc-68f0-4132-9a46-6b0e5c538c7e\") " pod="openstack/rabbitmq-cell1-server-0" Nov 28 13:42:35 crc kubenswrapper[4631]: I1128 13:42:35.836426 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/014519bc-68f0-4132-9a46-6b0e5c538c7e-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"014519bc-68f0-4132-9a46-6b0e5c538c7e\") " pod="openstack/rabbitmq-cell1-server-0" Nov 28 13:42:35 crc kubenswrapper[4631]: I1128 13:42:35.836453 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/014519bc-68f0-4132-9a46-6b0e5c538c7e-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"014519bc-68f0-4132-9a46-6b0e5c538c7e\") " pod="openstack/rabbitmq-cell1-server-0" Nov 28 13:42:35 crc kubenswrapper[4631]: I1128 13:42:35.836515 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/014519bc-68f0-4132-9a46-6b0e5c538c7e-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"014519bc-68f0-4132-9a46-6b0e5c538c7e\") " pod="openstack/rabbitmq-cell1-server-0" Nov 28 13:42:35 crc kubenswrapper[4631]: I1128 13:42:35.836568 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/014519bc-68f0-4132-9a46-6b0e5c538c7e-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"014519bc-68f0-4132-9a46-6b0e5c538c7e\") " pod="openstack/rabbitmq-cell1-server-0" Nov 28 13:42:35 crc kubenswrapper[4631]: I1128 13:42:35.836594 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/014519bc-68f0-4132-9a46-6b0e5c538c7e-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"014519bc-68f0-4132-9a46-6b0e5c538c7e\") " pod="openstack/rabbitmq-cell1-server-0" Nov 28 13:42:35 crc kubenswrapper[4631]: I1128 13:42:35.836616 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/014519bc-68f0-4132-9a46-6b0e5c538c7e-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"014519bc-68f0-4132-9a46-6b0e5c538c7e\") " pod="openstack/rabbitmq-cell1-server-0" Nov 28 13:42:35 crc kubenswrapper[4631]: I1128 13:42:35.837901 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/014519bc-68f0-4132-9a46-6b0e5c538c7e-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"014519bc-68f0-4132-9a46-6b0e5c538c7e\") " pod="openstack/rabbitmq-cell1-server-0" Nov 28 13:42:35 crc kubenswrapper[4631]: I1128 13:42:35.838647 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/014519bc-68f0-4132-9a46-6b0e5c538c7e-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"014519bc-68f0-4132-9a46-6b0e5c538c7e\") " pod="openstack/rabbitmq-cell1-server-0" Nov 28 13:42:35 crc kubenswrapper[4631]: I1128 13:42:35.839237 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/014519bc-68f0-4132-9a46-6b0e5c538c7e-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"014519bc-68f0-4132-9a46-6b0e5c538c7e\") " pod="openstack/rabbitmq-cell1-server-0" Nov 28 13:42:35 crc kubenswrapper[4631]: I1128 13:42:35.839520 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/014519bc-68f0-4132-9a46-6b0e5c538c7e-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"014519bc-68f0-4132-9a46-6b0e5c538c7e\") " pod="openstack/rabbitmq-cell1-server-0" Nov 28 13:42:35 crc kubenswrapper[4631]: I1128 13:42:35.839518 4631 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"014519bc-68f0-4132-9a46-6b0e5c538c7e\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/rabbitmq-cell1-server-0" Nov 28 13:42:35 crc kubenswrapper[4631]: I1128 13:42:35.839829 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/014519bc-68f0-4132-9a46-6b0e5c538c7e-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"014519bc-68f0-4132-9a46-6b0e5c538c7e\") " pod="openstack/rabbitmq-cell1-server-0" Nov 28 13:42:35 crc kubenswrapper[4631]: I1128 13:42:35.842592 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/014519bc-68f0-4132-9a46-6b0e5c538c7e-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"014519bc-68f0-4132-9a46-6b0e5c538c7e\") " pod="openstack/rabbitmq-cell1-server-0" Nov 28 13:42:35 crc kubenswrapper[4631]: I1128 13:42:35.843691 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/014519bc-68f0-4132-9a46-6b0e5c538c7e-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"014519bc-68f0-4132-9a46-6b0e5c538c7e\") " pod="openstack/rabbitmq-cell1-server-0" Nov 28 13:42:35 crc kubenswrapper[4631]: I1128 13:42:35.847578 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/014519bc-68f0-4132-9a46-6b0e5c538c7e-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"014519bc-68f0-4132-9a46-6b0e5c538c7e\") " pod="openstack/rabbitmq-cell1-server-0" Nov 28 13:42:35 crc kubenswrapper[4631]: I1128 13:42:35.850628 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/014519bc-68f0-4132-9a46-6b0e5c538c7e-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"014519bc-68f0-4132-9a46-6b0e5c538c7e\") " pod="openstack/rabbitmq-cell1-server-0" Nov 28 13:42:35 crc kubenswrapper[4631]: I1128 13:42:35.858431 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tvxxq\" (UniqueName: \"kubernetes.io/projected/014519bc-68f0-4132-9a46-6b0e5c538c7e-kube-api-access-tvxxq\") pod \"rabbitmq-cell1-server-0\" (UID: \"014519bc-68f0-4132-9a46-6b0e5c538c7e\") " pod="openstack/rabbitmq-cell1-server-0" Nov 28 13:42:35 crc kubenswrapper[4631]: I1128 13:42:35.869050 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"014519bc-68f0-4132-9a46-6b0e5c538c7e\") " pod="openstack/rabbitmq-cell1-server-0" Nov 28 13:42:35 crc kubenswrapper[4631]: I1128 13:42:35.911736 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 28 13:42:35 crc kubenswrapper[4631]: I1128 13:42:35.985495 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 28 13:42:36 crc kubenswrapper[4631]: I1128 13:42:36.267073 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 28 13:42:36 crc kubenswrapper[4631]: I1128 13:42:36.426639 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"fc5706c2-8b7d-439a-95b3-981907360c92","Type":"ContainerStarted","Data":"fce268ce457e26331dce6d72209b8a26b9d39e9642e9671db73a12479391bfdb"} Nov 28 13:42:36 crc kubenswrapper[4631]: I1128 13:42:36.630547 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 28 13:42:36 crc kubenswrapper[4631]: W1128 13:42:36.635132 4631 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod014519bc_68f0_4132_9a46_6b0e5c538c7e.slice/crio-3c2c087d3ae73b9809a82ccd59791f272ef2c65bc0a6a3bf12b7941ec9457dbd WatchSource:0}: Error finding container 3c2c087d3ae73b9809a82ccd59791f272ef2c65bc0a6a3bf12b7941ec9457dbd: Status 404 returned error can't find the container with id 3c2c087d3ae73b9809a82ccd59791f272ef2c65bc0a6a3bf12b7941ec9457dbd Nov 28 13:42:37 crc kubenswrapper[4631]: I1128 13:42:37.253625 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-79bd4cc8c9-4gxf7"] Nov 28 13:42:37 crc kubenswrapper[4631]: I1128 13:42:37.255755 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-79bd4cc8c9-4gxf7" Nov 28 13:42:37 crc kubenswrapper[4631]: I1128 13:42:37.261728 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-edpm-ipam" Nov 28 13:42:37 crc kubenswrapper[4631]: I1128 13:42:37.280392 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-79bd4cc8c9-4gxf7"] Nov 28 13:42:37 crc kubenswrapper[4631]: I1128 13:42:37.378215 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9fk57\" (UniqueName: \"kubernetes.io/projected/7179a7f7-1ee8-40a2-abcd-0cd373575b82-kube-api-access-9fk57\") pod \"dnsmasq-dns-79bd4cc8c9-4gxf7\" (UID: \"7179a7f7-1ee8-40a2-abcd-0cd373575b82\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-4gxf7" Nov 28 13:42:37 crc kubenswrapper[4631]: I1128 13:42:37.378513 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7179a7f7-1ee8-40a2-abcd-0cd373575b82-config\") pod \"dnsmasq-dns-79bd4cc8c9-4gxf7\" (UID: \"7179a7f7-1ee8-40a2-abcd-0cd373575b82\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-4gxf7" Nov 28 13:42:37 crc kubenswrapper[4631]: I1128 13:42:37.378603 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7179a7f7-1ee8-40a2-abcd-0cd373575b82-dns-svc\") pod \"dnsmasq-dns-79bd4cc8c9-4gxf7\" (UID: \"7179a7f7-1ee8-40a2-abcd-0cd373575b82\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-4gxf7" Nov 28 13:42:37 crc kubenswrapper[4631]: I1128 13:42:37.378729 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7179a7f7-1ee8-40a2-abcd-0cd373575b82-ovsdbserver-sb\") pod \"dnsmasq-dns-79bd4cc8c9-4gxf7\" (UID: \"7179a7f7-1ee8-40a2-abcd-0cd373575b82\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-4gxf7" Nov 28 13:42:37 crc kubenswrapper[4631]: I1128 13:42:37.378820 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/7179a7f7-1ee8-40a2-abcd-0cd373575b82-openstack-edpm-ipam\") pod \"dnsmasq-dns-79bd4cc8c9-4gxf7\" (UID: \"7179a7f7-1ee8-40a2-abcd-0cd373575b82\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-4gxf7" Nov 28 13:42:37 crc kubenswrapper[4631]: I1128 13:42:37.378899 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7179a7f7-1ee8-40a2-abcd-0cd373575b82-ovsdbserver-nb\") pod \"dnsmasq-dns-79bd4cc8c9-4gxf7\" (UID: \"7179a7f7-1ee8-40a2-abcd-0cd373575b82\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-4gxf7" Nov 28 13:42:37 crc kubenswrapper[4631]: I1128 13:42:37.378962 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7179a7f7-1ee8-40a2-abcd-0cd373575b82-dns-swift-storage-0\") pod \"dnsmasq-dns-79bd4cc8c9-4gxf7\" (UID: \"7179a7f7-1ee8-40a2-abcd-0cd373575b82\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-4gxf7" Nov 28 13:42:37 crc kubenswrapper[4631]: I1128 13:42:37.436728 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"014519bc-68f0-4132-9a46-6b0e5c538c7e","Type":"ContainerStarted","Data":"3c2c087d3ae73b9809a82ccd59791f272ef2c65bc0a6a3bf12b7941ec9457dbd"} Nov 28 13:42:37 crc kubenswrapper[4631]: I1128 13:42:37.481358 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7179a7f7-1ee8-40a2-abcd-0cd373575b82-ovsdbserver-sb\") pod \"dnsmasq-dns-79bd4cc8c9-4gxf7\" (UID: \"7179a7f7-1ee8-40a2-abcd-0cd373575b82\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-4gxf7" Nov 28 13:42:37 crc kubenswrapper[4631]: I1128 13:42:37.481411 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/7179a7f7-1ee8-40a2-abcd-0cd373575b82-openstack-edpm-ipam\") pod \"dnsmasq-dns-79bd4cc8c9-4gxf7\" (UID: \"7179a7f7-1ee8-40a2-abcd-0cd373575b82\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-4gxf7" Nov 28 13:42:37 crc kubenswrapper[4631]: I1128 13:42:37.481439 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7179a7f7-1ee8-40a2-abcd-0cd373575b82-ovsdbserver-nb\") pod \"dnsmasq-dns-79bd4cc8c9-4gxf7\" (UID: \"7179a7f7-1ee8-40a2-abcd-0cd373575b82\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-4gxf7" Nov 28 13:42:37 crc kubenswrapper[4631]: I1128 13:42:37.481455 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7179a7f7-1ee8-40a2-abcd-0cd373575b82-dns-swift-storage-0\") pod \"dnsmasq-dns-79bd4cc8c9-4gxf7\" (UID: \"7179a7f7-1ee8-40a2-abcd-0cd373575b82\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-4gxf7" Nov 28 13:42:37 crc kubenswrapper[4631]: I1128 13:42:37.481562 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9fk57\" (UniqueName: \"kubernetes.io/projected/7179a7f7-1ee8-40a2-abcd-0cd373575b82-kube-api-access-9fk57\") pod \"dnsmasq-dns-79bd4cc8c9-4gxf7\" (UID: \"7179a7f7-1ee8-40a2-abcd-0cd373575b82\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-4gxf7" Nov 28 13:42:37 crc kubenswrapper[4631]: I1128 13:42:37.481599 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7179a7f7-1ee8-40a2-abcd-0cd373575b82-config\") pod \"dnsmasq-dns-79bd4cc8c9-4gxf7\" (UID: \"7179a7f7-1ee8-40a2-abcd-0cd373575b82\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-4gxf7" Nov 28 13:42:37 crc kubenswrapper[4631]: I1128 13:42:37.481632 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7179a7f7-1ee8-40a2-abcd-0cd373575b82-dns-svc\") pod \"dnsmasq-dns-79bd4cc8c9-4gxf7\" (UID: \"7179a7f7-1ee8-40a2-abcd-0cd373575b82\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-4gxf7" Nov 28 13:42:37 crc kubenswrapper[4631]: I1128 13:42:37.482512 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7179a7f7-1ee8-40a2-abcd-0cd373575b82-dns-svc\") pod \"dnsmasq-dns-79bd4cc8c9-4gxf7\" (UID: \"7179a7f7-1ee8-40a2-abcd-0cd373575b82\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-4gxf7" Nov 28 13:42:37 crc kubenswrapper[4631]: I1128 13:42:37.483038 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7179a7f7-1ee8-40a2-abcd-0cd373575b82-ovsdbserver-sb\") pod \"dnsmasq-dns-79bd4cc8c9-4gxf7\" (UID: \"7179a7f7-1ee8-40a2-abcd-0cd373575b82\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-4gxf7" Nov 28 13:42:37 crc kubenswrapper[4631]: I1128 13:42:37.483547 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/7179a7f7-1ee8-40a2-abcd-0cd373575b82-openstack-edpm-ipam\") pod \"dnsmasq-dns-79bd4cc8c9-4gxf7\" (UID: \"7179a7f7-1ee8-40a2-abcd-0cd373575b82\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-4gxf7" Nov 28 13:42:37 crc kubenswrapper[4631]: I1128 13:42:37.484107 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7179a7f7-1ee8-40a2-abcd-0cd373575b82-ovsdbserver-nb\") pod \"dnsmasq-dns-79bd4cc8c9-4gxf7\" (UID: \"7179a7f7-1ee8-40a2-abcd-0cd373575b82\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-4gxf7" Nov 28 13:42:37 crc kubenswrapper[4631]: I1128 13:42:37.484634 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7179a7f7-1ee8-40a2-abcd-0cd373575b82-dns-swift-storage-0\") pod \"dnsmasq-dns-79bd4cc8c9-4gxf7\" (UID: \"7179a7f7-1ee8-40a2-abcd-0cd373575b82\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-4gxf7" Nov 28 13:42:37 crc kubenswrapper[4631]: I1128 13:42:37.485608 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7179a7f7-1ee8-40a2-abcd-0cd373575b82-config\") pod \"dnsmasq-dns-79bd4cc8c9-4gxf7\" (UID: \"7179a7f7-1ee8-40a2-abcd-0cd373575b82\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-4gxf7" Nov 28 13:42:37 crc kubenswrapper[4631]: I1128 13:42:37.526828 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9fk57\" (UniqueName: \"kubernetes.io/projected/7179a7f7-1ee8-40a2-abcd-0cd373575b82-kube-api-access-9fk57\") pod \"dnsmasq-dns-79bd4cc8c9-4gxf7\" (UID: \"7179a7f7-1ee8-40a2-abcd-0cd373575b82\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-4gxf7" Nov 28 13:42:37 crc kubenswrapper[4631]: I1128 13:42:37.533476 4631 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3a91e9fa-34b7-4688-9171-56a62043759d" path="/var/lib/kubelet/pods/3a91e9fa-34b7-4688-9171-56a62043759d/volumes" Nov 28 13:42:37 crc kubenswrapper[4631]: I1128 13:42:37.577913 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-79bd4cc8c9-4gxf7" Nov 28 13:42:38 crc kubenswrapper[4631]: I1128 13:42:38.150522 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-79bd4cc8c9-4gxf7"] Nov 28 13:42:38 crc kubenswrapper[4631]: W1128 13:42:38.220841 4631 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7179a7f7_1ee8_40a2_abcd_0cd373575b82.slice/crio-47f054a60628e02b8a4e2d376b8c43009ccc9ff6d84538cdb68450a089c966e9 WatchSource:0}: Error finding container 47f054a60628e02b8a4e2d376b8c43009ccc9ff6d84538cdb68450a089c966e9: Status 404 returned error can't find the container with id 47f054a60628e02b8a4e2d376b8c43009ccc9ff6d84538cdb68450a089c966e9 Nov 28 13:42:38 crc kubenswrapper[4631]: I1128 13:42:38.453537 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"014519bc-68f0-4132-9a46-6b0e5c538c7e","Type":"ContainerStarted","Data":"0ced32c9dda9f55eceb6c4e60abf6aa57be7ac93b6c4ef8ecac6bf27f0f3fbda"} Nov 28 13:42:38 crc kubenswrapper[4631]: I1128 13:42:38.457679 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"fc5706c2-8b7d-439a-95b3-981907360c92","Type":"ContainerStarted","Data":"400faf511747c4aecb8b263484297c384f64c4393d9b88fbc03ba79c021a9bee"} Nov 28 13:42:38 crc kubenswrapper[4631]: I1128 13:42:38.460049 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-79bd4cc8c9-4gxf7" event={"ID":"7179a7f7-1ee8-40a2-abcd-0cd373575b82","Type":"ContainerStarted","Data":"47f054a60628e02b8a4e2d376b8c43009ccc9ff6d84538cdb68450a089c966e9"} Nov 28 13:42:39 crc kubenswrapper[4631]: I1128 13:42:39.472168 4631 generic.go:334] "Generic (PLEG): container finished" podID="7179a7f7-1ee8-40a2-abcd-0cd373575b82" containerID="28d9cc47a5644ecb10afdbc2123a96f22c954216480070dffda06fa445deae2e" exitCode=0 Nov 28 13:42:39 crc kubenswrapper[4631]: I1128 13:42:39.472416 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-79bd4cc8c9-4gxf7" event={"ID":"7179a7f7-1ee8-40a2-abcd-0cd373575b82","Type":"ContainerDied","Data":"28d9cc47a5644ecb10afdbc2123a96f22c954216480070dffda06fa445deae2e"} Nov 28 13:42:40 crc kubenswrapper[4631]: I1128 13:42:40.505344 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-79bd4cc8c9-4gxf7" event={"ID":"7179a7f7-1ee8-40a2-abcd-0cd373575b82","Type":"ContainerStarted","Data":"8ad6f16d4e9ca7be24d22143e8977e1234b4a887dab4dbc3fc9568e901edc099"} Nov 28 13:42:40 crc kubenswrapper[4631]: I1128 13:42:40.507498 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-79bd4cc8c9-4gxf7" Nov 28 13:42:40 crc kubenswrapper[4631]: I1128 13:42:40.534561 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-79bd4cc8c9-4gxf7" podStartSLOduration=3.534519953 podStartE2EDuration="3.534519953s" podCreationTimestamp="2025-11-28 13:42:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 13:42:40.532886545 +0000 UTC m=+1317.340189899" watchObservedRunningTime="2025-11-28 13:42:40.534519953 +0000 UTC m=+1317.341823297" Nov 28 13:42:47 crc kubenswrapper[4631]: I1128 13:42:47.580673 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-79bd4cc8c9-4gxf7" Nov 28 13:42:47 crc kubenswrapper[4631]: I1128 13:42:47.689372 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-89c5cd4d5-qvl5b"] Nov 28 13:42:47 crc kubenswrapper[4631]: I1128 13:42:47.727322 4631 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-89c5cd4d5-qvl5b" podUID="931de4b9-12a1-45bf-b69e-b0498c6c770f" containerName="dnsmasq-dns" containerID="cri-o://4db6ded4946f647032503312a75a1009a07c3e44165a5c47a10ae0217d33bec5" gracePeriod=10 Nov 28 13:42:48 crc kubenswrapper[4631]: I1128 13:42:48.104889 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6cd9bffc9-x8stz"] Nov 28 13:42:48 crc kubenswrapper[4631]: I1128 13:42:48.107488 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6cd9bffc9-x8stz" Nov 28 13:42:48 crc kubenswrapper[4631]: I1128 13:42:48.173406 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6cd9bffc9-x8stz"] Nov 28 13:42:48 crc kubenswrapper[4631]: I1128 13:42:48.265634 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6b64d173-0e5b-41e3-b841-68fef937dfb4-ovsdbserver-sb\") pod \"dnsmasq-dns-6cd9bffc9-x8stz\" (UID: \"6b64d173-0e5b-41e3-b841-68fef937dfb4\") " pod="openstack/dnsmasq-dns-6cd9bffc9-x8stz" Nov 28 13:42:48 crc kubenswrapper[4631]: I1128 13:42:48.265713 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6b64d173-0e5b-41e3-b841-68fef937dfb4-ovsdbserver-nb\") pod \"dnsmasq-dns-6cd9bffc9-x8stz\" (UID: \"6b64d173-0e5b-41e3-b841-68fef937dfb4\") " pod="openstack/dnsmasq-dns-6cd9bffc9-x8stz" Nov 28 13:42:48 crc kubenswrapper[4631]: I1128 13:42:48.265749 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6b64d173-0e5b-41e3-b841-68fef937dfb4-dns-swift-storage-0\") pod \"dnsmasq-dns-6cd9bffc9-x8stz\" (UID: \"6b64d173-0e5b-41e3-b841-68fef937dfb4\") " pod="openstack/dnsmasq-dns-6cd9bffc9-x8stz" Nov 28 13:42:48 crc kubenswrapper[4631]: I1128 13:42:48.265806 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6b64d173-0e5b-41e3-b841-68fef937dfb4-dns-svc\") pod \"dnsmasq-dns-6cd9bffc9-x8stz\" (UID: \"6b64d173-0e5b-41e3-b841-68fef937dfb4\") " pod="openstack/dnsmasq-dns-6cd9bffc9-x8stz" Nov 28 13:42:48 crc kubenswrapper[4631]: I1128 13:42:48.265861 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zqkd8\" (UniqueName: \"kubernetes.io/projected/6b64d173-0e5b-41e3-b841-68fef937dfb4-kube-api-access-zqkd8\") pod \"dnsmasq-dns-6cd9bffc9-x8stz\" (UID: \"6b64d173-0e5b-41e3-b841-68fef937dfb4\") " pod="openstack/dnsmasq-dns-6cd9bffc9-x8stz" Nov 28 13:42:48 crc kubenswrapper[4631]: I1128 13:42:48.265880 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/6b64d173-0e5b-41e3-b841-68fef937dfb4-openstack-edpm-ipam\") pod \"dnsmasq-dns-6cd9bffc9-x8stz\" (UID: \"6b64d173-0e5b-41e3-b841-68fef937dfb4\") " pod="openstack/dnsmasq-dns-6cd9bffc9-x8stz" Nov 28 13:42:48 crc kubenswrapper[4631]: I1128 13:42:48.265894 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6b64d173-0e5b-41e3-b841-68fef937dfb4-config\") pod \"dnsmasq-dns-6cd9bffc9-x8stz\" (UID: \"6b64d173-0e5b-41e3-b841-68fef937dfb4\") " pod="openstack/dnsmasq-dns-6cd9bffc9-x8stz" Nov 28 13:42:48 crc kubenswrapper[4631]: I1128 13:42:48.367978 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6b64d173-0e5b-41e3-b841-68fef937dfb4-ovsdbserver-sb\") pod \"dnsmasq-dns-6cd9bffc9-x8stz\" (UID: \"6b64d173-0e5b-41e3-b841-68fef937dfb4\") " pod="openstack/dnsmasq-dns-6cd9bffc9-x8stz" Nov 28 13:42:48 crc kubenswrapper[4631]: I1128 13:42:48.368080 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6b64d173-0e5b-41e3-b841-68fef937dfb4-ovsdbserver-nb\") pod \"dnsmasq-dns-6cd9bffc9-x8stz\" (UID: \"6b64d173-0e5b-41e3-b841-68fef937dfb4\") " pod="openstack/dnsmasq-dns-6cd9bffc9-x8stz" Nov 28 13:42:48 crc kubenswrapper[4631]: I1128 13:42:48.368124 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6b64d173-0e5b-41e3-b841-68fef937dfb4-dns-swift-storage-0\") pod \"dnsmasq-dns-6cd9bffc9-x8stz\" (UID: \"6b64d173-0e5b-41e3-b841-68fef937dfb4\") " pod="openstack/dnsmasq-dns-6cd9bffc9-x8stz" Nov 28 13:42:48 crc kubenswrapper[4631]: I1128 13:42:48.368199 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6b64d173-0e5b-41e3-b841-68fef937dfb4-dns-svc\") pod \"dnsmasq-dns-6cd9bffc9-x8stz\" (UID: \"6b64d173-0e5b-41e3-b841-68fef937dfb4\") " pod="openstack/dnsmasq-dns-6cd9bffc9-x8stz" Nov 28 13:42:48 crc kubenswrapper[4631]: I1128 13:42:48.368299 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zqkd8\" (UniqueName: \"kubernetes.io/projected/6b64d173-0e5b-41e3-b841-68fef937dfb4-kube-api-access-zqkd8\") pod \"dnsmasq-dns-6cd9bffc9-x8stz\" (UID: \"6b64d173-0e5b-41e3-b841-68fef937dfb4\") " pod="openstack/dnsmasq-dns-6cd9bffc9-x8stz" Nov 28 13:42:48 crc kubenswrapper[4631]: I1128 13:42:48.368358 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/6b64d173-0e5b-41e3-b841-68fef937dfb4-openstack-edpm-ipam\") pod \"dnsmasq-dns-6cd9bffc9-x8stz\" (UID: \"6b64d173-0e5b-41e3-b841-68fef937dfb4\") " pod="openstack/dnsmasq-dns-6cd9bffc9-x8stz" Nov 28 13:42:48 crc kubenswrapper[4631]: I1128 13:42:48.368387 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6b64d173-0e5b-41e3-b841-68fef937dfb4-config\") pod \"dnsmasq-dns-6cd9bffc9-x8stz\" (UID: \"6b64d173-0e5b-41e3-b841-68fef937dfb4\") " pod="openstack/dnsmasq-dns-6cd9bffc9-x8stz" Nov 28 13:42:48 crc kubenswrapper[4631]: I1128 13:42:48.369771 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6b64d173-0e5b-41e3-b841-68fef937dfb4-dns-swift-storage-0\") pod \"dnsmasq-dns-6cd9bffc9-x8stz\" (UID: \"6b64d173-0e5b-41e3-b841-68fef937dfb4\") " pod="openstack/dnsmasq-dns-6cd9bffc9-x8stz" Nov 28 13:42:48 crc kubenswrapper[4631]: I1128 13:42:48.370057 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6b64d173-0e5b-41e3-b841-68fef937dfb4-config\") pod \"dnsmasq-dns-6cd9bffc9-x8stz\" (UID: \"6b64d173-0e5b-41e3-b841-68fef937dfb4\") " pod="openstack/dnsmasq-dns-6cd9bffc9-x8stz" Nov 28 13:42:48 crc kubenswrapper[4631]: I1128 13:42:48.370461 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6b64d173-0e5b-41e3-b841-68fef937dfb4-ovsdbserver-sb\") pod \"dnsmasq-dns-6cd9bffc9-x8stz\" (UID: \"6b64d173-0e5b-41e3-b841-68fef937dfb4\") " pod="openstack/dnsmasq-dns-6cd9bffc9-x8stz" Nov 28 13:42:48 crc kubenswrapper[4631]: I1128 13:42:48.370877 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6b64d173-0e5b-41e3-b841-68fef937dfb4-dns-svc\") pod \"dnsmasq-dns-6cd9bffc9-x8stz\" (UID: \"6b64d173-0e5b-41e3-b841-68fef937dfb4\") " pod="openstack/dnsmasq-dns-6cd9bffc9-x8stz" Nov 28 13:42:48 crc kubenswrapper[4631]: I1128 13:42:48.371011 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6b64d173-0e5b-41e3-b841-68fef937dfb4-ovsdbserver-nb\") pod \"dnsmasq-dns-6cd9bffc9-x8stz\" (UID: \"6b64d173-0e5b-41e3-b841-68fef937dfb4\") " pod="openstack/dnsmasq-dns-6cd9bffc9-x8stz" Nov 28 13:42:48 crc kubenswrapper[4631]: I1128 13:42:48.371627 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/6b64d173-0e5b-41e3-b841-68fef937dfb4-openstack-edpm-ipam\") pod \"dnsmasq-dns-6cd9bffc9-x8stz\" (UID: \"6b64d173-0e5b-41e3-b841-68fef937dfb4\") " pod="openstack/dnsmasq-dns-6cd9bffc9-x8stz" Nov 28 13:42:48 crc kubenswrapper[4631]: I1128 13:42:48.411656 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zqkd8\" (UniqueName: \"kubernetes.io/projected/6b64d173-0e5b-41e3-b841-68fef937dfb4-kube-api-access-zqkd8\") pod \"dnsmasq-dns-6cd9bffc9-x8stz\" (UID: \"6b64d173-0e5b-41e3-b841-68fef937dfb4\") " pod="openstack/dnsmasq-dns-6cd9bffc9-x8stz" Nov 28 13:42:48 crc kubenswrapper[4631]: I1128 13:42:48.495175 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-89c5cd4d5-qvl5b" Nov 28 13:42:48 crc kubenswrapper[4631]: I1128 13:42:48.499168 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6cd9bffc9-x8stz" Nov 28 13:42:48 crc kubenswrapper[4631]: I1128 13:42:48.585562 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z6dbm\" (UniqueName: \"kubernetes.io/projected/931de4b9-12a1-45bf-b69e-b0498c6c770f-kube-api-access-z6dbm\") pod \"931de4b9-12a1-45bf-b69e-b0498c6c770f\" (UID: \"931de4b9-12a1-45bf-b69e-b0498c6c770f\") " Nov 28 13:42:48 crc kubenswrapper[4631]: I1128 13:42:48.585671 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/931de4b9-12a1-45bf-b69e-b0498c6c770f-config\") pod \"931de4b9-12a1-45bf-b69e-b0498c6c770f\" (UID: \"931de4b9-12a1-45bf-b69e-b0498c6c770f\") " Nov 28 13:42:48 crc kubenswrapper[4631]: I1128 13:42:48.585797 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/931de4b9-12a1-45bf-b69e-b0498c6c770f-ovsdbserver-nb\") pod \"931de4b9-12a1-45bf-b69e-b0498c6c770f\" (UID: \"931de4b9-12a1-45bf-b69e-b0498c6c770f\") " Nov 28 13:42:48 crc kubenswrapper[4631]: I1128 13:42:48.585905 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/931de4b9-12a1-45bf-b69e-b0498c6c770f-ovsdbserver-sb\") pod \"931de4b9-12a1-45bf-b69e-b0498c6c770f\" (UID: \"931de4b9-12a1-45bf-b69e-b0498c6c770f\") " Nov 28 13:42:48 crc kubenswrapper[4631]: I1128 13:42:48.585924 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/931de4b9-12a1-45bf-b69e-b0498c6c770f-dns-swift-storage-0\") pod \"931de4b9-12a1-45bf-b69e-b0498c6c770f\" (UID: \"931de4b9-12a1-45bf-b69e-b0498c6c770f\") " Nov 28 13:42:48 crc kubenswrapper[4631]: I1128 13:42:48.585959 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/931de4b9-12a1-45bf-b69e-b0498c6c770f-dns-svc\") pod \"931de4b9-12a1-45bf-b69e-b0498c6c770f\" (UID: \"931de4b9-12a1-45bf-b69e-b0498c6c770f\") " Nov 28 13:42:48 crc kubenswrapper[4631]: I1128 13:42:48.598467 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/931de4b9-12a1-45bf-b69e-b0498c6c770f-kube-api-access-z6dbm" (OuterVolumeSpecName: "kube-api-access-z6dbm") pod "931de4b9-12a1-45bf-b69e-b0498c6c770f" (UID: "931de4b9-12a1-45bf-b69e-b0498c6c770f"). InnerVolumeSpecName "kube-api-access-z6dbm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:42:48 crc kubenswrapper[4631]: I1128 13:42:48.706383 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z6dbm\" (UniqueName: \"kubernetes.io/projected/931de4b9-12a1-45bf-b69e-b0498c6c770f-kube-api-access-z6dbm\") on node \"crc\" DevicePath \"\"" Nov 28 13:42:48 crc kubenswrapper[4631]: I1128 13:42:48.739682 4631 generic.go:334] "Generic (PLEG): container finished" podID="931de4b9-12a1-45bf-b69e-b0498c6c770f" containerID="4db6ded4946f647032503312a75a1009a07c3e44165a5c47a10ae0217d33bec5" exitCode=0 Nov 28 13:42:48 crc kubenswrapper[4631]: I1128 13:42:48.739738 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-89c5cd4d5-qvl5b" event={"ID":"931de4b9-12a1-45bf-b69e-b0498c6c770f","Type":"ContainerDied","Data":"4db6ded4946f647032503312a75a1009a07c3e44165a5c47a10ae0217d33bec5"} Nov 28 13:42:48 crc kubenswrapper[4631]: I1128 13:42:48.739772 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-89c5cd4d5-qvl5b" event={"ID":"931de4b9-12a1-45bf-b69e-b0498c6c770f","Type":"ContainerDied","Data":"2e70e7cb3fcb57c4c650696f213c0b32fe875a1662c34d60be6bbd8b0a354173"} Nov 28 13:42:48 crc kubenswrapper[4631]: I1128 13:42:48.739794 4631 scope.go:117] "RemoveContainer" containerID="4db6ded4946f647032503312a75a1009a07c3e44165a5c47a10ae0217d33bec5" Nov 28 13:42:48 crc kubenswrapper[4631]: I1128 13:42:48.740000 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-89c5cd4d5-qvl5b" Nov 28 13:42:48 crc kubenswrapper[4631]: I1128 13:42:48.770409 4631 scope.go:117] "RemoveContainer" containerID="584175cb653c5b5b9801311291b1034c9cab49768b9f11cbf85a176a8b325e03" Nov 28 13:42:48 crc kubenswrapper[4631]: I1128 13:42:48.794218 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/931de4b9-12a1-45bf-b69e-b0498c6c770f-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "931de4b9-12a1-45bf-b69e-b0498c6c770f" (UID: "931de4b9-12a1-45bf-b69e-b0498c6c770f"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 13:42:48 crc kubenswrapper[4631]: I1128 13:42:48.812203 4631 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/931de4b9-12a1-45bf-b69e-b0498c6c770f-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 28 13:42:48 crc kubenswrapper[4631]: I1128 13:42:48.813647 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/931de4b9-12a1-45bf-b69e-b0498c6c770f-config" (OuterVolumeSpecName: "config") pod "931de4b9-12a1-45bf-b69e-b0498c6c770f" (UID: "931de4b9-12a1-45bf-b69e-b0498c6c770f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 13:42:48 crc kubenswrapper[4631]: I1128 13:42:48.839424 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/931de4b9-12a1-45bf-b69e-b0498c6c770f-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "931de4b9-12a1-45bf-b69e-b0498c6c770f" (UID: "931de4b9-12a1-45bf-b69e-b0498c6c770f"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 13:42:48 crc kubenswrapper[4631]: I1128 13:42:48.840194 4631 scope.go:117] "RemoveContainer" containerID="4db6ded4946f647032503312a75a1009a07c3e44165a5c47a10ae0217d33bec5" Nov 28 13:42:48 crc kubenswrapper[4631]: E1128 13:42:48.846271 4631 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4db6ded4946f647032503312a75a1009a07c3e44165a5c47a10ae0217d33bec5\": container with ID starting with 4db6ded4946f647032503312a75a1009a07c3e44165a5c47a10ae0217d33bec5 not found: ID does not exist" containerID="4db6ded4946f647032503312a75a1009a07c3e44165a5c47a10ae0217d33bec5" Nov 28 13:42:48 crc kubenswrapper[4631]: I1128 13:42:48.846346 4631 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4db6ded4946f647032503312a75a1009a07c3e44165a5c47a10ae0217d33bec5"} err="failed to get container status \"4db6ded4946f647032503312a75a1009a07c3e44165a5c47a10ae0217d33bec5\": rpc error: code = NotFound desc = could not find container \"4db6ded4946f647032503312a75a1009a07c3e44165a5c47a10ae0217d33bec5\": container with ID starting with 4db6ded4946f647032503312a75a1009a07c3e44165a5c47a10ae0217d33bec5 not found: ID does not exist" Nov 28 13:42:48 crc kubenswrapper[4631]: I1128 13:42:48.846389 4631 scope.go:117] "RemoveContainer" containerID="584175cb653c5b5b9801311291b1034c9cab49768b9f11cbf85a176a8b325e03" Nov 28 13:42:48 crc kubenswrapper[4631]: I1128 13:42:48.859053 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/931de4b9-12a1-45bf-b69e-b0498c6c770f-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "931de4b9-12a1-45bf-b69e-b0498c6c770f" (UID: "931de4b9-12a1-45bf-b69e-b0498c6c770f"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 13:42:48 crc kubenswrapper[4631]: E1128 13:42:48.859207 4631 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"584175cb653c5b5b9801311291b1034c9cab49768b9f11cbf85a176a8b325e03\": container with ID starting with 584175cb653c5b5b9801311291b1034c9cab49768b9f11cbf85a176a8b325e03 not found: ID does not exist" containerID="584175cb653c5b5b9801311291b1034c9cab49768b9f11cbf85a176a8b325e03" Nov 28 13:42:48 crc kubenswrapper[4631]: I1128 13:42:48.859247 4631 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"584175cb653c5b5b9801311291b1034c9cab49768b9f11cbf85a176a8b325e03"} err="failed to get container status \"584175cb653c5b5b9801311291b1034c9cab49768b9f11cbf85a176a8b325e03\": rpc error: code = NotFound desc = could not find container \"584175cb653c5b5b9801311291b1034c9cab49768b9f11cbf85a176a8b325e03\": container with ID starting with 584175cb653c5b5b9801311291b1034c9cab49768b9f11cbf85a176a8b325e03 not found: ID does not exist" Nov 28 13:42:48 crc kubenswrapper[4631]: I1128 13:42:48.870198 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/931de4b9-12a1-45bf-b69e-b0498c6c770f-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "931de4b9-12a1-45bf-b69e-b0498c6c770f" (UID: "931de4b9-12a1-45bf-b69e-b0498c6c770f"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 13:42:48 crc kubenswrapper[4631]: I1128 13:42:48.915228 4631 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/931de4b9-12a1-45bf-b69e-b0498c6c770f-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 28 13:42:48 crc kubenswrapper[4631]: I1128 13:42:48.915714 4631 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/931de4b9-12a1-45bf-b69e-b0498c6c770f-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 28 13:42:48 crc kubenswrapper[4631]: I1128 13:42:48.915728 4631 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/931de4b9-12a1-45bf-b69e-b0498c6c770f-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 28 13:42:48 crc kubenswrapper[4631]: I1128 13:42:48.915738 4631 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/931de4b9-12a1-45bf-b69e-b0498c6c770f-config\") on node \"crc\" DevicePath \"\"" Nov 28 13:42:49 crc kubenswrapper[4631]: I1128 13:42:49.088934 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-89c5cd4d5-qvl5b"] Nov 28 13:42:49 crc kubenswrapper[4631]: I1128 13:42:49.102147 4631 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-89c5cd4d5-qvl5b"] Nov 28 13:42:49 crc kubenswrapper[4631]: I1128 13:42:49.196137 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6cd9bffc9-x8stz"] Nov 28 13:42:49 crc kubenswrapper[4631]: W1128 13:42:49.223759 4631 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6b64d173_0e5b_41e3_b841_68fef937dfb4.slice/crio-3a660291042c9c1cec0d6f4fc5a12a01a47f940315ecce4e26f15b55d96f17c6 WatchSource:0}: Error finding container 3a660291042c9c1cec0d6f4fc5a12a01a47f940315ecce4e26f15b55d96f17c6: Status 404 returned error can't find the container with id 3a660291042c9c1cec0d6f4fc5a12a01a47f940315ecce4e26f15b55d96f17c6 Nov 28 13:42:49 crc kubenswrapper[4631]: I1128 13:42:49.533447 4631 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="931de4b9-12a1-45bf-b69e-b0498c6c770f" path="/var/lib/kubelet/pods/931de4b9-12a1-45bf-b69e-b0498c6c770f/volumes" Nov 28 13:42:49 crc kubenswrapper[4631]: I1128 13:42:49.751774 4631 generic.go:334] "Generic (PLEG): container finished" podID="6b64d173-0e5b-41e3-b841-68fef937dfb4" containerID="38bb91ef4c9dd4b7238c92a5f3d4a04b6b3ab2fc09a62e2e41c559fdf9c4fb8d" exitCode=0 Nov 28 13:42:49 crc kubenswrapper[4631]: I1128 13:42:49.751869 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6cd9bffc9-x8stz" event={"ID":"6b64d173-0e5b-41e3-b841-68fef937dfb4","Type":"ContainerDied","Data":"38bb91ef4c9dd4b7238c92a5f3d4a04b6b3ab2fc09a62e2e41c559fdf9c4fb8d"} Nov 28 13:42:49 crc kubenswrapper[4631]: I1128 13:42:49.752657 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6cd9bffc9-x8stz" event={"ID":"6b64d173-0e5b-41e3-b841-68fef937dfb4","Type":"ContainerStarted","Data":"3a660291042c9c1cec0d6f4fc5a12a01a47f940315ecce4e26f15b55d96f17c6"} Nov 28 13:42:50 crc kubenswrapper[4631]: I1128 13:42:50.773228 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6cd9bffc9-x8stz" event={"ID":"6b64d173-0e5b-41e3-b841-68fef937dfb4","Type":"ContainerStarted","Data":"b8f9eb4864dbed8d35b3c1da7dda1e533d2d070a3cc85940b419a2f9557ae322"} Nov 28 13:42:50 crc kubenswrapper[4631]: I1128 13:42:50.773943 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6cd9bffc9-x8stz" Nov 28 13:42:50 crc kubenswrapper[4631]: I1128 13:42:50.818466 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6cd9bffc9-x8stz" podStartSLOduration=3.8184179719999998 podStartE2EDuration="3.818417972s" podCreationTimestamp="2025-11-28 13:42:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 13:42:50.812161266 +0000 UTC m=+1327.619464610" watchObservedRunningTime="2025-11-28 13:42:50.818417972 +0000 UTC m=+1327.625721316" Nov 28 13:42:58 crc kubenswrapper[4631]: I1128 13:42:58.503495 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6cd9bffc9-x8stz" Nov 28 13:42:58 crc kubenswrapper[4631]: I1128 13:42:58.606325 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-79bd4cc8c9-4gxf7"] Nov 28 13:42:58 crc kubenswrapper[4631]: I1128 13:42:58.606620 4631 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-79bd4cc8c9-4gxf7" podUID="7179a7f7-1ee8-40a2-abcd-0cd373575b82" containerName="dnsmasq-dns" containerID="cri-o://8ad6f16d4e9ca7be24d22143e8977e1234b4a887dab4dbc3fc9568e901edc099" gracePeriod=10 Nov 28 13:42:58 crc kubenswrapper[4631]: I1128 13:42:58.880253 4631 generic.go:334] "Generic (PLEG): container finished" podID="7179a7f7-1ee8-40a2-abcd-0cd373575b82" containerID="8ad6f16d4e9ca7be24d22143e8977e1234b4a887dab4dbc3fc9568e901edc099" exitCode=0 Nov 28 13:42:58 crc kubenswrapper[4631]: I1128 13:42:58.880330 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-79bd4cc8c9-4gxf7" event={"ID":"7179a7f7-1ee8-40a2-abcd-0cd373575b82","Type":"ContainerDied","Data":"8ad6f16d4e9ca7be24d22143e8977e1234b4a887dab4dbc3fc9568e901edc099"} Nov 28 13:42:59 crc kubenswrapper[4631]: I1128 13:42:59.170062 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-79bd4cc8c9-4gxf7" Nov 28 13:42:59 crc kubenswrapper[4631]: I1128 13:42:59.254765 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9fk57\" (UniqueName: \"kubernetes.io/projected/7179a7f7-1ee8-40a2-abcd-0cd373575b82-kube-api-access-9fk57\") pod \"7179a7f7-1ee8-40a2-abcd-0cd373575b82\" (UID: \"7179a7f7-1ee8-40a2-abcd-0cd373575b82\") " Nov 28 13:42:59 crc kubenswrapper[4631]: I1128 13:42:59.255013 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7179a7f7-1ee8-40a2-abcd-0cd373575b82-ovsdbserver-sb\") pod \"7179a7f7-1ee8-40a2-abcd-0cd373575b82\" (UID: \"7179a7f7-1ee8-40a2-abcd-0cd373575b82\") " Nov 28 13:42:59 crc kubenswrapper[4631]: I1128 13:42:59.255466 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7179a7f7-1ee8-40a2-abcd-0cd373575b82-dns-svc\") pod \"7179a7f7-1ee8-40a2-abcd-0cd373575b82\" (UID: \"7179a7f7-1ee8-40a2-abcd-0cd373575b82\") " Nov 28 13:42:59 crc kubenswrapper[4631]: I1128 13:42:59.255540 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7179a7f7-1ee8-40a2-abcd-0cd373575b82-ovsdbserver-nb\") pod \"7179a7f7-1ee8-40a2-abcd-0cd373575b82\" (UID: \"7179a7f7-1ee8-40a2-abcd-0cd373575b82\") " Nov 28 13:42:59 crc kubenswrapper[4631]: I1128 13:42:59.255570 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7179a7f7-1ee8-40a2-abcd-0cd373575b82-dns-swift-storage-0\") pod \"7179a7f7-1ee8-40a2-abcd-0cd373575b82\" (UID: \"7179a7f7-1ee8-40a2-abcd-0cd373575b82\") " Nov 28 13:42:59 crc kubenswrapper[4631]: I1128 13:42:59.257096 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/7179a7f7-1ee8-40a2-abcd-0cd373575b82-openstack-edpm-ipam\") pod \"7179a7f7-1ee8-40a2-abcd-0cd373575b82\" (UID: \"7179a7f7-1ee8-40a2-abcd-0cd373575b82\") " Nov 28 13:42:59 crc kubenswrapper[4631]: I1128 13:42:59.257221 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7179a7f7-1ee8-40a2-abcd-0cd373575b82-config\") pod \"7179a7f7-1ee8-40a2-abcd-0cd373575b82\" (UID: \"7179a7f7-1ee8-40a2-abcd-0cd373575b82\") " Nov 28 13:42:59 crc kubenswrapper[4631]: I1128 13:42:59.283418 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7179a7f7-1ee8-40a2-abcd-0cd373575b82-kube-api-access-9fk57" (OuterVolumeSpecName: "kube-api-access-9fk57") pod "7179a7f7-1ee8-40a2-abcd-0cd373575b82" (UID: "7179a7f7-1ee8-40a2-abcd-0cd373575b82"). InnerVolumeSpecName "kube-api-access-9fk57". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:42:59 crc kubenswrapper[4631]: I1128 13:42:59.316393 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7179a7f7-1ee8-40a2-abcd-0cd373575b82-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "7179a7f7-1ee8-40a2-abcd-0cd373575b82" (UID: "7179a7f7-1ee8-40a2-abcd-0cd373575b82"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 13:42:59 crc kubenswrapper[4631]: I1128 13:42:59.319489 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7179a7f7-1ee8-40a2-abcd-0cd373575b82-openstack-edpm-ipam" (OuterVolumeSpecName: "openstack-edpm-ipam") pod "7179a7f7-1ee8-40a2-abcd-0cd373575b82" (UID: "7179a7f7-1ee8-40a2-abcd-0cd373575b82"). InnerVolumeSpecName "openstack-edpm-ipam". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 13:42:59 crc kubenswrapper[4631]: I1128 13:42:59.329909 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7179a7f7-1ee8-40a2-abcd-0cd373575b82-config" (OuterVolumeSpecName: "config") pod "7179a7f7-1ee8-40a2-abcd-0cd373575b82" (UID: "7179a7f7-1ee8-40a2-abcd-0cd373575b82"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 13:42:59 crc kubenswrapper[4631]: I1128 13:42:59.340222 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7179a7f7-1ee8-40a2-abcd-0cd373575b82-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "7179a7f7-1ee8-40a2-abcd-0cd373575b82" (UID: "7179a7f7-1ee8-40a2-abcd-0cd373575b82"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 13:42:59 crc kubenswrapper[4631]: I1128 13:42:59.347252 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7179a7f7-1ee8-40a2-abcd-0cd373575b82-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "7179a7f7-1ee8-40a2-abcd-0cd373575b82" (UID: "7179a7f7-1ee8-40a2-abcd-0cd373575b82"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 13:42:59 crc kubenswrapper[4631]: I1128 13:42:59.364885 4631 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/7179a7f7-1ee8-40a2-abcd-0cd373575b82-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Nov 28 13:42:59 crc kubenswrapper[4631]: I1128 13:42:59.365019 4631 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7179a7f7-1ee8-40a2-abcd-0cd373575b82-config\") on node \"crc\" DevicePath \"\"" Nov 28 13:42:59 crc kubenswrapper[4631]: I1128 13:42:59.365109 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9fk57\" (UniqueName: \"kubernetes.io/projected/7179a7f7-1ee8-40a2-abcd-0cd373575b82-kube-api-access-9fk57\") on node \"crc\" DevicePath \"\"" Nov 28 13:42:59 crc kubenswrapper[4631]: I1128 13:42:59.365205 4631 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7179a7f7-1ee8-40a2-abcd-0cd373575b82-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 28 13:42:59 crc kubenswrapper[4631]: I1128 13:42:59.365272 4631 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7179a7f7-1ee8-40a2-abcd-0cd373575b82-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 28 13:42:59 crc kubenswrapper[4631]: I1128 13:42:59.365349 4631 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7179a7f7-1ee8-40a2-abcd-0cd373575b82-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 28 13:42:59 crc kubenswrapper[4631]: I1128 13:42:59.394681 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7179a7f7-1ee8-40a2-abcd-0cd373575b82-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "7179a7f7-1ee8-40a2-abcd-0cd373575b82" (UID: "7179a7f7-1ee8-40a2-abcd-0cd373575b82"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 13:42:59 crc kubenswrapper[4631]: I1128 13:42:59.468076 4631 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7179a7f7-1ee8-40a2-abcd-0cd373575b82-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 28 13:42:59 crc kubenswrapper[4631]: I1128 13:42:59.893671 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-79bd4cc8c9-4gxf7" event={"ID":"7179a7f7-1ee8-40a2-abcd-0cd373575b82","Type":"ContainerDied","Data":"47f054a60628e02b8a4e2d376b8c43009ccc9ff6d84538cdb68450a089c966e9"} Nov 28 13:42:59 crc kubenswrapper[4631]: I1128 13:42:59.893735 4631 scope.go:117] "RemoveContainer" containerID="8ad6f16d4e9ca7be24d22143e8977e1234b4a887dab4dbc3fc9568e901edc099" Nov 28 13:42:59 crc kubenswrapper[4631]: I1128 13:42:59.893922 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-79bd4cc8c9-4gxf7" Nov 28 13:42:59 crc kubenswrapper[4631]: I1128 13:42:59.923744 4631 scope.go:117] "RemoveContainer" containerID="28d9cc47a5644ecb10afdbc2123a96f22c954216480070dffda06fa445deae2e" Nov 28 13:42:59 crc kubenswrapper[4631]: I1128 13:42:59.948478 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-79bd4cc8c9-4gxf7"] Nov 28 13:42:59 crc kubenswrapper[4631]: I1128 13:42:59.961770 4631 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-79bd4cc8c9-4gxf7"] Nov 28 13:43:01 crc kubenswrapper[4631]: I1128 13:43:01.526450 4631 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7179a7f7-1ee8-40a2-abcd-0cd373575b82" path="/var/lib/kubelet/pods/7179a7f7-1ee8-40a2-abcd-0cd373575b82/volumes" Nov 28 13:43:02 crc kubenswrapper[4631]: I1128 13:43:02.474584 4631 scope.go:117] "RemoveContainer" containerID="12413999ef935ad5f2fe03a148a79c915c22824ce8578cc03526415f1f365612" Nov 28 13:43:11 crc kubenswrapper[4631]: I1128 13:43:11.008697 4631 generic.go:334] "Generic (PLEG): container finished" podID="fc5706c2-8b7d-439a-95b3-981907360c92" containerID="400faf511747c4aecb8b263484297c384f64c4393d9b88fbc03ba79c021a9bee" exitCode=0 Nov 28 13:43:11 crc kubenswrapper[4631]: I1128 13:43:11.008787 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"fc5706c2-8b7d-439a-95b3-981907360c92","Type":"ContainerDied","Data":"400faf511747c4aecb8b263484297c384f64c4393d9b88fbc03ba79c021a9bee"} Nov 28 13:43:11 crc kubenswrapper[4631]: I1128 13:43:11.013888 4631 generic.go:334] "Generic (PLEG): container finished" podID="014519bc-68f0-4132-9a46-6b0e5c538c7e" containerID="0ced32c9dda9f55eceb6c4e60abf6aa57be7ac93b6c4ef8ecac6bf27f0f3fbda" exitCode=0 Nov 28 13:43:11 crc kubenswrapper[4631]: I1128 13:43:11.013938 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"014519bc-68f0-4132-9a46-6b0e5c538c7e","Type":"ContainerDied","Data":"0ced32c9dda9f55eceb6c4e60abf6aa57be7ac93b6c4ef8ecac6bf27f0f3fbda"} Nov 28 13:43:12 crc kubenswrapper[4631]: I1128 13:43:12.035438 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"fc5706c2-8b7d-439a-95b3-981907360c92","Type":"ContainerStarted","Data":"1476f9a6c6ed6c8dadce8fa1304a31a2ee47421b2e6c1023e13677bdb9fde2df"} Nov 28 13:43:12 crc kubenswrapper[4631]: I1128 13:43:12.038156 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Nov 28 13:43:12 crc kubenswrapper[4631]: I1128 13:43:12.040436 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"014519bc-68f0-4132-9a46-6b0e5c538c7e","Type":"ContainerStarted","Data":"86c7be075f00b520ceeab21a57753af161b802972f7e2b8a41450352e2d3f713"} Nov 28 13:43:12 crc kubenswrapper[4631]: I1128 13:43:12.040878 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Nov 28 13:43:12 crc kubenswrapper[4631]: I1128 13:43:12.079460 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=37.079432952 podStartE2EDuration="37.079432952s" podCreationTimestamp="2025-11-28 13:42:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 13:43:12.075714046 +0000 UTC m=+1348.883017400" watchObservedRunningTime="2025-11-28 13:43:12.079432952 +0000 UTC m=+1348.886736296" Nov 28 13:43:12 crc kubenswrapper[4631]: I1128 13:43:12.119479 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=37.119454032 podStartE2EDuration="37.119454032s" podCreationTimestamp="2025-11-28 13:42:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 13:43:12.116814362 +0000 UTC m=+1348.924117716" watchObservedRunningTime="2025-11-28 13:43:12.119454032 +0000 UTC m=+1348.926757376" Nov 28 13:43:16 crc kubenswrapper[4631]: I1128 13:43:16.130175 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-slcpp"] Nov 28 13:43:16 crc kubenswrapper[4631]: E1128 13:43:16.131054 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="931de4b9-12a1-45bf-b69e-b0498c6c770f" containerName="init" Nov 28 13:43:16 crc kubenswrapper[4631]: I1128 13:43:16.131068 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="931de4b9-12a1-45bf-b69e-b0498c6c770f" containerName="init" Nov 28 13:43:16 crc kubenswrapper[4631]: E1128 13:43:16.131091 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7179a7f7-1ee8-40a2-abcd-0cd373575b82" containerName="dnsmasq-dns" Nov 28 13:43:16 crc kubenswrapper[4631]: I1128 13:43:16.131096 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="7179a7f7-1ee8-40a2-abcd-0cd373575b82" containerName="dnsmasq-dns" Nov 28 13:43:16 crc kubenswrapper[4631]: E1128 13:43:16.131112 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="931de4b9-12a1-45bf-b69e-b0498c6c770f" containerName="dnsmasq-dns" Nov 28 13:43:16 crc kubenswrapper[4631]: I1128 13:43:16.131118 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="931de4b9-12a1-45bf-b69e-b0498c6c770f" containerName="dnsmasq-dns" Nov 28 13:43:16 crc kubenswrapper[4631]: E1128 13:43:16.131133 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7179a7f7-1ee8-40a2-abcd-0cd373575b82" containerName="init" Nov 28 13:43:16 crc kubenswrapper[4631]: I1128 13:43:16.131139 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="7179a7f7-1ee8-40a2-abcd-0cd373575b82" containerName="init" Nov 28 13:43:16 crc kubenswrapper[4631]: I1128 13:43:16.131379 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="931de4b9-12a1-45bf-b69e-b0498c6c770f" containerName="dnsmasq-dns" Nov 28 13:43:16 crc kubenswrapper[4631]: I1128 13:43:16.131390 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="7179a7f7-1ee8-40a2-abcd-0cd373575b82" containerName="dnsmasq-dns" Nov 28 13:43:16 crc kubenswrapper[4631]: I1128 13:43:16.132126 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-slcpp" Nov 28 13:43:16 crc kubenswrapper[4631]: I1128 13:43:16.135395 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-vp4ns" Nov 28 13:43:16 crc kubenswrapper[4631]: I1128 13:43:16.135639 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 28 13:43:16 crc kubenswrapper[4631]: I1128 13:43:16.138620 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 28 13:43:16 crc kubenswrapper[4631]: I1128 13:43:16.142879 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 28 13:43:16 crc kubenswrapper[4631]: I1128 13:43:16.147700 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-slcpp"] Nov 28 13:43:16 crc kubenswrapper[4631]: I1128 13:43:16.250702 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0527a324-2259-4aac-a64b-33747a2e24ec-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-slcpp\" (UID: \"0527a324-2259-4aac-a64b-33747a2e24ec\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-slcpp" Nov 28 13:43:16 crc kubenswrapper[4631]: I1128 13:43:16.251122 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0527a324-2259-4aac-a64b-33747a2e24ec-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-slcpp\" (UID: \"0527a324-2259-4aac-a64b-33747a2e24ec\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-slcpp" Nov 28 13:43:16 crc kubenswrapper[4631]: I1128 13:43:16.251416 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kfnn5\" (UniqueName: \"kubernetes.io/projected/0527a324-2259-4aac-a64b-33747a2e24ec-kube-api-access-kfnn5\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-slcpp\" (UID: \"0527a324-2259-4aac-a64b-33747a2e24ec\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-slcpp" Nov 28 13:43:16 crc kubenswrapper[4631]: I1128 13:43:16.251487 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0527a324-2259-4aac-a64b-33747a2e24ec-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-slcpp\" (UID: \"0527a324-2259-4aac-a64b-33747a2e24ec\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-slcpp" Nov 28 13:43:16 crc kubenswrapper[4631]: I1128 13:43:16.353923 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0527a324-2259-4aac-a64b-33747a2e24ec-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-slcpp\" (UID: \"0527a324-2259-4aac-a64b-33747a2e24ec\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-slcpp" Nov 28 13:43:16 crc kubenswrapper[4631]: I1128 13:43:16.354022 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kfnn5\" (UniqueName: \"kubernetes.io/projected/0527a324-2259-4aac-a64b-33747a2e24ec-kube-api-access-kfnn5\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-slcpp\" (UID: \"0527a324-2259-4aac-a64b-33747a2e24ec\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-slcpp" Nov 28 13:43:16 crc kubenswrapper[4631]: I1128 13:43:16.354054 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0527a324-2259-4aac-a64b-33747a2e24ec-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-slcpp\" (UID: \"0527a324-2259-4aac-a64b-33747a2e24ec\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-slcpp" Nov 28 13:43:16 crc kubenswrapper[4631]: I1128 13:43:16.354119 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0527a324-2259-4aac-a64b-33747a2e24ec-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-slcpp\" (UID: \"0527a324-2259-4aac-a64b-33747a2e24ec\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-slcpp" Nov 28 13:43:16 crc kubenswrapper[4631]: I1128 13:43:16.361685 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0527a324-2259-4aac-a64b-33747a2e24ec-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-slcpp\" (UID: \"0527a324-2259-4aac-a64b-33747a2e24ec\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-slcpp" Nov 28 13:43:16 crc kubenswrapper[4631]: I1128 13:43:16.362500 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0527a324-2259-4aac-a64b-33747a2e24ec-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-slcpp\" (UID: \"0527a324-2259-4aac-a64b-33747a2e24ec\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-slcpp" Nov 28 13:43:16 crc kubenswrapper[4631]: I1128 13:43:16.363633 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0527a324-2259-4aac-a64b-33747a2e24ec-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-slcpp\" (UID: \"0527a324-2259-4aac-a64b-33747a2e24ec\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-slcpp" Nov 28 13:43:16 crc kubenswrapper[4631]: I1128 13:43:16.389626 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kfnn5\" (UniqueName: \"kubernetes.io/projected/0527a324-2259-4aac-a64b-33747a2e24ec-kube-api-access-kfnn5\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-slcpp\" (UID: \"0527a324-2259-4aac-a64b-33747a2e24ec\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-slcpp" Nov 28 13:43:16 crc kubenswrapper[4631]: I1128 13:43:16.457647 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-slcpp" Nov 28 13:43:17 crc kubenswrapper[4631]: I1128 13:43:17.136660 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-slcpp"] Nov 28 13:43:18 crc kubenswrapper[4631]: I1128 13:43:18.122937 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-slcpp" event={"ID":"0527a324-2259-4aac-a64b-33747a2e24ec","Type":"ContainerStarted","Data":"f8ed12daaa9dcdade37d54b395c859302dcfef74407b97a2441240e2c2fe6699"} Nov 28 13:43:19 crc kubenswrapper[4631]: I1128 13:43:19.634978 4631 patch_prober.go:28] interesting pod/machine-config-daemon-47sxw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 28 13:43:19 crc kubenswrapper[4631]: I1128 13:43:19.635344 4631 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" podUID="8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 28 13:43:25 crc kubenswrapper[4631]: I1128 13:43:25.918100 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Nov 28 13:43:25 crc kubenswrapper[4631]: I1128 13:43:25.989521 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Nov 28 13:43:30 crc kubenswrapper[4631]: I1128 13:43:30.286047 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-slcpp" event={"ID":"0527a324-2259-4aac-a64b-33747a2e24ec","Type":"ContainerStarted","Data":"fb00b7600ea6b4b236bcef7317cd96616e01f25d576600c2c5cdf1adabe90db4"} Nov 28 13:43:30 crc kubenswrapper[4631]: I1128 13:43:30.324103 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-slcpp" podStartSLOduration=2.135347802 podStartE2EDuration="14.324049001s" podCreationTimestamp="2025-11-28 13:43:16 +0000 UTC" firstStartedPulling="2025-11-28 13:43:17.145787434 +0000 UTC m=+1353.953090778" lastFinishedPulling="2025-11-28 13:43:29.334488633 +0000 UTC m=+1366.141791977" observedRunningTime="2025-11-28 13:43:30.309005356 +0000 UTC m=+1367.116308720" watchObservedRunningTime="2025-11-28 13:43:30.324049001 +0000 UTC m=+1367.131352345" Nov 28 13:43:42 crc kubenswrapper[4631]: I1128 13:43:42.451998 4631 generic.go:334] "Generic (PLEG): container finished" podID="0527a324-2259-4aac-a64b-33747a2e24ec" containerID="fb00b7600ea6b4b236bcef7317cd96616e01f25d576600c2c5cdf1adabe90db4" exitCode=0 Nov 28 13:43:42 crc kubenswrapper[4631]: I1128 13:43:42.452166 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-slcpp" event={"ID":"0527a324-2259-4aac-a64b-33747a2e24ec","Type":"ContainerDied","Data":"fb00b7600ea6b4b236bcef7317cd96616e01f25d576600c2c5cdf1adabe90db4"} Nov 28 13:43:43 crc kubenswrapper[4631]: I1128 13:43:43.993306 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-slcpp" Nov 28 13:43:44 crc kubenswrapper[4631]: I1128 13:43:44.067874 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfnn5\" (UniqueName: \"kubernetes.io/projected/0527a324-2259-4aac-a64b-33747a2e24ec-kube-api-access-kfnn5\") pod \"0527a324-2259-4aac-a64b-33747a2e24ec\" (UID: \"0527a324-2259-4aac-a64b-33747a2e24ec\") " Nov 28 13:43:44 crc kubenswrapper[4631]: I1128 13:43:44.068009 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0527a324-2259-4aac-a64b-33747a2e24ec-inventory\") pod \"0527a324-2259-4aac-a64b-33747a2e24ec\" (UID: \"0527a324-2259-4aac-a64b-33747a2e24ec\") " Nov 28 13:43:44 crc kubenswrapper[4631]: I1128 13:43:44.068092 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0527a324-2259-4aac-a64b-33747a2e24ec-ssh-key\") pod \"0527a324-2259-4aac-a64b-33747a2e24ec\" (UID: \"0527a324-2259-4aac-a64b-33747a2e24ec\") " Nov 28 13:43:44 crc kubenswrapper[4631]: I1128 13:43:44.068116 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0527a324-2259-4aac-a64b-33747a2e24ec-repo-setup-combined-ca-bundle\") pod \"0527a324-2259-4aac-a64b-33747a2e24ec\" (UID: \"0527a324-2259-4aac-a64b-33747a2e24ec\") " Nov 28 13:43:44 crc kubenswrapper[4631]: I1128 13:43:44.085708 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0527a324-2259-4aac-a64b-33747a2e24ec-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "0527a324-2259-4aac-a64b-33747a2e24ec" (UID: "0527a324-2259-4aac-a64b-33747a2e24ec"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:43:44 crc kubenswrapper[4631]: I1128 13:43:44.087865 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0527a324-2259-4aac-a64b-33747a2e24ec-kube-api-access-kfnn5" (OuterVolumeSpecName: "kube-api-access-kfnn5") pod "0527a324-2259-4aac-a64b-33747a2e24ec" (UID: "0527a324-2259-4aac-a64b-33747a2e24ec"). InnerVolumeSpecName "kube-api-access-kfnn5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:43:44 crc kubenswrapper[4631]: I1128 13:43:44.112710 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0527a324-2259-4aac-a64b-33747a2e24ec-inventory" (OuterVolumeSpecName: "inventory") pod "0527a324-2259-4aac-a64b-33747a2e24ec" (UID: "0527a324-2259-4aac-a64b-33747a2e24ec"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:43:44 crc kubenswrapper[4631]: I1128 13:43:44.113098 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0527a324-2259-4aac-a64b-33747a2e24ec-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "0527a324-2259-4aac-a64b-33747a2e24ec" (UID: "0527a324-2259-4aac-a64b-33747a2e24ec"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:43:44 crc kubenswrapper[4631]: I1128 13:43:44.169992 4631 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0527a324-2259-4aac-a64b-33747a2e24ec-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 28 13:43:44 crc kubenswrapper[4631]: I1128 13:43:44.170093 4631 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0527a324-2259-4aac-a64b-33747a2e24ec-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 13:43:44 crc kubenswrapper[4631]: I1128 13:43:44.170113 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfnn5\" (UniqueName: \"kubernetes.io/projected/0527a324-2259-4aac-a64b-33747a2e24ec-kube-api-access-kfnn5\") on node \"crc\" DevicePath \"\"" Nov 28 13:43:44 crc kubenswrapper[4631]: I1128 13:43:44.170128 4631 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0527a324-2259-4aac-a64b-33747a2e24ec-inventory\") on node \"crc\" DevicePath \"\"" Nov 28 13:43:44 crc kubenswrapper[4631]: I1128 13:43:44.487511 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-slcpp" event={"ID":"0527a324-2259-4aac-a64b-33747a2e24ec","Type":"ContainerDied","Data":"f8ed12daaa9dcdade37d54b395c859302dcfef74407b97a2441240e2c2fe6699"} Nov 28 13:43:44 crc kubenswrapper[4631]: I1128 13:43:44.487560 4631 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f8ed12daaa9dcdade37d54b395c859302dcfef74407b97a2441240e2c2fe6699" Nov 28 13:43:44 crc kubenswrapper[4631]: I1128 13:43:44.487605 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-slcpp" Nov 28 13:43:44 crc kubenswrapper[4631]: I1128 13:43:44.621042 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-s9hz7"] Nov 28 13:43:44 crc kubenswrapper[4631]: E1128 13:43:44.623106 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0527a324-2259-4aac-a64b-33747a2e24ec" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Nov 28 13:43:44 crc kubenswrapper[4631]: I1128 13:43:44.623223 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="0527a324-2259-4aac-a64b-33747a2e24ec" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Nov 28 13:43:44 crc kubenswrapper[4631]: I1128 13:43:44.623880 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="0527a324-2259-4aac-a64b-33747a2e24ec" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Nov 28 13:43:44 crc kubenswrapper[4631]: I1128 13:43:44.625417 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-s9hz7" Nov 28 13:43:44 crc kubenswrapper[4631]: I1128 13:43:44.639576 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-s9hz7"] Nov 28 13:43:44 crc kubenswrapper[4631]: I1128 13:43:44.679753 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 28 13:43:44 crc kubenswrapper[4631]: I1128 13:43:44.681639 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 28 13:43:44 crc kubenswrapper[4631]: I1128 13:43:44.682962 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-vp4ns" Nov 28 13:43:44 crc kubenswrapper[4631]: I1128 13:43:44.683118 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 28 13:43:44 crc kubenswrapper[4631]: I1128 13:43:44.700837 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xm9ls\" (UniqueName: \"kubernetes.io/projected/2b15b142-3dc9-4d33-96e7-5ba3e357ee59-kube-api-access-xm9ls\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-s9hz7\" (UID: \"2b15b142-3dc9-4d33-96e7-5ba3e357ee59\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-s9hz7" Nov 28 13:43:44 crc kubenswrapper[4631]: I1128 13:43:44.700906 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2b15b142-3dc9-4d33-96e7-5ba3e357ee59-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-s9hz7\" (UID: \"2b15b142-3dc9-4d33-96e7-5ba3e357ee59\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-s9hz7" Nov 28 13:43:44 crc kubenswrapper[4631]: I1128 13:43:44.700983 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2b15b142-3dc9-4d33-96e7-5ba3e357ee59-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-s9hz7\" (UID: \"2b15b142-3dc9-4d33-96e7-5ba3e357ee59\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-s9hz7" Nov 28 13:43:44 crc kubenswrapper[4631]: I1128 13:43:44.804491 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xm9ls\" (UniqueName: \"kubernetes.io/projected/2b15b142-3dc9-4d33-96e7-5ba3e357ee59-kube-api-access-xm9ls\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-s9hz7\" (UID: \"2b15b142-3dc9-4d33-96e7-5ba3e357ee59\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-s9hz7" Nov 28 13:43:44 crc kubenswrapper[4631]: I1128 13:43:44.804622 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2b15b142-3dc9-4d33-96e7-5ba3e357ee59-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-s9hz7\" (UID: \"2b15b142-3dc9-4d33-96e7-5ba3e357ee59\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-s9hz7" Nov 28 13:43:44 crc kubenswrapper[4631]: I1128 13:43:44.804733 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2b15b142-3dc9-4d33-96e7-5ba3e357ee59-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-s9hz7\" (UID: \"2b15b142-3dc9-4d33-96e7-5ba3e357ee59\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-s9hz7" Nov 28 13:43:44 crc kubenswrapper[4631]: I1128 13:43:44.812907 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2b15b142-3dc9-4d33-96e7-5ba3e357ee59-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-s9hz7\" (UID: \"2b15b142-3dc9-4d33-96e7-5ba3e357ee59\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-s9hz7" Nov 28 13:43:44 crc kubenswrapper[4631]: I1128 13:43:44.824073 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2b15b142-3dc9-4d33-96e7-5ba3e357ee59-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-s9hz7\" (UID: \"2b15b142-3dc9-4d33-96e7-5ba3e357ee59\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-s9hz7" Nov 28 13:43:44 crc kubenswrapper[4631]: I1128 13:43:44.829255 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xm9ls\" (UniqueName: \"kubernetes.io/projected/2b15b142-3dc9-4d33-96e7-5ba3e357ee59-kube-api-access-xm9ls\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-s9hz7\" (UID: \"2b15b142-3dc9-4d33-96e7-5ba3e357ee59\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-s9hz7" Nov 28 13:43:45 crc kubenswrapper[4631]: I1128 13:43:45.012479 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-s9hz7" Nov 28 13:43:45 crc kubenswrapper[4631]: I1128 13:43:45.597164 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-s9hz7"] Nov 28 13:43:46 crc kubenswrapper[4631]: I1128 13:43:46.508739 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-s9hz7" event={"ID":"2b15b142-3dc9-4d33-96e7-5ba3e357ee59","Type":"ContainerStarted","Data":"d64df0007bb42e2ba9c84b6656fbaf60c3bbd5bc2fb250da71d5a59627dd4412"} Nov 28 13:43:47 crc kubenswrapper[4631]: I1128 13:43:47.535243 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-s9hz7" event={"ID":"2b15b142-3dc9-4d33-96e7-5ba3e357ee59","Type":"ContainerStarted","Data":"451c535439eb4142cf8fcc9ed3a6dc7bfa7d1230f6f66a305197d031b181faa4"} Nov 28 13:43:47 crc kubenswrapper[4631]: I1128 13:43:47.570610 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-s9hz7" podStartSLOduration=2.78354932 podStartE2EDuration="3.570584242s" podCreationTimestamp="2025-11-28 13:43:44 +0000 UTC" firstStartedPulling="2025-11-28 13:43:45.636957241 +0000 UTC m=+1382.444260585" lastFinishedPulling="2025-11-28 13:43:46.423992163 +0000 UTC m=+1383.231295507" observedRunningTime="2025-11-28 13:43:47.565112923 +0000 UTC m=+1384.372416267" watchObservedRunningTime="2025-11-28 13:43:47.570584242 +0000 UTC m=+1384.377887586" Nov 28 13:43:49 crc kubenswrapper[4631]: I1128 13:43:49.562645 4631 generic.go:334] "Generic (PLEG): container finished" podID="2b15b142-3dc9-4d33-96e7-5ba3e357ee59" containerID="451c535439eb4142cf8fcc9ed3a6dc7bfa7d1230f6f66a305197d031b181faa4" exitCode=0 Nov 28 13:43:49 crc kubenswrapper[4631]: I1128 13:43:49.562913 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-s9hz7" event={"ID":"2b15b142-3dc9-4d33-96e7-5ba3e357ee59","Type":"ContainerDied","Data":"451c535439eb4142cf8fcc9ed3a6dc7bfa7d1230f6f66a305197d031b181faa4"} Nov 28 13:43:49 crc kubenswrapper[4631]: I1128 13:43:49.635048 4631 patch_prober.go:28] interesting pod/machine-config-daemon-47sxw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 28 13:43:49 crc kubenswrapper[4631]: I1128 13:43:49.635120 4631 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" podUID="8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 28 13:43:51 crc kubenswrapper[4631]: I1128 13:43:51.013709 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-s9hz7" Nov 28 13:43:51 crc kubenswrapper[4631]: I1128 13:43:51.071492 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xm9ls\" (UniqueName: \"kubernetes.io/projected/2b15b142-3dc9-4d33-96e7-5ba3e357ee59-kube-api-access-xm9ls\") pod \"2b15b142-3dc9-4d33-96e7-5ba3e357ee59\" (UID: \"2b15b142-3dc9-4d33-96e7-5ba3e357ee59\") " Nov 28 13:43:51 crc kubenswrapper[4631]: I1128 13:43:51.071533 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2b15b142-3dc9-4d33-96e7-5ba3e357ee59-inventory\") pod \"2b15b142-3dc9-4d33-96e7-5ba3e357ee59\" (UID: \"2b15b142-3dc9-4d33-96e7-5ba3e357ee59\") " Nov 28 13:43:51 crc kubenswrapper[4631]: I1128 13:43:51.071623 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2b15b142-3dc9-4d33-96e7-5ba3e357ee59-ssh-key\") pod \"2b15b142-3dc9-4d33-96e7-5ba3e357ee59\" (UID: \"2b15b142-3dc9-4d33-96e7-5ba3e357ee59\") " Nov 28 13:43:51 crc kubenswrapper[4631]: I1128 13:43:51.089892 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2b15b142-3dc9-4d33-96e7-5ba3e357ee59-kube-api-access-xm9ls" (OuterVolumeSpecName: "kube-api-access-xm9ls") pod "2b15b142-3dc9-4d33-96e7-5ba3e357ee59" (UID: "2b15b142-3dc9-4d33-96e7-5ba3e357ee59"). InnerVolumeSpecName "kube-api-access-xm9ls". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:43:51 crc kubenswrapper[4631]: I1128 13:43:51.102274 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2b15b142-3dc9-4d33-96e7-5ba3e357ee59-inventory" (OuterVolumeSpecName: "inventory") pod "2b15b142-3dc9-4d33-96e7-5ba3e357ee59" (UID: "2b15b142-3dc9-4d33-96e7-5ba3e357ee59"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:43:51 crc kubenswrapper[4631]: I1128 13:43:51.105052 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2b15b142-3dc9-4d33-96e7-5ba3e357ee59-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "2b15b142-3dc9-4d33-96e7-5ba3e357ee59" (UID: "2b15b142-3dc9-4d33-96e7-5ba3e357ee59"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:43:51 crc kubenswrapper[4631]: I1128 13:43:51.173801 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xm9ls\" (UniqueName: \"kubernetes.io/projected/2b15b142-3dc9-4d33-96e7-5ba3e357ee59-kube-api-access-xm9ls\") on node \"crc\" DevicePath \"\"" Nov 28 13:43:51 crc kubenswrapper[4631]: I1128 13:43:51.173860 4631 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2b15b142-3dc9-4d33-96e7-5ba3e357ee59-inventory\") on node \"crc\" DevicePath \"\"" Nov 28 13:43:51 crc kubenswrapper[4631]: I1128 13:43:51.173869 4631 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2b15b142-3dc9-4d33-96e7-5ba3e357ee59-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 28 13:43:51 crc kubenswrapper[4631]: I1128 13:43:51.589602 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-s9hz7" event={"ID":"2b15b142-3dc9-4d33-96e7-5ba3e357ee59","Type":"ContainerDied","Data":"d64df0007bb42e2ba9c84b6656fbaf60c3bbd5bc2fb250da71d5a59627dd4412"} Nov 28 13:43:51 crc kubenswrapper[4631]: I1128 13:43:51.589960 4631 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d64df0007bb42e2ba9c84b6656fbaf60c3bbd5bc2fb250da71d5a59627dd4412" Nov 28 13:43:51 crc kubenswrapper[4631]: I1128 13:43:51.589659 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-s9hz7" Nov 28 13:43:51 crc kubenswrapper[4631]: I1128 13:43:51.672965 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-xz7jm"] Nov 28 13:43:51 crc kubenswrapper[4631]: E1128 13:43:51.673543 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2b15b142-3dc9-4d33-96e7-5ba3e357ee59" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Nov 28 13:43:51 crc kubenswrapper[4631]: I1128 13:43:51.673565 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="2b15b142-3dc9-4d33-96e7-5ba3e357ee59" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Nov 28 13:43:51 crc kubenswrapper[4631]: I1128 13:43:51.673811 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="2b15b142-3dc9-4d33-96e7-5ba3e357ee59" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Nov 28 13:43:51 crc kubenswrapper[4631]: I1128 13:43:51.674493 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-xz7jm" Nov 28 13:43:51 crc kubenswrapper[4631]: I1128 13:43:51.679709 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 28 13:43:51 crc kubenswrapper[4631]: I1128 13:43:51.680189 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 28 13:43:51 crc kubenswrapper[4631]: I1128 13:43:51.680798 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-vp4ns" Nov 28 13:43:51 crc kubenswrapper[4631]: I1128 13:43:51.681099 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 28 13:43:51 crc kubenswrapper[4631]: I1128 13:43:51.693538 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-xz7jm"] Nov 28 13:43:51 crc kubenswrapper[4631]: I1128 13:43:51.787816 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/44ae011d-6420-4400-9c7c-21463e3c82b1-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-xz7jm\" (UID: \"44ae011d-6420-4400-9c7c-21463e3c82b1\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-xz7jm" Nov 28 13:43:51 crc kubenswrapper[4631]: I1128 13:43:51.787920 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n5l9z\" (UniqueName: \"kubernetes.io/projected/44ae011d-6420-4400-9c7c-21463e3c82b1-kube-api-access-n5l9z\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-xz7jm\" (UID: \"44ae011d-6420-4400-9c7c-21463e3c82b1\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-xz7jm" Nov 28 13:43:51 crc kubenswrapper[4631]: I1128 13:43:51.788022 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/44ae011d-6420-4400-9c7c-21463e3c82b1-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-xz7jm\" (UID: \"44ae011d-6420-4400-9c7c-21463e3c82b1\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-xz7jm" Nov 28 13:43:51 crc kubenswrapper[4631]: I1128 13:43:51.788125 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/44ae011d-6420-4400-9c7c-21463e3c82b1-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-xz7jm\" (UID: \"44ae011d-6420-4400-9c7c-21463e3c82b1\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-xz7jm" Nov 28 13:43:51 crc kubenswrapper[4631]: I1128 13:43:51.890811 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/44ae011d-6420-4400-9c7c-21463e3c82b1-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-xz7jm\" (UID: \"44ae011d-6420-4400-9c7c-21463e3c82b1\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-xz7jm" Nov 28 13:43:51 crc kubenswrapper[4631]: I1128 13:43:51.890872 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/44ae011d-6420-4400-9c7c-21463e3c82b1-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-xz7jm\" (UID: \"44ae011d-6420-4400-9c7c-21463e3c82b1\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-xz7jm" Nov 28 13:43:51 crc kubenswrapper[4631]: I1128 13:43:51.890926 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n5l9z\" (UniqueName: \"kubernetes.io/projected/44ae011d-6420-4400-9c7c-21463e3c82b1-kube-api-access-n5l9z\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-xz7jm\" (UID: \"44ae011d-6420-4400-9c7c-21463e3c82b1\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-xz7jm" Nov 28 13:43:51 crc kubenswrapper[4631]: I1128 13:43:51.891025 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/44ae011d-6420-4400-9c7c-21463e3c82b1-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-xz7jm\" (UID: \"44ae011d-6420-4400-9c7c-21463e3c82b1\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-xz7jm" Nov 28 13:43:51 crc kubenswrapper[4631]: I1128 13:43:51.896200 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/44ae011d-6420-4400-9c7c-21463e3c82b1-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-xz7jm\" (UID: \"44ae011d-6420-4400-9c7c-21463e3c82b1\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-xz7jm" Nov 28 13:43:51 crc kubenswrapper[4631]: I1128 13:43:51.901630 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/44ae011d-6420-4400-9c7c-21463e3c82b1-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-xz7jm\" (UID: \"44ae011d-6420-4400-9c7c-21463e3c82b1\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-xz7jm" Nov 28 13:43:51 crc kubenswrapper[4631]: I1128 13:43:51.910133 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/44ae011d-6420-4400-9c7c-21463e3c82b1-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-xz7jm\" (UID: \"44ae011d-6420-4400-9c7c-21463e3c82b1\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-xz7jm" Nov 28 13:43:51 crc kubenswrapper[4631]: I1128 13:43:51.913724 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n5l9z\" (UniqueName: \"kubernetes.io/projected/44ae011d-6420-4400-9c7c-21463e3c82b1-kube-api-access-n5l9z\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-xz7jm\" (UID: \"44ae011d-6420-4400-9c7c-21463e3c82b1\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-xz7jm" Nov 28 13:43:51 crc kubenswrapper[4631]: I1128 13:43:51.992134 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-xz7jm" Nov 28 13:43:52 crc kubenswrapper[4631]: I1128 13:43:52.546016 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-xz7jm"] Nov 28 13:43:52 crc kubenswrapper[4631]: I1128 13:43:52.606787 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-xz7jm" event={"ID":"44ae011d-6420-4400-9c7c-21463e3c82b1","Type":"ContainerStarted","Data":"3bafea57cfda07b8698ecbc6b31bcd09cb79c7abc3c7216606d86e4d223c3053"} Nov 28 13:43:53 crc kubenswrapper[4631]: I1128 13:43:53.619491 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-xz7jm" event={"ID":"44ae011d-6420-4400-9c7c-21463e3c82b1","Type":"ContainerStarted","Data":"812c7c40987259fd36752aa82b116fec7562779d3e5df6f7f438155ff00d0820"} Nov 28 13:43:53 crc kubenswrapper[4631]: I1128 13:43:53.641220 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-xz7jm" podStartSLOduration=2.104462451 podStartE2EDuration="2.641196015s" podCreationTimestamp="2025-11-28 13:43:51 +0000 UTC" firstStartedPulling="2025-11-28 13:43:52.554471354 +0000 UTC m=+1389.361774698" lastFinishedPulling="2025-11-28 13:43:53.091204918 +0000 UTC m=+1389.898508262" observedRunningTime="2025-11-28 13:43:53.636870833 +0000 UTC m=+1390.444174197" watchObservedRunningTime="2025-11-28 13:43:53.641196015 +0000 UTC m=+1390.448499359" Nov 28 13:44:02 crc kubenswrapper[4631]: I1128 13:44:02.657605 4631 scope.go:117] "RemoveContainer" containerID="2011893ca0d1cfaff73961c600f7e3a4976a22a496820c12fdceadc1ef74ba32" Nov 28 13:44:19 crc kubenswrapper[4631]: I1128 13:44:19.634319 4631 patch_prober.go:28] interesting pod/machine-config-daemon-47sxw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 28 13:44:19 crc kubenswrapper[4631]: I1128 13:44:19.635162 4631 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" podUID="8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 28 13:44:19 crc kubenswrapper[4631]: I1128 13:44:19.635252 4631 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" Nov 28 13:44:19 crc kubenswrapper[4631]: I1128 13:44:19.636527 4631 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"05e6feabd7664ef6344f7bae4443e46fb2b0ab0814dcbdcb00623d5d82759c9c"} pod="openshift-machine-config-operator/machine-config-daemon-47sxw" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 28 13:44:19 crc kubenswrapper[4631]: I1128 13:44:19.636592 4631 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" podUID="8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1" containerName="machine-config-daemon" containerID="cri-o://05e6feabd7664ef6344f7bae4443e46fb2b0ab0814dcbdcb00623d5d82759c9c" gracePeriod=600 Nov 28 13:44:19 crc kubenswrapper[4631]: I1128 13:44:19.930278 4631 generic.go:334] "Generic (PLEG): container finished" podID="8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1" containerID="05e6feabd7664ef6344f7bae4443e46fb2b0ab0814dcbdcb00623d5d82759c9c" exitCode=0 Nov 28 13:44:19 crc kubenswrapper[4631]: I1128 13:44:19.930367 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" event={"ID":"8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1","Type":"ContainerDied","Data":"05e6feabd7664ef6344f7bae4443e46fb2b0ab0814dcbdcb00623d5d82759c9c"} Nov 28 13:44:19 crc kubenswrapper[4631]: I1128 13:44:19.930462 4631 scope.go:117] "RemoveContainer" containerID="ae892f616f5c89a5af75dda3b0799a9e1ba33e9ae4579727d53b425382bdbdc2" Nov 28 13:44:20 crc kubenswrapper[4631]: I1128 13:44:20.945250 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" event={"ID":"8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1","Type":"ContainerStarted","Data":"54d38280b4f1c123c61719ed71e612b2bf974f3bf3d8ff13fe19d70fa5ca5198"} Nov 28 13:44:40 crc kubenswrapper[4631]: I1128 13:44:40.753201 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-wg5bg"] Nov 28 13:44:40 crc kubenswrapper[4631]: I1128 13:44:40.759665 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wg5bg" Nov 28 13:44:40 crc kubenswrapper[4631]: I1128 13:44:40.775190 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c9nj7\" (UniqueName: \"kubernetes.io/projected/6c545b0d-6224-497d-b190-e9f619d340b2-kube-api-access-c9nj7\") pod \"redhat-operators-wg5bg\" (UID: \"6c545b0d-6224-497d-b190-e9f619d340b2\") " pod="openshift-marketplace/redhat-operators-wg5bg" Nov 28 13:44:40 crc kubenswrapper[4631]: I1128 13:44:40.776117 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6c545b0d-6224-497d-b190-e9f619d340b2-catalog-content\") pod \"redhat-operators-wg5bg\" (UID: \"6c545b0d-6224-497d-b190-e9f619d340b2\") " pod="openshift-marketplace/redhat-operators-wg5bg" Nov 28 13:44:40 crc kubenswrapper[4631]: I1128 13:44:40.776279 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6c545b0d-6224-497d-b190-e9f619d340b2-utilities\") pod \"redhat-operators-wg5bg\" (UID: \"6c545b0d-6224-497d-b190-e9f619d340b2\") " pod="openshift-marketplace/redhat-operators-wg5bg" Nov 28 13:44:40 crc kubenswrapper[4631]: I1128 13:44:40.790418 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-wg5bg"] Nov 28 13:44:40 crc kubenswrapper[4631]: I1128 13:44:40.879233 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c9nj7\" (UniqueName: \"kubernetes.io/projected/6c545b0d-6224-497d-b190-e9f619d340b2-kube-api-access-c9nj7\") pod \"redhat-operators-wg5bg\" (UID: \"6c545b0d-6224-497d-b190-e9f619d340b2\") " pod="openshift-marketplace/redhat-operators-wg5bg" Nov 28 13:44:40 crc kubenswrapper[4631]: I1128 13:44:40.879381 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6c545b0d-6224-497d-b190-e9f619d340b2-catalog-content\") pod \"redhat-operators-wg5bg\" (UID: \"6c545b0d-6224-497d-b190-e9f619d340b2\") " pod="openshift-marketplace/redhat-operators-wg5bg" Nov 28 13:44:40 crc kubenswrapper[4631]: I1128 13:44:40.879472 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6c545b0d-6224-497d-b190-e9f619d340b2-utilities\") pod \"redhat-operators-wg5bg\" (UID: \"6c545b0d-6224-497d-b190-e9f619d340b2\") " pod="openshift-marketplace/redhat-operators-wg5bg" Nov 28 13:44:40 crc kubenswrapper[4631]: I1128 13:44:40.880022 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6c545b0d-6224-497d-b190-e9f619d340b2-utilities\") pod \"redhat-operators-wg5bg\" (UID: \"6c545b0d-6224-497d-b190-e9f619d340b2\") " pod="openshift-marketplace/redhat-operators-wg5bg" Nov 28 13:44:40 crc kubenswrapper[4631]: I1128 13:44:40.880896 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6c545b0d-6224-497d-b190-e9f619d340b2-catalog-content\") pod \"redhat-operators-wg5bg\" (UID: \"6c545b0d-6224-497d-b190-e9f619d340b2\") " pod="openshift-marketplace/redhat-operators-wg5bg" Nov 28 13:44:40 crc kubenswrapper[4631]: I1128 13:44:40.906587 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c9nj7\" (UniqueName: \"kubernetes.io/projected/6c545b0d-6224-497d-b190-e9f619d340b2-kube-api-access-c9nj7\") pod \"redhat-operators-wg5bg\" (UID: \"6c545b0d-6224-497d-b190-e9f619d340b2\") " pod="openshift-marketplace/redhat-operators-wg5bg" Nov 28 13:44:41 crc kubenswrapper[4631]: I1128 13:44:41.106739 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wg5bg" Nov 28 13:44:41 crc kubenswrapper[4631]: I1128 13:44:41.638466 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-wg5bg"] Nov 28 13:44:42 crc kubenswrapper[4631]: I1128 13:44:42.225364 4631 generic.go:334] "Generic (PLEG): container finished" podID="6c545b0d-6224-497d-b190-e9f619d340b2" containerID="7c37ecdf0fbc5e39f49ae4ad205831cedc0d9100393516343642ace7813acbf7" exitCode=0 Nov 28 13:44:42 crc kubenswrapper[4631]: I1128 13:44:42.227042 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wg5bg" event={"ID":"6c545b0d-6224-497d-b190-e9f619d340b2","Type":"ContainerDied","Data":"7c37ecdf0fbc5e39f49ae4ad205831cedc0d9100393516343642ace7813acbf7"} Nov 28 13:44:42 crc kubenswrapper[4631]: I1128 13:44:42.227079 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wg5bg" event={"ID":"6c545b0d-6224-497d-b190-e9f619d340b2","Type":"ContainerStarted","Data":"71ceecc7f09b2810a809c93779bd44b7ed450a6c40810c2cfeaec52a91241606"} Nov 28 13:44:42 crc kubenswrapper[4631]: I1128 13:44:42.228485 4631 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 28 13:44:44 crc kubenswrapper[4631]: I1128 13:44:44.253880 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wg5bg" event={"ID":"6c545b0d-6224-497d-b190-e9f619d340b2","Type":"ContainerStarted","Data":"7e00b92e6e594984f05d2d7a6501956b32472077089937f74f54aa6f7f2e1f61"} Nov 28 13:44:48 crc kubenswrapper[4631]: I1128 13:44:48.341351 4631 generic.go:334] "Generic (PLEG): container finished" podID="6c545b0d-6224-497d-b190-e9f619d340b2" containerID="7e00b92e6e594984f05d2d7a6501956b32472077089937f74f54aa6f7f2e1f61" exitCode=0 Nov 28 13:44:48 crc kubenswrapper[4631]: I1128 13:44:48.341476 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wg5bg" event={"ID":"6c545b0d-6224-497d-b190-e9f619d340b2","Type":"ContainerDied","Data":"7e00b92e6e594984f05d2d7a6501956b32472077089937f74f54aa6f7f2e1f61"} Nov 28 13:44:49 crc kubenswrapper[4631]: I1128 13:44:49.354616 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wg5bg" event={"ID":"6c545b0d-6224-497d-b190-e9f619d340b2","Type":"ContainerStarted","Data":"a5cdb60a8c7b33ae92566eaffc9e17c69b1322b3fb5638789fe13d9bf0e7f786"} Nov 28 13:44:49 crc kubenswrapper[4631]: I1128 13:44:49.379461 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-wg5bg" podStartSLOduration=2.734929364 podStartE2EDuration="9.37944159s" podCreationTimestamp="2025-11-28 13:44:40 +0000 UTC" firstStartedPulling="2025-11-28 13:44:42.228252112 +0000 UTC m=+1439.035555456" lastFinishedPulling="2025-11-28 13:44:48.872764348 +0000 UTC m=+1445.680067682" observedRunningTime="2025-11-28 13:44:49.37477689 +0000 UTC m=+1446.182080234" watchObservedRunningTime="2025-11-28 13:44:49.37944159 +0000 UTC m=+1446.186744934" Nov 28 13:44:51 crc kubenswrapper[4631]: I1128 13:44:51.107077 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-wg5bg" Nov 28 13:44:51 crc kubenswrapper[4631]: I1128 13:44:51.107967 4631 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-wg5bg" Nov 28 13:44:52 crc kubenswrapper[4631]: I1128 13:44:52.152136 4631 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-wg5bg" podUID="6c545b0d-6224-497d-b190-e9f619d340b2" containerName="registry-server" probeResult="failure" output=< Nov 28 13:44:52 crc kubenswrapper[4631]: timeout: failed to connect service ":50051" within 1s Nov 28 13:44:52 crc kubenswrapper[4631]: > Nov 28 13:44:57 crc kubenswrapper[4631]: I1128 13:44:57.268126 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-c8xj5"] Nov 28 13:44:57 crc kubenswrapper[4631]: I1128 13:44:57.271708 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-c8xj5" Nov 28 13:44:57 crc kubenswrapper[4631]: I1128 13:44:57.280816 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-c8xj5"] Nov 28 13:44:57 crc kubenswrapper[4631]: I1128 13:44:57.388522 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lljw2\" (UniqueName: \"kubernetes.io/projected/0a640440-9c58-4899-967e-164e0f3ab927-kube-api-access-lljw2\") pod \"certified-operators-c8xj5\" (UID: \"0a640440-9c58-4899-967e-164e0f3ab927\") " pod="openshift-marketplace/certified-operators-c8xj5" Nov 28 13:44:57 crc kubenswrapper[4631]: I1128 13:44:57.388679 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0a640440-9c58-4899-967e-164e0f3ab927-catalog-content\") pod \"certified-operators-c8xj5\" (UID: \"0a640440-9c58-4899-967e-164e0f3ab927\") " pod="openshift-marketplace/certified-operators-c8xj5" Nov 28 13:44:57 crc kubenswrapper[4631]: I1128 13:44:57.388748 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0a640440-9c58-4899-967e-164e0f3ab927-utilities\") pod \"certified-operators-c8xj5\" (UID: \"0a640440-9c58-4899-967e-164e0f3ab927\") " pod="openshift-marketplace/certified-operators-c8xj5" Nov 28 13:44:57 crc kubenswrapper[4631]: I1128 13:44:57.491184 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lljw2\" (UniqueName: \"kubernetes.io/projected/0a640440-9c58-4899-967e-164e0f3ab927-kube-api-access-lljw2\") pod \"certified-operators-c8xj5\" (UID: \"0a640440-9c58-4899-967e-164e0f3ab927\") " pod="openshift-marketplace/certified-operators-c8xj5" Nov 28 13:44:57 crc kubenswrapper[4631]: I1128 13:44:57.491315 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0a640440-9c58-4899-967e-164e0f3ab927-catalog-content\") pod \"certified-operators-c8xj5\" (UID: \"0a640440-9c58-4899-967e-164e0f3ab927\") " pod="openshift-marketplace/certified-operators-c8xj5" Nov 28 13:44:57 crc kubenswrapper[4631]: I1128 13:44:57.491368 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0a640440-9c58-4899-967e-164e0f3ab927-utilities\") pod \"certified-operators-c8xj5\" (UID: \"0a640440-9c58-4899-967e-164e0f3ab927\") " pod="openshift-marketplace/certified-operators-c8xj5" Nov 28 13:44:57 crc kubenswrapper[4631]: I1128 13:44:57.491801 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0a640440-9c58-4899-967e-164e0f3ab927-catalog-content\") pod \"certified-operators-c8xj5\" (UID: \"0a640440-9c58-4899-967e-164e0f3ab927\") " pod="openshift-marketplace/certified-operators-c8xj5" Nov 28 13:44:57 crc kubenswrapper[4631]: I1128 13:44:57.491910 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0a640440-9c58-4899-967e-164e0f3ab927-utilities\") pod \"certified-operators-c8xj5\" (UID: \"0a640440-9c58-4899-967e-164e0f3ab927\") " pod="openshift-marketplace/certified-operators-c8xj5" Nov 28 13:44:57 crc kubenswrapper[4631]: I1128 13:44:57.520380 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lljw2\" (UniqueName: \"kubernetes.io/projected/0a640440-9c58-4899-967e-164e0f3ab927-kube-api-access-lljw2\") pod \"certified-operators-c8xj5\" (UID: \"0a640440-9c58-4899-967e-164e0f3ab927\") " pod="openshift-marketplace/certified-operators-c8xj5" Nov 28 13:44:57 crc kubenswrapper[4631]: I1128 13:44:57.593450 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-c8xj5" Nov 28 13:44:58 crc kubenswrapper[4631]: I1128 13:44:58.370751 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-c8xj5"] Nov 28 13:44:58 crc kubenswrapper[4631]: I1128 13:44:58.463261 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-c8xj5" event={"ID":"0a640440-9c58-4899-967e-164e0f3ab927","Type":"ContainerStarted","Data":"d35ea7fb9be621f9addf2c73d5143dad96f69a53cbae95f9448e6ef5dfc75cef"} Nov 28 13:44:59 crc kubenswrapper[4631]: I1128 13:44:59.477377 4631 generic.go:334] "Generic (PLEG): container finished" podID="0a640440-9c58-4899-967e-164e0f3ab927" containerID="f1a120bfde8f3fd81dd0f611449df3af3c75cfc45cfc20af575485a021fc7215" exitCode=0 Nov 28 13:44:59 crc kubenswrapper[4631]: I1128 13:44:59.477462 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-c8xj5" event={"ID":"0a640440-9c58-4899-967e-164e0f3ab927","Type":"ContainerDied","Data":"f1a120bfde8f3fd81dd0f611449df3af3c75cfc45cfc20af575485a021fc7215"} Nov 28 13:45:00 crc kubenswrapper[4631]: I1128 13:45:00.208340 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29405625-ncrbh"] Nov 28 13:45:00 crc kubenswrapper[4631]: I1128 13:45:00.210172 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29405625-ncrbh" Nov 28 13:45:00 crc kubenswrapper[4631]: I1128 13:45:00.214775 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 28 13:45:00 crc kubenswrapper[4631]: I1128 13:45:00.214968 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 28 13:45:00 crc kubenswrapper[4631]: I1128 13:45:00.253735 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29405625-ncrbh"] Nov 28 13:45:00 crc kubenswrapper[4631]: I1128 13:45:00.275117 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/9583083a-b161-492e-93c6-088fb16c06fe-secret-volume\") pod \"collect-profiles-29405625-ncrbh\" (UID: \"9583083a-b161-492e-93c6-088fb16c06fe\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405625-ncrbh" Nov 28 13:45:00 crc kubenswrapper[4631]: I1128 13:45:00.275367 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gg827\" (UniqueName: \"kubernetes.io/projected/9583083a-b161-492e-93c6-088fb16c06fe-kube-api-access-gg827\") pod \"collect-profiles-29405625-ncrbh\" (UID: \"9583083a-b161-492e-93c6-088fb16c06fe\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405625-ncrbh" Nov 28 13:45:00 crc kubenswrapper[4631]: I1128 13:45:00.275394 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9583083a-b161-492e-93c6-088fb16c06fe-config-volume\") pod \"collect-profiles-29405625-ncrbh\" (UID: \"9583083a-b161-492e-93c6-088fb16c06fe\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405625-ncrbh" Nov 28 13:45:00 crc kubenswrapper[4631]: I1128 13:45:00.378130 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gg827\" (UniqueName: \"kubernetes.io/projected/9583083a-b161-492e-93c6-088fb16c06fe-kube-api-access-gg827\") pod \"collect-profiles-29405625-ncrbh\" (UID: \"9583083a-b161-492e-93c6-088fb16c06fe\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405625-ncrbh" Nov 28 13:45:00 crc kubenswrapper[4631]: I1128 13:45:00.378191 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9583083a-b161-492e-93c6-088fb16c06fe-config-volume\") pod \"collect-profiles-29405625-ncrbh\" (UID: \"9583083a-b161-492e-93c6-088fb16c06fe\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405625-ncrbh" Nov 28 13:45:00 crc kubenswrapper[4631]: I1128 13:45:00.378239 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/9583083a-b161-492e-93c6-088fb16c06fe-secret-volume\") pod \"collect-profiles-29405625-ncrbh\" (UID: \"9583083a-b161-492e-93c6-088fb16c06fe\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405625-ncrbh" Nov 28 13:45:00 crc kubenswrapper[4631]: I1128 13:45:00.380648 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9583083a-b161-492e-93c6-088fb16c06fe-config-volume\") pod \"collect-profiles-29405625-ncrbh\" (UID: \"9583083a-b161-492e-93c6-088fb16c06fe\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405625-ncrbh" Nov 28 13:45:00 crc kubenswrapper[4631]: I1128 13:45:00.388096 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/9583083a-b161-492e-93c6-088fb16c06fe-secret-volume\") pod \"collect-profiles-29405625-ncrbh\" (UID: \"9583083a-b161-492e-93c6-088fb16c06fe\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405625-ncrbh" Nov 28 13:45:00 crc kubenswrapper[4631]: I1128 13:45:00.396873 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gg827\" (UniqueName: \"kubernetes.io/projected/9583083a-b161-492e-93c6-088fb16c06fe-kube-api-access-gg827\") pod \"collect-profiles-29405625-ncrbh\" (UID: \"9583083a-b161-492e-93c6-088fb16c06fe\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405625-ncrbh" Nov 28 13:45:00 crc kubenswrapper[4631]: I1128 13:45:00.489663 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-c8xj5" event={"ID":"0a640440-9c58-4899-967e-164e0f3ab927","Type":"ContainerStarted","Data":"77721bfbe18c79b2e27da386e8e5a052bb3c51f68fb1ad03b405f415e128e531"} Nov 28 13:45:00 crc kubenswrapper[4631]: I1128 13:45:00.548138 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29405625-ncrbh" Nov 28 13:45:01 crc kubenswrapper[4631]: I1128 13:45:01.310032 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29405625-ncrbh"] Nov 28 13:45:01 crc kubenswrapper[4631]: W1128 13:45:01.319517 4631 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9583083a_b161_492e_93c6_088fb16c06fe.slice/crio-9cbc1eb6fbd73b6dcd2f2ba7d38aed758690ab7aa228e63636eacfdba4a3d440 WatchSource:0}: Error finding container 9cbc1eb6fbd73b6dcd2f2ba7d38aed758690ab7aa228e63636eacfdba4a3d440: Status 404 returned error can't find the container with id 9cbc1eb6fbd73b6dcd2f2ba7d38aed758690ab7aa228e63636eacfdba4a3d440 Nov 28 13:45:01 crc kubenswrapper[4631]: I1128 13:45:01.504176 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29405625-ncrbh" event={"ID":"9583083a-b161-492e-93c6-088fb16c06fe","Type":"ContainerStarted","Data":"9cbc1eb6fbd73b6dcd2f2ba7d38aed758690ab7aa228e63636eacfdba4a3d440"} Nov 28 13:45:02 crc kubenswrapper[4631]: I1128 13:45:02.235697 4631 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-wg5bg" podUID="6c545b0d-6224-497d-b190-e9f619d340b2" containerName="registry-server" probeResult="failure" output=< Nov 28 13:45:02 crc kubenswrapper[4631]: timeout: failed to connect service ":50051" within 1s Nov 28 13:45:02 crc kubenswrapper[4631]: > Nov 28 13:45:02 crc kubenswrapper[4631]: I1128 13:45:02.521984 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29405625-ncrbh" event={"ID":"9583083a-b161-492e-93c6-088fb16c06fe","Type":"ContainerStarted","Data":"19f40940349ba6d21797abd21b448bca573afc98fe0bfafd6332b170448653c7"} Nov 28 13:45:02 crc kubenswrapper[4631]: I1128 13:45:02.540006 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29405625-ncrbh" podStartSLOduration=2.539978557 podStartE2EDuration="2.539978557s" podCreationTimestamp="2025-11-28 13:45:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 13:45:02.537625262 +0000 UTC m=+1459.344928606" watchObservedRunningTime="2025-11-28 13:45:02.539978557 +0000 UTC m=+1459.347281901" Nov 28 13:45:03 crc kubenswrapper[4631]: I1128 13:45:03.563846 4631 generic.go:334] "Generic (PLEG): container finished" podID="9583083a-b161-492e-93c6-088fb16c06fe" containerID="19f40940349ba6d21797abd21b448bca573afc98fe0bfafd6332b170448653c7" exitCode=0 Nov 28 13:45:03 crc kubenswrapper[4631]: I1128 13:45:03.563938 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29405625-ncrbh" event={"ID":"9583083a-b161-492e-93c6-088fb16c06fe","Type":"ContainerDied","Data":"19f40940349ba6d21797abd21b448bca573afc98fe0bfafd6332b170448653c7"} Nov 28 13:45:03 crc kubenswrapper[4631]: I1128 13:45:03.568016 4631 generic.go:334] "Generic (PLEG): container finished" podID="0a640440-9c58-4899-967e-164e0f3ab927" containerID="77721bfbe18c79b2e27da386e8e5a052bb3c51f68fb1ad03b405f415e128e531" exitCode=0 Nov 28 13:45:03 crc kubenswrapper[4631]: I1128 13:45:03.568078 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-c8xj5" event={"ID":"0a640440-9c58-4899-967e-164e0f3ab927","Type":"ContainerDied","Data":"77721bfbe18c79b2e27da386e8e5a052bb3c51f68fb1ad03b405f415e128e531"} Nov 28 13:45:05 crc kubenswrapper[4631]: I1128 13:45:05.036826 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29405625-ncrbh" Nov 28 13:45:05 crc kubenswrapper[4631]: I1128 13:45:05.085423 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9583083a-b161-492e-93c6-088fb16c06fe-config-volume\") pod \"9583083a-b161-492e-93c6-088fb16c06fe\" (UID: \"9583083a-b161-492e-93c6-088fb16c06fe\") " Nov 28 13:45:05 crc kubenswrapper[4631]: I1128 13:45:05.085555 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/9583083a-b161-492e-93c6-088fb16c06fe-secret-volume\") pod \"9583083a-b161-492e-93c6-088fb16c06fe\" (UID: \"9583083a-b161-492e-93c6-088fb16c06fe\") " Nov 28 13:45:05 crc kubenswrapper[4631]: I1128 13:45:05.085612 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gg827\" (UniqueName: \"kubernetes.io/projected/9583083a-b161-492e-93c6-088fb16c06fe-kube-api-access-gg827\") pod \"9583083a-b161-492e-93c6-088fb16c06fe\" (UID: \"9583083a-b161-492e-93c6-088fb16c06fe\") " Nov 28 13:45:05 crc kubenswrapper[4631]: I1128 13:45:05.086188 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9583083a-b161-492e-93c6-088fb16c06fe-config-volume" (OuterVolumeSpecName: "config-volume") pod "9583083a-b161-492e-93c6-088fb16c06fe" (UID: "9583083a-b161-492e-93c6-088fb16c06fe"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 13:45:05 crc kubenswrapper[4631]: I1128 13:45:05.092179 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9583083a-b161-492e-93c6-088fb16c06fe-kube-api-access-gg827" (OuterVolumeSpecName: "kube-api-access-gg827") pod "9583083a-b161-492e-93c6-088fb16c06fe" (UID: "9583083a-b161-492e-93c6-088fb16c06fe"). InnerVolumeSpecName "kube-api-access-gg827". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:45:05 crc kubenswrapper[4631]: I1128 13:45:05.094392 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9583083a-b161-492e-93c6-088fb16c06fe-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "9583083a-b161-492e-93c6-088fb16c06fe" (UID: "9583083a-b161-492e-93c6-088fb16c06fe"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:45:05 crc kubenswrapper[4631]: I1128 13:45:05.187911 4631 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/9583083a-b161-492e-93c6-088fb16c06fe-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 28 13:45:05 crc kubenswrapper[4631]: I1128 13:45:05.188186 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gg827\" (UniqueName: \"kubernetes.io/projected/9583083a-b161-492e-93c6-088fb16c06fe-kube-api-access-gg827\") on node \"crc\" DevicePath \"\"" Nov 28 13:45:05 crc kubenswrapper[4631]: I1128 13:45:05.188263 4631 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9583083a-b161-492e-93c6-088fb16c06fe-config-volume\") on node \"crc\" DevicePath \"\"" Nov 28 13:45:05 crc kubenswrapper[4631]: I1128 13:45:05.607396 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-c8xj5" event={"ID":"0a640440-9c58-4899-967e-164e0f3ab927","Type":"ContainerStarted","Data":"78f9f2873c2eb631c7aa658384e0cb3671c196bab57405be8071a78c0081f0d4"} Nov 28 13:45:05 crc kubenswrapper[4631]: I1128 13:45:05.608832 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29405625-ncrbh" event={"ID":"9583083a-b161-492e-93c6-088fb16c06fe","Type":"ContainerDied","Data":"9cbc1eb6fbd73b6dcd2f2ba7d38aed758690ab7aa228e63636eacfdba4a3d440"} Nov 28 13:45:05 crc kubenswrapper[4631]: I1128 13:45:05.608859 4631 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9cbc1eb6fbd73b6dcd2f2ba7d38aed758690ab7aa228e63636eacfdba4a3d440" Nov 28 13:45:05 crc kubenswrapper[4631]: I1128 13:45:05.608874 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29405625-ncrbh" Nov 28 13:45:05 crc kubenswrapper[4631]: I1128 13:45:05.637488 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-c8xj5" podStartSLOduration=3.654233291 podStartE2EDuration="8.637462776s" podCreationTimestamp="2025-11-28 13:44:57 +0000 UTC" firstStartedPulling="2025-11-28 13:44:59.481038732 +0000 UTC m=+1456.288342086" lastFinishedPulling="2025-11-28 13:45:04.464268227 +0000 UTC m=+1461.271571571" observedRunningTime="2025-11-28 13:45:05.626360573 +0000 UTC m=+1462.433663927" watchObservedRunningTime="2025-11-28 13:45:05.637462776 +0000 UTC m=+1462.444766110" Nov 28 13:45:07 crc kubenswrapper[4631]: I1128 13:45:07.593746 4631 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-c8xj5" Nov 28 13:45:07 crc kubenswrapper[4631]: I1128 13:45:07.594146 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-c8xj5" Nov 28 13:45:07 crc kubenswrapper[4631]: I1128 13:45:07.648978 4631 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-c8xj5" Nov 28 13:45:11 crc kubenswrapper[4631]: I1128 13:45:11.185303 4631 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-wg5bg" Nov 28 13:45:11 crc kubenswrapper[4631]: I1128 13:45:11.247777 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-wg5bg" Nov 28 13:45:11 crc kubenswrapper[4631]: I1128 13:45:11.965567 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-wg5bg"] Nov 28 13:45:12 crc kubenswrapper[4631]: I1128 13:45:12.694383 4631 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-wg5bg" podUID="6c545b0d-6224-497d-b190-e9f619d340b2" containerName="registry-server" containerID="cri-o://a5cdb60a8c7b33ae92566eaffc9e17c69b1322b3fb5638789fe13d9bf0e7f786" gracePeriod=2 Nov 28 13:45:13 crc kubenswrapper[4631]: I1128 13:45:13.232408 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wg5bg" Nov 28 13:45:13 crc kubenswrapper[4631]: I1128 13:45:13.283443 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6c545b0d-6224-497d-b190-e9f619d340b2-catalog-content\") pod \"6c545b0d-6224-497d-b190-e9f619d340b2\" (UID: \"6c545b0d-6224-497d-b190-e9f619d340b2\") " Nov 28 13:45:13 crc kubenswrapper[4631]: I1128 13:45:13.283544 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6c545b0d-6224-497d-b190-e9f619d340b2-utilities\") pod \"6c545b0d-6224-497d-b190-e9f619d340b2\" (UID: \"6c545b0d-6224-497d-b190-e9f619d340b2\") " Nov 28 13:45:13 crc kubenswrapper[4631]: I1128 13:45:13.283657 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c9nj7\" (UniqueName: \"kubernetes.io/projected/6c545b0d-6224-497d-b190-e9f619d340b2-kube-api-access-c9nj7\") pod \"6c545b0d-6224-497d-b190-e9f619d340b2\" (UID: \"6c545b0d-6224-497d-b190-e9f619d340b2\") " Nov 28 13:45:13 crc kubenswrapper[4631]: I1128 13:45:13.284353 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6c545b0d-6224-497d-b190-e9f619d340b2-utilities" (OuterVolumeSpecName: "utilities") pod "6c545b0d-6224-497d-b190-e9f619d340b2" (UID: "6c545b0d-6224-497d-b190-e9f619d340b2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 13:45:13 crc kubenswrapper[4631]: I1128 13:45:13.311894 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6c545b0d-6224-497d-b190-e9f619d340b2-kube-api-access-c9nj7" (OuterVolumeSpecName: "kube-api-access-c9nj7") pod "6c545b0d-6224-497d-b190-e9f619d340b2" (UID: "6c545b0d-6224-497d-b190-e9f619d340b2"). InnerVolumeSpecName "kube-api-access-c9nj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:45:13 crc kubenswrapper[4631]: I1128 13:45:13.386704 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c9nj7\" (UniqueName: \"kubernetes.io/projected/6c545b0d-6224-497d-b190-e9f619d340b2-kube-api-access-c9nj7\") on node \"crc\" DevicePath \"\"" Nov 28 13:45:13 crc kubenswrapper[4631]: I1128 13:45:13.387067 4631 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6c545b0d-6224-497d-b190-e9f619d340b2-utilities\") on node \"crc\" DevicePath \"\"" Nov 28 13:45:13 crc kubenswrapper[4631]: I1128 13:45:13.419104 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6c545b0d-6224-497d-b190-e9f619d340b2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6c545b0d-6224-497d-b190-e9f619d340b2" (UID: "6c545b0d-6224-497d-b190-e9f619d340b2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 13:45:13 crc kubenswrapper[4631]: I1128 13:45:13.489066 4631 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6c545b0d-6224-497d-b190-e9f619d340b2-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 28 13:45:13 crc kubenswrapper[4631]: I1128 13:45:13.710674 4631 generic.go:334] "Generic (PLEG): container finished" podID="6c545b0d-6224-497d-b190-e9f619d340b2" containerID="a5cdb60a8c7b33ae92566eaffc9e17c69b1322b3fb5638789fe13d9bf0e7f786" exitCode=0 Nov 28 13:45:13 crc kubenswrapper[4631]: I1128 13:45:13.710743 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wg5bg" event={"ID":"6c545b0d-6224-497d-b190-e9f619d340b2","Type":"ContainerDied","Data":"a5cdb60a8c7b33ae92566eaffc9e17c69b1322b3fb5638789fe13d9bf0e7f786"} Nov 28 13:45:13 crc kubenswrapper[4631]: I1128 13:45:13.710804 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wg5bg" event={"ID":"6c545b0d-6224-497d-b190-e9f619d340b2","Type":"ContainerDied","Data":"71ceecc7f09b2810a809c93779bd44b7ed450a6c40810c2cfeaec52a91241606"} Nov 28 13:45:13 crc kubenswrapper[4631]: I1128 13:45:13.710832 4631 scope.go:117] "RemoveContainer" containerID="a5cdb60a8c7b33ae92566eaffc9e17c69b1322b3fb5638789fe13d9bf0e7f786" Nov 28 13:45:13 crc kubenswrapper[4631]: I1128 13:45:13.710829 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wg5bg" Nov 28 13:45:13 crc kubenswrapper[4631]: I1128 13:45:13.734328 4631 scope.go:117] "RemoveContainer" containerID="7e00b92e6e594984f05d2d7a6501956b32472077089937f74f54aa6f7f2e1f61" Nov 28 13:45:13 crc kubenswrapper[4631]: I1128 13:45:13.747658 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-wg5bg"] Nov 28 13:45:13 crc kubenswrapper[4631]: I1128 13:45:13.756541 4631 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-wg5bg"] Nov 28 13:45:13 crc kubenswrapper[4631]: I1128 13:45:13.769093 4631 scope.go:117] "RemoveContainer" containerID="7c37ecdf0fbc5e39f49ae4ad205831cedc0d9100393516343642ace7813acbf7" Nov 28 13:45:13 crc kubenswrapper[4631]: I1128 13:45:13.815812 4631 scope.go:117] "RemoveContainer" containerID="a5cdb60a8c7b33ae92566eaffc9e17c69b1322b3fb5638789fe13d9bf0e7f786" Nov 28 13:45:13 crc kubenswrapper[4631]: E1128 13:45:13.816460 4631 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a5cdb60a8c7b33ae92566eaffc9e17c69b1322b3fb5638789fe13d9bf0e7f786\": container with ID starting with a5cdb60a8c7b33ae92566eaffc9e17c69b1322b3fb5638789fe13d9bf0e7f786 not found: ID does not exist" containerID="a5cdb60a8c7b33ae92566eaffc9e17c69b1322b3fb5638789fe13d9bf0e7f786" Nov 28 13:45:13 crc kubenswrapper[4631]: I1128 13:45:13.816552 4631 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a5cdb60a8c7b33ae92566eaffc9e17c69b1322b3fb5638789fe13d9bf0e7f786"} err="failed to get container status \"a5cdb60a8c7b33ae92566eaffc9e17c69b1322b3fb5638789fe13d9bf0e7f786\": rpc error: code = NotFound desc = could not find container \"a5cdb60a8c7b33ae92566eaffc9e17c69b1322b3fb5638789fe13d9bf0e7f786\": container with ID starting with a5cdb60a8c7b33ae92566eaffc9e17c69b1322b3fb5638789fe13d9bf0e7f786 not found: ID does not exist" Nov 28 13:45:13 crc kubenswrapper[4631]: I1128 13:45:13.816598 4631 scope.go:117] "RemoveContainer" containerID="7e00b92e6e594984f05d2d7a6501956b32472077089937f74f54aa6f7f2e1f61" Nov 28 13:45:13 crc kubenswrapper[4631]: E1128 13:45:13.817405 4631 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7e00b92e6e594984f05d2d7a6501956b32472077089937f74f54aa6f7f2e1f61\": container with ID starting with 7e00b92e6e594984f05d2d7a6501956b32472077089937f74f54aa6f7f2e1f61 not found: ID does not exist" containerID="7e00b92e6e594984f05d2d7a6501956b32472077089937f74f54aa6f7f2e1f61" Nov 28 13:45:13 crc kubenswrapper[4631]: I1128 13:45:13.817461 4631 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7e00b92e6e594984f05d2d7a6501956b32472077089937f74f54aa6f7f2e1f61"} err="failed to get container status \"7e00b92e6e594984f05d2d7a6501956b32472077089937f74f54aa6f7f2e1f61\": rpc error: code = NotFound desc = could not find container \"7e00b92e6e594984f05d2d7a6501956b32472077089937f74f54aa6f7f2e1f61\": container with ID starting with 7e00b92e6e594984f05d2d7a6501956b32472077089937f74f54aa6f7f2e1f61 not found: ID does not exist" Nov 28 13:45:13 crc kubenswrapper[4631]: I1128 13:45:13.817493 4631 scope.go:117] "RemoveContainer" containerID="7c37ecdf0fbc5e39f49ae4ad205831cedc0d9100393516343642ace7813acbf7" Nov 28 13:45:13 crc kubenswrapper[4631]: E1128 13:45:13.817762 4631 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7c37ecdf0fbc5e39f49ae4ad205831cedc0d9100393516343642ace7813acbf7\": container with ID starting with 7c37ecdf0fbc5e39f49ae4ad205831cedc0d9100393516343642ace7813acbf7 not found: ID does not exist" containerID="7c37ecdf0fbc5e39f49ae4ad205831cedc0d9100393516343642ace7813acbf7" Nov 28 13:45:13 crc kubenswrapper[4631]: I1128 13:45:13.817788 4631 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7c37ecdf0fbc5e39f49ae4ad205831cedc0d9100393516343642ace7813acbf7"} err="failed to get container status \"7c37ecdf0fbc5e39f49ae4ad205831cedc0d9100393516343642ace7813acbf7\": rpc error: code = NotFound desc = could not find container \"7c37ecdf0fbc5e39f49ae4ad205831cedc0d9100393516343642ace7813acbf7\": container with ID starting with 7c37ecdf0fbc5e39f49ae4ad205831cedc0d9100393516343642ace7813acbf7 not found: ID does not exist" Nov 28 13:45:15 crc kubenswrapper[4631]: I1128 13:45:15.548614 4631 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6c545b0d-6224-497d-b190-e9f619d340b2" path="/var/lib/kubelet/pods/6c545b0d-6224-497d-b190-e9f619d340b2/volumes" Nov 28 13:45:17 crc kubenswrapper[4631]: I1128 13:45:17.647844 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-c8xj5" Nov 28 13:45:17 crc kubenswrapper[4631]: I1128 13:45:17.719348 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-c8xj5"] Nov 28 13:45:17 crc kubenswrapper[4631]: I1128 13:45:17.760069 4631 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-c8xj5" podUID="0a640440-9c58-4899-967e-164e0f3ab927" containerName="registry-server" containerID="cri-o://78f9f2873c2eb631c7aa658384e0cb3671c196bab57405be8071a78c0081f0d4" gracePeriod=2 Nov 28 13:45:18 crc kubenswrapper[4631]: I1128 13:45:18.254836 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-c8xj5" Nov 28 13:45:18 crc kubenswrapper[4631]: I1128 13:45:18.305795 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0a640440-9c58-4899-967e-164e0f3ab927-catalog-content\") pod \"0a640440-9c58-4899-967e-164e0f3ab927\" (UID: \"0a640440-9c58-4899-967e-164e0f3ab927\") " Nov 28 13:45:18 crc kubenswrapper[4631]: I1128 13:45:18.305926 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lljw2\" (UniqueName: \"kubernetes.io/projected/0a640440-9c58-4899-967e-164e0f3ab927-kube-api-access-lljw2\") pod \"0a640440-9c58-4899-967e-164e0f3ab927\" (UID: \"0a640440-9c58-4899-967e-164e0f3ab927\") " Nov 28 13:45:18 crc kubenswrapper[4631]: I1128 13:45:18.306022 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0a640440-9c58-4899-967e-164e0f3ab927-utilities\") pod \"0a640440-9c58-4899-967e-164e0f3ab927\" (UID: \"0a640440-9c58-4899-967e-164e0f3ab927\") " Nov 28 13:45:18 crc kubenswrapper[4631]: I1128 13:45:18.307018 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0a640440-9c58-4899-967e-164e0f3ab927-utilities" (OuterVolumeSpecName: "utilities") pod "0a640440-9c58-4899-967e-164e0f3ab927" (UID: "0a640440-9c58-4899-967e-164e0f3ab927"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 13:45:18 crc kubenswrapper[4631]: I1128 13:45:18.345992 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0a640440-9c58-4899-967e-164e0f3ab927-kube-api-access-lljw2" (OuterVolumeSpecName: "kube-api-access-lljw2") pod "0a640440-9c58-4899-967e-164e0f3ab927" (UID: "0a640440-9c58-4899-967e-164e0f3ab927"). InnerVolumeSpecName "kube-api-access-lljw2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:45:18 crc kubenswrapper[4631]: I1128 13:45:18.395117 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0a640440-9c58-4899-967e-164e0f3ab927-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0a640440-9c58-4899-967e-164e0f3ab927" (UID: "0a640440-9c58-4899-967e-164e0f3ab927"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 13:45:18 crc kubenswrapper[4631]: I1128 13:45:18.409639 4631 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0a640440-9c58-4899-967e-164e0f3ab927-utilities\") on node \"crc\" DevicePath \"\"" Nov 28 13:45:18 crc kubenswrapper[4631]: I1128 13:45:18.409680 4631 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0a640440-9c58-4899-967e-164e0f3ab927-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 28 13:45:18 crc kubenswrapper[4631]: I1128 13:45:18.409696 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lljw2\" (UniqueName: \"kubernetes.io/projected/0a640440-9c58-4899-967e-164e0f3ab927-kube-api-access-lljw2\") on node \"crc\" DevicePath \"\"" Nov 28 13:45:18 crc kubenswrapper[4631]: I1128 13:45:18.774631 4631 generic.go:334] "Generic (PLEG): container finished" podID="0a640440-9c58-4899-967e-164e0f3ab927" containerID="78f9f2873c2eb631c7aa658384e0cb3671c196bab57405be8071a78c0081f0d4" exitCode=0 Nov 28 13:45:18 crc kubenswrapper[4631]: I1128 13:45:18.774693 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-c8xj5" event={"ID":"0a640440-9c58-4899-967e-164e0f3ab927","Type":"ContainerDied","Data":"78f9f2873c2eb631c7aa658384e0cb3671c196bab57405be8071a78c0081f0d4"} Nov 28 13:45:18 crc kubenswrapper[4631]: I1128 13:45:18.774732 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-c8xj5" event={"ID":"0a640440-9c58-4899-967e-164e0f3ab927","Type":"ContainerDied","Data":"d35ea7fb9be621f9addf2c73d5143dad96f69a53cbae95f9448e6ef5dfc75cef"} Nov 28 13:45:18 crc kubenswrapper[4631]: I1128 13:45:18.774756 4631 scope.go:117] "RemoveContainer" containerID="78f9f2873c2eb631c7aa658384e0cb3671c196bab57405be8071a78c0081f0d4" Nov 28 13:45:18 crc kubenswrapper[4631]: I1128 13:45:18.774930 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-c8xj5" Nov 28 13:45:18 crc kubenswrapper[4631]: I1128 13:45:18.815676 4631 scope.go:117] "RemoveContainer" containerID="77721bfbe18c79b2e27da386e8e5a052bb3c51f68fb1ad03b405f415e128e531" Nov 28 13:45:18 crc kubenswrapper[4631]: I1128 13:45:18.819057 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-c8xj5"] Nov 28 13:45:18 crc kubenswrapper[4631]: I1128 13:45:18.832896 4631 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-c8xj5"] Nov 28 13:45:18 crc kubenswrapper[4631]: I1128 13:45:18.843883 4631 scope.go:117] "RemoveContainer" containerID="f1a120bfde8f3fd81dd0f611449df3af3c75cfc45cfc20af575485a021fc7215" Nov 28 13:45:18 crc kubenswrapper[4631]: I1128 13:45:18.893619 4631 scope.go:117] "RemoveContainer" containerID="78f9f2873c2eb631c7aa658384e0cb3671c196bab57405be8071a78c0081f0d4" Nov 28 13:45:18 crc kubenswrapper[4631]: E1128 13:45:18.894349 4631 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"78f9f2873c2eb631c7aa658384e0cb3671c196bab57405be8071a78c0081f0d4\": container with ID starting with 78f9f2873c2eb631c7aa658384e0cb3671c196bab57405be8071a78c0081f0d4 not found: ID does not exist" containerID="78f9f2873c2eb631c7aa658384e0cb3671c196bab57405be8071a78c0081f0d4" Nov 28 13:45:18 crc kubenswrapper[4631]: I1128 13:45:18.894429 4631 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"78f9f2873c2eb631c7aa658384e0cb3671c196bab57405be8071a78c0081f0d4"} err="failed to get container status \"78f9f2873c2eb631c7aa658384e0cb3671c196bab57405be8071a78c0081f0d4\": rpc error: code = NotFound desc = could not find container \"78f9f2873c2eb631c7aa658384e0cb3671c196bab57405be8071a78c0081f0d4\": container with ID starting with 78f9f2873c2eb631c7aa658384e0cb3671c196bab57405be8071a78c0081f0d4 not found: ID does not exist" Nov 28 13:45:18 crc kubenswrapper[4631]: I1128 13:45:18.894462 4631 scope.go:117] "RemoveContainer" containerID="77721bfbe18c79b2e27da386e8e5a052bb3c51f68fb1ad03b405f415e128e531" Nov 28 13:45:18 crc kubenswrapper[4631]: E1128 13:45:18.895146 4631 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"77721bfbe18c79b2e27da386e8e5a052bb3c51f68fb1ad03b405f415e128e531\": container with ID starting with 77721bfbe18c79b2e27da386e8e5a052bb3c51f68fb1ad03b405f415e128e531 not found: ID does not exist" containerID="77721bfbe18c79b2e27da386e8e5a052bb3c51f68fb1ad03b405f415e128e531" Nov 28 13:45:18 crc kubenswrapper[4631]: I1128 13:45:18.895214 4631 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"77721bfbe18c79b2e27da386e8e5a052bb3c51f68fb1ad03b405f415e128e531"} err="failed to get container status \"77721bfbe18c79b2e27da386e8e5a052bb3c51f68fb1ad03b405f415e128e531\": rpc error: code = NotFound desc = could not find container \"77721bfbe18c79b2e27da386e8e5a052bb3c51f68fb1ad03b405f415e128e531\": container with ID starting with 77721bfbe18c79b2e27da386e8e5a052bb3c51f68fb1ad03b405f415e128e531 not found: ID does not exist" Nov 28 13:45:18 crc kubenswrapper[4631]: I1128 13:45:18.895260 4631 scope.go:117] "RemoveContainer" containerID="f1a120bfde8f3fd81dd0f611449df3af3c75cfc45cfc20af575485a021fc7215" Nov 28 13:45:18 crc kubenswrapper[4631]: E1128 13:45:18.895746 4631 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f1a120bfde8f3fd81dd0f611449df3af3c75cfc45cfc20af575485a021fc7215\": container with ID starting with f1a120bfde8f3fd81dd0f611449df3af3c75cfc45cfc20af575485a021fc7215 not found: ID does not exist" containerID="f1a120bfde8f3fd81dd0f611449df3af3c75cfc45cfc20af575485a021fc7215" Nov 28 13:45:18 crc kubenswrapper[4631]: I1128 13:45:18.895858 4631 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f1a120bfde8f3fd81dd0f611449df3af3c75cfc45cfc20af575485a021fc7215"} err="failed to get container status \"f1a120bfde8f3fd81dd0f611449df3af3c75cfc45cfc20af575485a021fc7215\": rpc error: code = NotFound desc = could not find container \"f1a120bfde8f3fd81dd0f611449df3af3c75cfc45cfc20af575485a021fc7215\": container with ID starting with f1a120bfde8f3fd81dd0f611449df3af3c75cfc45cfc20af575485a021fc7215 not found: ID does not exist" Nov 28 13:45:19 crc kubenswrapper[4631]: I1128 13:45:19.528513 4631 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0a640440-9c58-4899-967e-164e0f3ab927" path="/var/lib/kubelet/pods/0a640440-9c58-4899-967e-164e0f3ab927/volumes" Nov 28 13:45:27 crc kubenswrapper[4631]: I1128 13:45:27.230181 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-mtc24"] Nov 28 13:45:27 crc kubenswrapper[4631]: E1128 13:45:27.231394 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6c545b0d-6224-497d-b190-e9f619d340b2" containerName="registry-server" Nov 28 13:45:27 crc kubenswrapper[4631]: I1128 13:45:27.231410 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="6c545b0d-6224-497d-b190-e9f619d340b2" containerName="registry-server" Nov 28 13:45:27 crc kubenswrapper[4631]: E1128 13:45:27.231423 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6c545b0d-6224-497d-b190-e9f619d340b2" containerName="extract-content" Nov 28 13:45:27 crc kubenswrapper[4631]: I1128 13:45:27.231429 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="6c545b0d-6224-497d-b190-e9f619d340b2" containerName="extract-content" Nov 28 13:45:27 crc kubenswrapper[4631]: E1128 13:45:27.231443 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a640440-9c58-4899-967e-164e0f3ab927" containerName="registry-server" Nov 28 13:45:27 crc kubenswrapper[4631]: I1128 13:45:27.231450 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a640440-9c58-4899-967e-164e0f3ab927" containerName="registry-server" Nov 28 13:45:27 crc kubenswrapper[4631]: E1128 13:45:27.231473 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a640440-9c58-4899-967e-164e0f3ab927" containerName="extract-content" Nov 28 13:45:27 crc kubenswrapper[4631]: I1128 13:45:27.231481 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a640440-9c58-4899-967e-164e0f3ab927" containerName="extract-content" Nov 28 13:45:27 crc kubenswrapper[4631]: E1128 13:45:27.231500 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a640440-9c58-4899-967e-164e0f3ab927" containerName="extract-utilities" Nov 28 13:45:27 crc kubenswrapper[4631]: I1128 13:45:27.231508 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a640440-9c58-4899-967e-164e0f3ab927" containerName="extract-utilities" Nov 28 13:45:27 crc kubenswrapper[4631]: E1128 13:45:27.231524 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6c545b0d-6224-497d-b190-e9f619d340b2" containerName="extract-utilities" Nov 28 13:45:27 crc kubenswrapper[4631]: I1128 13:45:27.231531 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="6c545b0d-6224-497d-b190-e9f619d340b2" containerName="extract-utilities" Nov 28 13:45:27 crc kubenswrapper[4631]: E1128 13:45:27.231553 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9583083a-b161-492e-93c6-088fb16c06fe" containerName="collect-profiles" Nov 28 13:45:27 crc kubenswrapper[4631]: I1128 13:45:27.231560 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="9583083a-b161-492e-93c6-088fb16c06fe" containerName="collect-profiles" Nov 28 13:45:27 crc kubenswrapper[4631]: I1128 13:45:27.231748 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="0a640440-9c58-4899-967e-164e0f3ab927" containerName="registry-server" Nov 28 13:45:27 crc kubenswrapper[4631]: I1128 13:45:27.231769 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="6c545b0d-6224-497d-b190-e9f619d340b2" containerName="registry-server" Nov 28 13:45:27 crc kubenswrapper[4631]: I1128 13:45:27.231786 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="9583083a-b161-492e-93c6-088fb16c06fe" containerName="collect-profiles" Nov 28 13:45:27 crc kubenswrapper[4631]: I1128 13:45:27.233210 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mtc24" Nov 28 13:45:27 crc kubenswrapper[4631]: I1128 13:45:27.256072 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-mtc24"] Nov 28 13:45:27 crc kubenswrapper[4631]: I1128 13:45:27.329595 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2gbsl\" (UniqueName: \"kubernetes.io/projected/1f7a85b3-c1a6-4d7b-af11-8ed258e20818-kube-api-access-2gbsl\") pod \"redhat-marketplace-mtc24\" (UID: \"1f7a85b3-c1a6-4d7b-af11-8ed258e20818\") " pod="openshift-marketplace/redhat-marketplace-mtc24" Nov 28 13:45:27 crc kubenswrapper[4631]: I1128 13:45:27.329652 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1f7a85b3-c1a6-4d7b-af11-8ed258e20818-catalog-content\") pod \"redhat-marketplace-mtc24\" (UID: \"1f7a85b3-c1a6-4d7b-af11-8ed258e20818\") " pod="openshift-marketplace/redhat-marketplace-mtc24" Nov 28 13:45:27 crc kubenswrapper[4631]: I1128 13:45:27.329695 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1f7a85b3-c1a6-4d7b-af11-8ed258e20818-utilities\") pod \"redhat-marketplace-mtc24\" (UID: \"1f7a85b3-c1a6-4d7b-af11-8ed258e20818\") " pod="openshift-marketplace/redhat-marketplace-mtc24" Nov 28 13:45:27 crc kubenswrapper[4631]: I1128 13:45:27.432712 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2gbsl\" (UniqueName: \"kubernetes.io/projected/1f7a85b3-c1a6-4d7b-af11-8ed258e20818-kube-api-access-2gbsl\") pod \"redhat-marketplace-mtc24\" (UID: \"1f7a85b3-c1a6-4d7b-af11-8ed258e20818\") " pod="openshift-marketplace/redhat-marketplace-mtc24" Nov 28 13:45:27 crc kubenswrapper[4631]: I1128 13:45:27.433681 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1f7a85b3-c1a6-4d7b-af11-8ed258e20818-catalog-content\") pod \"redhat-marketplace-mtc24\" (UID: \"1f7a85b3-c1a6-4d7b-af11-8ed258e20818\") " pod="openshift-marketplace/redhat-marketplace-mtc24" Nov 28 13:45:27 crc kubenswrapper[4631]: I1128 13:45:27.433812 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1f7a85b3-c1a6-4d7b-af11-8ed258e20818-utilities\") pod \"redhat-marketplace-mtc24\" (UID: \"1f7a85b3-c1a6-4d7b-af11-8ed258e20818\") " pod="openshift-marketplace/redhat-marketplace-mtc24" Nov 28 13:45:27 crc kubenswrapper[4631]: I1128 13:45:27.434356 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1f7a85b3-c1a6-4d7b-af11-8ed258e20818-catalog-content\") pod \"redhat-marketplace-mtc24\" (UID: \"1f7a85b3-c1a6-4d7b-af11-8ed258e20818\") " pod="openshift-marketplace/redhat-marketplace-mtc24" Nov 28 13:45:27 crc kubenswrapper[4631]: I1128 13:45:27.434480 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1f7a85b3-c1a6-4d7b-af11-8ed258e20818-utilities\") pod \"redhat-marketplace-mtc24\" (UID: \"1f7a85b3-c1a6-4d7b-af11-8ed258e20818\") " pod="openshift-marketplace/redhat-marketplace-mtc24" Nov 28 13:45:27 crc kubenswrapper[4631]: I1128 13:45:27.459782 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2gbsl\" (UniqueName: \"kubernetes.io/projected/1f7a85b3-c1a6-4d7b-af11-8ed258e20818-kube-api-access-2gbsl\") pod \"redhat-marketplace-mtc24\" (UID: \"1f7a85b3-c1a6-4d7b-af11-8ed258e20818\") " pod="openshift-marketplace/redhat-marketplace-mtc24" Nov 28 13:45:27 crc kubenswrapper[4631]: I1128 13:45:27.554527 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mtc24" Nov 28 13:45:28 crc kubenswrapper[4631]: I1128 13:45:28.149882 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-mtc24"] Nov 28 13:45:28 crc kubenswrapper[4631]: I1128 13:45:28.893207 4631 generic.go:334] "Generic (PLEG): container finished" podID="1f7a85b3-c1a6-4d7b-af11-8ed258e20818" containerID="103678cf9023bc7f034cbda61452385a177c7a55ec0b69f286d022e6ef4288f7" exitCode=0 Nov 28 13:45:28 crc kubenswrapper[4631]: I1128 13:45:28.893310 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mtc24" event={"ID":"1f7a85b3-c1a6-4d7b-af11-8ed258e20818","Type":"ContainerDied","Data":"103678cf9023bc7f034cbda61452385a177c7a55ec0b69f286d022e6ef4288f7"} Nov 28 13:45:28 crc kubenswrapper[4631]: I1128 13:45:28.893767 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mtc24" event={"ID":"1f7a85b3-c1a6-4d7b-af11-8ed258e20818","Type":"ContainerStarted","Data":"c091f597dfad1e271d3003e828ac52d806f3af841efe24fc2b14ae370ef39ed1"} Nov 28 13:45:30 crc kubenswrapper[4631]: I1128 13:45:30.924623 4631 generic.go:334] "Generic (PLEG): container finished" podID="1f7a85b3-c1a6-4d7b-af11-8ed258e20818" containerID="58294307a8aa3b5c27e7792cfd44995aa925389897ba90302ad1c6d99d3d69ad" exitCode=0 Nov 28 13:45:30 crc kubenswrapper[4631]: I1128 13:45:30.924800 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mtc24" event={"ID":"1f7a85b3-c1a6-4d7b-af11-8ed258e20818","Type":"ContainerDied","Data":"58294307a8aa3b5c27e7792cfd44995aa925389897ba90302ad1c6d99d3d69ad"} Nov 28 13:45:31 crc kubenswrapper[4631]: I1128 13:45:31.010931 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-cnr7x"] Nov 28 13:45:31 crc kubenswrapper[4631]: I1128 13:45:31.013551 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cnr7x" Nov 28 13:45:31 crc kubenswrapper[4631]: I1128 13:45:31.026499 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-cnr7x"] Nov 28 13:45:31 crc kubenswrapper[4631]: I1128 13:45:31.113177 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f9dnw\" (UniqueName: \"kubernetes.io/projected/2007e6df-db00-4c6d-8632-07e934613379-kube-api-access-f9dnw\") pod \"community-operators-cnr7x\" (UID: \"2007e6df-db00-4c6d-8632-07e934613379\") " pod="openshift-marketplace/community-operators-cnr7x" Nov 28 13:45:31 crc kubenswrapper[4631]: I1128 13:45:31.113239 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2007e6df-db00-4c6d-8632-07e934613379-catalog-content\") pod \"community-operators-cnr7x\" (UID: \"2007e6df-db00-4c6d-8632-07e934613379\") " pod="openshift-marketplace/community-operators-cnr7x" Nov 28 13:45:31 crc kubenswrapper[4631]: I1128 13:45:31.113273 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2007e6df-db00-4c6d-8632-07e934613379-utilities\") pod \"community-operators-cnr7x\" (UID: \"2007e6df-db00-4c6d-8632-07e934613379\") " pod="openshift-marketplace/community-operators-cnr7x" Nov 28 13:45:31 crc kubenswrapper[4631]: I1128 13:45:31.215121 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f9dnw\" (UniqueName: \"kubernetes.io/projected/2007e6df-db00-4c6d-8632-07e934613379-kube-api-access-f9dnw\") pod \"community-operators-cnr7x\" (UID: \"2007e6df-db00-4c6d-8632-07e934613379\") " pod="openshift-marketplace/community-operators-cnr7x" Nov 28 13:45:31 crc kubenswrapper[4631]: I1128 13:45:31.215195 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2007e6df-db00-4c6d-8632-07e934613379-catalog-content\") pod \"community-operators-cnr7x\" (UID: \"2007e6df-db00-4c6d-8632-07e934613379\") " pod="openshift-marketplace/community-operators-cnr7x" Nov 28 13:45:31 crc kubenswrapper[4631]: I1128 13:45:31.215466 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2007e6df-db00-4c6d-8632-07e934613379-utilities\") pod \"community-operators-cnr7x\" (UID: \"2007e6df-db00-4c6d-8632-07e934613379\") " pod="openshift-marketplace/community-operators-cnr7x" Nov 28 13:45:31 crc kubenswrapper[4631]: I1128 13:45:31.215900 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2007e6df-db00-4c6d-8632-07e934613379-catalog-content\") pod \"community-operators-cnr7x\" (UID: \"2007e6df-db00-4c6d-8632-07e934613379\") " pod="openshift-marketplace/community-operators-cnr7x" Nov 28 13:45:31 crc kubenswrapper[4631]: I1128 13:45:31.216061 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2007e6df-db00-4c6d-8632-07e934613379-utilities\") pod \"community-operators-cnr7x\" (UID: \"2007e6df-db00-4c6d-8632-07e934613379\") " pod="openshift-marketplace/community-operators-cnr7x" Nov 28 13:45:31 crc kubenswrapper[4631]: I1128 13:45:31.241171 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f9dnw\" (UniqueName: \"kubernetes.io/projected/2007e6df-db00-4c6d-8632-07e934613379-kube-api-access-f9dnw\") pod \"community-operators-cnr7x\" (UID: \"2007e6df-db00-4c6d-8632-07e934613379\") " pod="openshift-marketplace/community-operators-cnr7x" Nov 28 13:45:31 crc kubenswrapper[4631]: I1128 13:45:31.332205 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cnr7x" Nov 28 13:45:31 crc kubenswrapper[4631]: I1128 13:45:31.888837 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-cnr7x"] Nov 28 13:45:31 crc kubenswrapper[4631]: I1128 13:45:31.942544 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cnr7x" event={"ID":"2007e6df-db00-4c6d-8632-07e934613379","Type":"ContainerStarted","Data":"6e5b039ad8f8de92a1a2b7cec88e81a387e1f64d53aa6669ef97f5a306f4e8f2"} Nov 28 13:45:31 crc kubenswrapper[4631]: I1128 13:45:31.950686 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mtc24" event={"ID":"1f7a85b3-c1a6-4d7b-af11-8ed258e20818","Type":"ContainerStarted","Data":"9a37b50e202f6b2b89e5204cfb371da286c594bbf3fc1385060e508d3e6a28cc"} Nov 28 13:45:32 crc kubenswrapper[4631]: I1128 13:45:32.962829 4631 generic.go:334] "Generic (PLEG): container finished" podID="2007e6df-db00-4c6d-8632-07e934613379" containerID="4eb99330c33ffc765ef2d76e009172630e8f9c2d448cffa3a4a0f169b0118a09" exitCode=0 Nov 28 13:45:32 crc kubenswrapper[4631]: I1128 13:45:32.962918 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cnr7x" event={"ID":"2007e6df-db00-4c6d-8632-07e934613379","Type":"ContainerDied","Data":"4eb99330c33ffc765ef2d76e009172630e8f9c2d448cffa3a4a0f169b0118a09"} Nov 28 13:45:33 crc kubenswrapper[4631]: I1128 13:45:33.031042 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-mtc24" podStartSLOduration=3.502132845 podStartE2EDuration="6.031005546s" podCreationTimestamp="2025-11-28 13:45:27 +0000 UTC" firstStartedPulling="2025-11-28 13:45:28.896839422 +0000 UTC m=+1485.704142766" lastFinishedPulling="2025-11-28 13:45:31.425712123 +0000 UTC m=+1488.233015467" observedRunningTime="2025-11-28 13:45:31.979932141 +0000 UTC m=+1488.787235495" watchObservedRunningTime="2025-11-28 13:45:33.031005546 +0000 UTC m=+1489.838308890" Nov 28 13:45:35 crc kubenswrapper[4631]: I1128 13:45:35.037471 4631 generic.go:334] "Generic (PLEG): container finished" podID="2007e6df-db00-4c6d-8632-07e934613379" containerID="14f060b32cc6c8c54088d4cd7a5131a4fafba27be925b150be4d2c314e98d98e" exitCode=0 Nov 28 13:45:35 crc kubenswrapper[4631]: I1128 13:45:35.037559 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cnr7x" event={"ID":"2007e6df-db00-4c6d-8632-07e934613379","Type":"ContainerDied","Data":"14f060b32cc6c8c54088d4cd7a5131a4fafba27be925b150be4d2c314e98d98e"} Nov 28 13:45:36 crc kubenswrapper[4631]: I1128 13:45:36.052960 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cnr7x" event={"ID":"2007e6df-db00-4c6d-8632-07e934613379","Type":"ContainerStarted","Data":"3b51fef12d68144d1c3040a73fa2dff6e366dbc0f51b4a8326124a4ed6c5bbe3"} Nov 28 13:45:36 crc kubenswrapper[4631]: I1128 13:45:36.087905 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-cnr7x" podStartSLOduration=3.24264224 podStartE2EDuration="6.087883373s" podCreationTimestamp="2025-11-28 13:45:30 +0000 UTC" firstStartedPulling="2025-11-28 13:45:32.96615013 +0000 UTC m=+1489.773453474" lastFinishedPulling="2025-11-28 13:45:35.811391263 +0000 UTC m=+1492.618694607" observedRunningTime="2025-11-28 13:45:36.075806646 +0000 UTC m=+1492.883110010" watchObservedRunningTime="2025-11-28 13:45:36.087883373 +0000 UTC m=+1492.895186717" Nov 28 13:45:37 crc kubenswrapper[4631]: I1128 13:45:37.554651 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-mtc24" Nov 28 13:45:37 crc kubenswrapper[4631]: I1128 13:45:37.555504 4631 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-mtc24" Nov 28 13:45:37 crc kubenswrapper[4631]: I1128 13:45:37.727976 4631 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-mtc24" Nov 28 13:45:38 crc kubenswrapper[4631]: I1128 13:45:38.130249 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-mtc24" Nov 28 13:45:38 crc kubenswrapper[4631]: I1128 13:45:38.794637 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-mtc24"] Nov 28 13:45:40 crc kubenswrapper[4631]: I1128 13:45:40.090840 4631 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-mtc24" podUID="1f7a85b3-c1a6-4d7b-af11-8ed258e20818" containerName="registry-server" containerID="cri-o://9a37b50e202f6b2b89e5204cfb371da286c594bbf3fc1385060e508d3e6a28cc" gracePeriod=2 Nov 28 13:45:41 crc kubenswrapper[4631]: I1128 13:45:41.027892 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mtc24" Nov 28 13:45:41 crc kubenswrapper[4631]: I1128 13:45:41.104423 4631 generic.go:334] "Generic (PLEG): container finished" podID="1f7a85b3-c1a6-4d7b-af11-8ed258e20818" containerID="9a37b50e202f6b2b89e5204cfb371da286c594bbf3fc1385060e508d3e6a28cc" exitCode=0 Nov 28 13:45:41 crc kubenswrapper[4631]: I1128 13:45:41.104468 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mtc24" event={"ID":"1f7a85b3-c1a6-4d7b-af11-8ed258e20818","Type":"ContainerDied","Data":"9a37b50e202f6b2b89e5204cfb371da286c594bbf3fc1385060e508d3e6a28cc"} Nov 28 13:45:41 crc kubenswrapper[4631]: I1128 13:45:41.104494 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mtc24" event={"ID":"1f7a85b3-c1a6-4d7b-af11-8ed258e20818","Type":"ContainerDied","Data":"c091f597dfad1e271d3003e828ac52d806f3af841efe24fc2b14ae370ef39ed1"} Nov 28 13:45:41 crc kubenswrapper[4631]: I1128 13:45:41.104513 4631 scope.go:117] "RemoveContainer" containerID="9a37b50e202f6b2b89e5204cfb371da286c594bbf3fc1385060e508d3e6a28cc" Nov 28 13:45:41 crc kubenswrapper[4631]: I1128 13:45:41.104639 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mtc24" Nov 28 13:45:41 crc kubenswrapper[4631]: I1128 13:45:41.127184 4631 scope.go:117] "RemoveContainer" containerID="58294307a8aa3b5c27e7792cfd44995aa925389897ba90302ad1c6d99d3d69ad" Nov 28 13:45:41 crc kubenswrapper[4631]: I1128 13:45:41.148047 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1f7a85b3-c1a6-4d7b-af11-8ed258e20818-utilities\") pod \"1f7a85b3-c1a6-4d7b-af11-8ed258e20818\" (UID: \"1f7a85b3-c1a6-4d7b-af11-8ed258e20818\") " Nov 28 13:45:41 crc kubenswrapper[4631]: I1128 13:45:41.148347 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1f7a85b3-c1a6-4d7b-af11-8ed258e20818-catalog-content\") pod \"1f7a85b3-c1a6-4d7b-af11-8ed258e20818\" (UID: \"1f7a85b3-c1a6-4d7b-af11-8ed258e20818\") " Nov 28 13:45:41 crc kubenswrapper[4631]: I1128 13:45:41.148458 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2gbsl\" (UniqueName: \"kubernetes.io/projected/1f7a85b3-c1a6-4d7b-af11-8ed258e20818-kube-api-access-2gbsl\") pod \"1f7a85b3-c1a6-4d7b-af11-8ed258e20818\" (UID: \"1f7a85b3-c1a6-4d7b-af11-8ed258e20818\") " Nov 28 13:45:41 crc kubenswrapper[4631]: I1128 13:45:41.152169 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1f7a85b3-c1a6-4d7b-af11-8ed258e20818-utilities" (OuterVolumeSpecName: "utilities") pod "1f7a85b3-c1a6-4d7b-af11-8ed258e20818" (UID: "1f7a85b3-c1a6-4d7b-af11-8ed258e20818"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 13:45:41 crc kubenswrapper[4631]: I1128 13:45:41.155095 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1f7a85b3-c1a6-4d7b-af11-8ed258e20818-kube-api-access-2gbsl" (OuterVolumeSpecName: "kube-api-access-2gbsl") pod "1f7a85b3-c1a6-4d7b-af11-8ed258e20818" (UID: "1f7a85b3-c1a6-4d7b-af11-8ed258e20818"). InnerVolumeSpecName "kube-api-access-2gbsl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:45:41 crc kubenswrapper[4631]: I1128 13:45:41.170125 4631 scope.go:117] "RemoveContainer" containerID="103678cf9023bc7f034cbda61452385a177c7a55ec0b69f286d022e6ef4288f7" Nov 28 13:45:41 crc kubenswrapper[4631]: I1128 13:45:41.172946 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1f7a85b3-c1a6-4d7b-af11-8ed258e20818-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1f7a85b3-c1a6-4d7b-af11-8ed258e20818" (UID: "1f7a85b3-c1a6-4d7b-af11-8ed258e20818"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 13:45:41 crc kubenswrapper[4631]: I1128 13:45:41.236340 4631 scope.go:117] "RemoveContainer" containerID="9a37b50e202f6b2b89e5204cfb371da286c594bbf3fc1385060e508d3e6a28cc" Nov 28 13:45:41 crc kubenswrapper[4631]: E1128 13:45:41.237108 4631 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9a37b50e202f6b2b89e5204cfb371da286c594bbf3fc1385060e508d3e6a28cc\": container with ID starting with 9a37b50e202f6b2b89e5204cfb371da286c594bbf3fc1385060e508d3e6a28cc not found: ID does not exist" containerID="9a37b50e202f6b2b89e5204cfb371da286c594bbf3fc1385060e508d3e6a28cc" Nov 28 13:45:41 crc kubenswrapper[4631]: I1128 13:45:41.237149 4631 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9a37b50e202f6b2b89e5204cfb371da286c594bbf3fc1385060e508d3e6a28cc"} err="failed to get container status \"9a37b50e202f6b2b89e5204cfb371da286c594bbf3fc1385060e508d3e6a28cc\": rpc error: code = NotFound desc = could not find container \"9a37b50e202f6b2b89e5204cfb371da286c594bbf3fc1385060e508d3e6a28cc\": container with ID starting with 9a37b50e202f6b2b89e5204cfb371da286c594bbf3fc1385060e508d3e6a28cc not found: ID does not exist" Nov 28 13:45:41 crc kubenswrapper[4631]: I1128 13:45:41.237209 4631 scope.go:117] "RemoveContainer" containerID="58294307a8aa3b5c27e7792cfd44995aa925389897ba90302ad1c6d99d3d69ad" Nov 28 13:45:41 crc kubenswrapper[4631]: E1128 13:45:41.237686 4631 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"58294307a8aa3b5c27e7792cfd44995aa925389897ba90302ad1c6d99d3d69ad\": container with ID starting with 58294307a8aa3b5c27e7792cfd44995aa925389897ba90302ad1c6d99d3d69ad not found: ID does not exist" containerID="58294307a8aa3b5c27e7792cfd44995aa925389897ba90302ad1c6d99d3d69ad" Nov 28 13:45:41 crc kubenswrapper[4631]: I1128 13:45:41.237728 4631 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"58294307a8aa3b5c27e7792cfd44995aa925389897ba90302ad1c6d99d3d69ad"} err="failed to get container status \"58294307a8aa3b5c27e7792cfd44995aa925389897ba90302ad1c6d99d3d69ad\": rpc error: code = NotFound desc = could not find container \"58294307a8aa3b5c27e7792cfd44995aa925389897ba90302ad1c6d99d3d69ad\": container with ID starting with 58294307a8aa3b5c27e7792cfd44995aa925389897ba90302ad1c6d99d3d69ad not found: ID does not exist" Nov 28 13:45:41 crc kubenswrapper[4631]: I1128 13:45:41.237745 4631 scope.go:117] "RemoveContainer" containerID="103678cf9023bc7f034cbda61452385a177c7a55ec0b69f286d022e6ef4288f7" Nov 28 13:45:41 crc kubenswrapper[4631]: E1128 13:45:41.238076 4631 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"103678cf9023bc7f034cbda61452385a177c7a55ec0b69f286d022e6ef4288f7\": container with ID starting with 103678cf9023bc7f034cbda61452385a177c7a55ec0b69f286d022e6ef4288f7 not found: ID does not exist" containerID="103678cf9023bc7f034cbda61452385a177c7a55ec0b69f286d022e6ef4288f7" Nov 28 13:45:41 crc kubenswrapper[4631]: I1128 13:45:41.238119 4631 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"103678cf9023bc7f034cbda61452385a177c7a55ec0b69f286d022e6ef4288f7"} err="failed to get container status \"103678cf9023bc7f034cbda61452385a177c7a55ec0b69f286d022e6ef4288f7\": rpc error: code = NotFound desc = could not find container \"103678cf9023bc7f034cbda61452385a177c7a55ec0b69f286d022e6ef4288f7\": container with ID starting with 103678cf9023bc7f034cbda61452385a177c7a55ec0b69f286d022e6ef4288f7 not found: ID does not exist" Nov 28 13:45:41 crc kubenswrapper[4631]: I1128 13:45:41.251042 4631 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1f7a85b3-c1a6-4d7b-af11-8ed258e20818-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 28 13:45:41 crc kubenswrapper[4631]: I1128 13:45:41.251106 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2gbsl\" (UniqueName: \"kubernetes.io/projected/1f7a85b3-c1a6-4d7b-af11-8ed258e20818-kube-api-access-2gbsl\") on node \"crc\" DevicePath \"\"" Nov 28 13:45:41 crc kubenswrapper[4631]: I1128 13:45:41.251121 4631 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1f7a85b3-c1a6-4d7b-af11-8ed258e20818-utilities\") on node \"crc\" DevicePath \"\"" Nov 28 13:45:41 crc kubenswrapper[4631]: I1128 13:45:41.332843 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-cnr7x" Nov 28 13:45:41 crc kubenswrapper[4631]: I1128 13:45:41.332915 4631 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-cnr7x" Nov 28 13:45:41 crc kubenswrapper[4631]: I1128 13:45:41.392884 4631 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-cnr7x" Nov 28 13:45:41 crc kubenswrapper[4631]: I1128 13:45:41.439144 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-mtc24"] Nov 28 13:45:41 crc kubenswrapper[4631]: I1128 13:45:41.448317 4631 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-mtc24"] Nov 28 13:45:41 crc kubenswrapper[4631]: I1128 13:45:41.523827 4631 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1f7a85b3-c1a6-4d7b-af11-8ed258e20818" path="/var/lib/kubelet/pods/1f7a85b3-c1a6-4d7b-af11-8ed258e20818/volumes" Nov 28 13:45:42 crc kubenswrapper[4631]: I1128 13:45:42.190122 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-cnr7x" Nov 28 13:45:43 crc kubenswrapper[4631]: I1128 13:45:43.793374 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-cnr7x"] Nov 28 13:45:44 crc kubenswrapper[4631]: I1128 13:45:44.142562 4631 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-cnr7x" podUID="2007e6df-db00-4c6d-8632-07e934613379" containerName="registry-server" containerID="cri-o://3b51fef12d68144d1c3040a73fa2dff6e366dbc0f51b4a8326124a4ed6c5bbe3" gracePeriod=2 Nov 28 13:45:44 crc kubenswrapper[4631]: I1128 13:45:44.598492 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cnr7x" Nov 28 13:45:44 crc kubenswrapper[4631]: I1128 13:45:44.728034 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f9dnw\" (UniqueName: \"kubernetes.io/projected/2007e6df-db00-4c6d-8632-07e934613379-kube-api-access-f9dnw\") pod \"2007e6df-db00-4c6d-8632-07e934613379\" (UID: \"2007e6df-db00-4c6d-8632-07e934613379\") " Nov 28 13:45:44 crc kubenswrapper[4631]: I1128 13:45:44.728138 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2007e6df-db00-4c6d-8632-07e934613379-utilities\") pod \"2007e6df-db00-4c6d-8632-07e934613379\" (UID: \"2007e6df-db00-4c6d-8632-07e934613379\") " Nov 28 13:45:44 crc kubenswrapper[4631]: I1128 13:45:44.728230 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2007e6df-db00-4c6d-8632-07e934613379-catalog-content\") pod \"2007e6df-db00-4c6d-8632-07e934613379\" (UID: \"2007e6df-db00-4c6d-8632-07e934613379\") " Nov 28 13:45:44 crc kubenswrapper[4631]: I1128 13:45:44.730322 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2007e6df-db00-4c6d-8632-07e934613379-utilities" (OuterVolumeSpecName: "utilities") pod "2007e6df-db00-4c6d-8632-07e934613379" (UID: "2007e6df-db00-4c6d-8632-07e934613379"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 13:45:44 crc kubenswrapper[4631]: I1128 13:45:44.734943 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2007e6df-db00-4c6d-8632-07e934613379-kube-api-access-f9dnw" (OuterVolumeSpecName: "kube-api-access-f9dnw") pod "2007e6df-db00-4c6d-8632-07e934613379" (UID: "2007e6df-db00-4c6d-8632-07e934613379"). InnerVolumeSpecName "kube-api-access-f9dnw". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:45:44 crc kubenswrapper[4631]: I1128 13:45:44.784395 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2007e6df-db00-4c6d-8632-07e934613379-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2007e6df-db00-4c6d-8632-07e934613379" (UID: "2007e6df-db00-4c6d-8632-07e934613379"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 13:45:44 crc kubenswrapper[4631]: I1128 13:45:44.831541 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f9dnw\" (UniqueName: \"kubernetes.io/projected/2007e6df-db00-4c6d-8632-07e934613379-kube-api-access-f9dnw\") on node \"crc\" DevicePath \"\"" Nov 28 13:45:44 crc kubenswrapper[4631]: I1128 13:45:44.831604 4631 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2007e6df-db00-4c6d-8632-07e934613379-utilities\") on node \"crc\" DevicePath \"\"" Nov 28 13:45:44 crc kubenswrapper[4631]: I1128 13:45:44.831674 4631 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2007e6df-db00-4c6d-8632-07e934613379-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 28 13:45:45 crc kubenswrapper[4631]: I1128 13:45:45.158622 4631 generic.go:334] "Generic (PLEG): container finished" podID="2007e6df-db00-4c6d-8632-07e934613379" containerID="3b51fef12d68144d1c3040a73fa2dff6e366dbc0f51b4a8326124a4ed6c5bbe3" exitCode=0 Nov 28 13:45:45 crc kubenswrapper[4631]: I1128 13:45:45.158709 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cnr7x" event={"ID":"2007e6df-db00-4c6d-8632-07e934613379","Type":"ContainerDied","Data":"3b51fef12d68144d1c3040a73fa2dff6e366dbc0f51b4a8326124a4ed6c5bbe3"} Nov 28 13:45:45 crc kubenswrapper[4631]: I1128 13:45:45.158727 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cnr7x" Nov 28 13:45:45 crc kubenswrapper[4631]: I1128 13:45:45.158945 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cnr7x" event={"ID":"2007e6df-db00-4c6d-8632-07e934613379","Type":"ContainerDied","Data":"6e5b039ad8f8de92a1a2b7cec88e81a387e1f64d53aa6669ef97f5a306f4e8f2"} Nov 28 13:45:45 crc kubenswrapper[4631]: I1128 13:45:45.158965 4631 scope.go:117] "RemoveContainer" containerID="3b51fef12d68144d1c3040a73fa2dff6e366dbc0f51b4a8326124a4ed6c5bbe3" Nov 28 13:45:45 crc kubenswrapper[4631]: I1128 13:45:45.199850 4631 scope.go:117] "RemoveContainer" containerID="14f060b32cc6c8c54088d4cd7a5131a4fafba27be925b150be4d2c314e98d98e" Nov 28 13:45:45 crc kubenswrapper[4631]: I1128 13:45:45.203431 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-cnr7x"] Nov 28 13:45:45 crc kubenswrapper[4631]: I1128 13:45:45.213001 4631 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-cnr7x"] Nov 28 13:45:45 crc kubenswrapper[4631]: I1128 13:45:45.225149 4631 scope.go:117] "RemoveContainer" containerID="4eb99330c33ffc765ef2d76e009172630e8f9c2d448cffa3a4a0f169b0118a09" Nov 28 13:45:45 crc kubenswrapper[4631]: I1128 13:45:45.269205 4631 scope.go:117] "RemoveContainer" containerID="3b51fef12d68144d1c3040a73fa2dff6e366dbc0f51b4a8326124a4ed6c5bbe3" Nov 28 13:45:45 crc kubenswrapper[4631]: E1128 13:45:45.269649 4631 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3b51fef12d68144d1c3040a73fa2dff6e366dbc0f51b4a8326124a4ed6c5bbe3\": container with ID starting with 3b51fef12d68144d1c3040a73fa2dff6e366dbc0f51b4a8326124a4ed6c5bbe3 not found: ID does not exist" containerID="3b51fef12d68144d1c3040a73fa2dff6e366dbc0f51b4a8326124a4ed6c5bbe3" Nov 28 13:45:45 crc kubenswrapper[4631]: I1128 13:45:45.269697 4631 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3b51fef12d68144d1c3040a73fa2dff6e366dbc0f51b4a8326124a4ed6c5bbe3"} err="failed to get container status \"3b51fef12d68144d1c3040a73fa2dff6e366dbc0f51b4a8326124a4ed6c5bbe3\": rpc error: code = NotFound desc = could not find container \"3b51fef12d68144d1c3040a73fa2dff6e366dbc0f51b4a8326124a4ed6c5bbe3\": container with ID starting with 3b51fef12d68144d1c3040a73fa2dff6e366dbc0f51b4a8326124a4ed6c5bbe3 not found: ID does not exist" Nov 28 13:45:45 crc kubenswrapper[4631]: I1128 13:45:45.269731 4631 scope.go:117] "RemoveContainer" containerID="14f060b32cc6c8c54088d4cd7a5131a4fafba27be925b150be4d2c314e98d98e" Nov 28 13:45:45 crc kubenswrapper[4631]: E1128 13:45:45.270537 4631 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"14f060b32cc6c8c54088d4cd7a5131a4fafba27be925b150be4d2c314e98d98e\": container with ID starting with 14f060b32cc6c8c54088d4cd7a5131a4fafba27be925b150be4d2c314e98d98e not found: ID does not exist" containerID="14f060b32cc6c8c54088d4cd7a5131a4fafba27be925b150be4d2c314e98d98e" Nov 28 13:45:45 crc kubenswrapper[4631]: I1128 13:45:45.270669 4631 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"14f060b32cc6c8c54088d4cd7a5131a4fafba27be925b150be4d2c314e98d98e"} err="failed to get container status \"14f060b32cc6c8c54088d4cd7a5131a4fafba27be925b150be4d2c314e98d98e\": rpc error: code = NotFound desc = could not find container \"14f060b32cc6c8c54088d4cd7a5131a4fafba27be925b150be4d2c314e98d98e\": container with ID starting with 14f060b32cc6c8c54088d4cd7a5131a4fafba27be925b150be4d2c314e98d98e not found: ID does not exist" Nov 28 13:45:45 crc kubenswrapper[4631]: I1128 13:45:45.270805 4631 scope.go:117] "RemoveContainer" containerID="4eb99330c33ffc765ef2d76e009172630e8f9c2d448cffa3a4a0f169b0118a09" Nov 28 13:45:45 crc kubenswrapper[4631]: E1128 13:45:45.271233 4631 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4eb99330c33ffc765ef2d76e009172630e8f9c2d448cffa3a4a0f169b0118a09\": container with ID starting with 4eb99330c33ffc765ef2d76e009172630e8f9c2d448cffa3a4a0f169b0118a09 not found: ID does not exist" containerID="4eb99330c33ffc765ef2d76e009172630e8f9c2d448cffa3a4a0f169b0118a09" Nov 28 13:45:45 crc kubenswrapper[4631]: I1128 13:45:45.271424 4631 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4eb99330c33ffc765ef2d76e009172630e8f9c2d448cffa3a4a0f169b0118a09"} err="failed to get container status \"4eb99330c33ffc765ef2d76e009172630e8f9c2d448cffa3a4a0f169b0118a09\": rpc error: code = NotFound desc = could not find container \"4eb99330c33ffc765ef2d76e009172630e8f9c2d448cffa3a4a0f169b0118a09\": container with ID starting with 4eb99330c33ffc765ef2d76e009172630e8f9c2d448cffa3a4a0f169b0118a09 not found: ID does not exist" Nov 28 13:45:45 crc kubenswrapper[4631]: I1128 13:45:45.525711 4631 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2007e6df-db00-4c6d-8632-07e934613379" path="/var/lib/kubelet/pods/2007e6df-db00-4c6d-8632-07e934613379/volumes" Nov 28 13:46:10 crc kubenswrapper[4631]: I1128 13:46:10.057499 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-68dgq"] Nov 28 13:46:10 crc kubenswrapper[4631]: I1128 13:46:10.068070 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-d5ea-account-create-update-cw75f"] Nov 28 13:46:10 crc kubenswrapper[4631]: I1128 13:46:10.080766 4631 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-68dgq"] Nov 28 13:46:10 crc kubenswrapper[4631]: I1128 13:46:10.093709 4631 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-d5ea-account-create-update-cw75f"] Nov 28 13:46:11 crc kubenswrapper[4631]: I1128 13:46:11.042673 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-9m2n2"] Nov 28 13:46:11 crc kubenswrapper[4631]: I1128 13:46:11.058253 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-addf-account-create-update-px8w7"] Nov 28 13:46:11 crc kubenswrapper[4631]: I1128 13:46:11.072378 4631 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-9m2n2"] Nov 28 13:46:11 crc kubenswrapper[4631]: I1128 13:46:11.082628 4631 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-addf-account-create-update-px8w7"] Nov 28 13:46:11 crc kubenswrapper[4631]: I1128 13:46:11.526761 4631 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="07e8a124-1828-4a61-9b39-e1c8cf4cb2cc" path="/var/lib/kubelet/pods/07e8a124-1828-4a61-9b39-e1c8cf4cb2cc/volumes" Nov 28 13:46:11 crc kubenswrapper[4631]: I1128 13:46:11.528318 4631 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="122f3637-af63-467a-9e98-6b5f48570d9d" path="/var/lib/kubelet/pods/122f3637-af63-467a-9e98-6b5f48570d9d/volumes" Nov 28 13:46:11 crc kubenswrapper[4631]: I1128 13:46:11.530354 4631 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3edb21b5-935d-4b42-bc88-d590b7fb0e3d" path="/var/lib/kubelet/pods/3edb21b5-935d-4b42-bc88-d590b7fb0e3d/volumes" Nov 28 13:46:11 crc kubenswrapper[4631]: I1128 13:46:11.531293 4631 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dd84f393-e43d-45e8-8760-7e0643f7c30e" path="/var/lib/kubelet/pods/dd84f393-e43d-45e8-8760-7e0643f7c30e/volumes" Nov 28 13:46:15 crc kubenswrapper[4631]: I1128 13:46:15.040168 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-gn4cj"] Nov 28 13:46:15 crc kubenswrapper[4631]: I1128 13:46:15.047982 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-771b-account-create-update-wzs8f"] Nov 28 13:46:15 crc kubenswrapper[4631]: I1128 13:46:15.056644 4631 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-gn4cj"] Nov 28 13:46:15 crc kubenswrapper[4631]: I1128 13:46:15.066429 4631 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-771b-account-create-update-wzs8f"] Nov 28 13:46:15 crc kubenswrapper[4631]: I1128 13:46:15.528973 4631 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20f86721-2a44-4cca-976b-9d3132583903" path="/var/lib/kubelet/pods/20f86721-2a44-4cca-976b-9d3132583903/volumes" Nov 28 13:46:15 crc kubenswrapper[4631]: I1128 13:46:15.530627 4631 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c20311ad-58ca-4d2a-b19d-3b5291464558" path="/var/lib/kubelet/pods/c20311ad-58ca-4d2a-b19d-3b5291464558/volumes" Nov 28 13:46:23 crc kubenswrapper[4631]: I1128 13:46:23.662726 4631 patch_prober.go:28] interesting pod/controller-manager-66d87f89f6-5g9jj container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.64:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Nov 28 13:46:23 crc kubenswrapper[4631]: I1128 13:46:23.663441 4631 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-66d87f89f6-5g9jj" podUID="a2c8105b-a4d6-487c-af19-668a70b2964e" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.64:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Nov 28 13:46:23 crc kubenswrapper[4631]: I1128 13:46:23.678510 4631 patch_prober.go:28] interesting pod/controller-manager-66d87f89f6-5g9jj container/controller-manager namespace/openshift-controller-manager: Liveness probe status=failure output="Get \"https://10.217.0.64:8443/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Nov 28 13:46:23 crc kubenswrapper[4631]: I1128 13:46:23.678586 4631 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-controller-manager/controller-manager-66d87f89f6-5g9jj" podUID="a2c8105b-a4d6-487c-af19-668a70b2964e" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.64:8443/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 28 13:46:47 crc kubenswrapper[4631]: I1128 13:46:47.048363 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-g7926"] Nov 28 13:46:47 crc kubenswrapper[4631]: I1128 13:46:47.057683 4631 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-g7926"] Nov 28 13:46:47 crc kubenswrapper[4631]: I1128 13:46:47.531502 4631 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="387c22fd-1a37-4031-a53c-1a5d9e322bed" path="/var/lib/kubelet/pods/387c22fd-1a37-4031-a53c-1a5d9e322bed/volumes" Nov 28 13:46:49 crc kubenswrapper[4631]: I1128 13:46:49.634482 4631 patch_prober.go:28] interesting pod/machine-config-daemon-47sxw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 28 13:46:49 crc kubenswrapper[4631]: I1128 13:46:49.634767 4631 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" podUID="8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 28 13:47:00 crc kubenswrapper[4631]: I1128 13:47:00.032708 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-hplqp"] Nov 28 13:47:00 crc kubenswrapper[4631]: I1128 13:47:00.041557 4631 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-hplqp"] Nov 28 13:47:01 crc kubenswrapper[4631]: I1128 13:47:01.073468 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-99b5-account-create-update-67zx8"] Nov 28 13:47:01 crc kubenswrapper[4631]: I1128 13:47:01.089247 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-c506-account-create-update-nr5s5"] Nov 28 13:47:01 crc kubenswrapper[4631]: I1128 13:47:01.098339 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-m4zmz"] Nov 28 13:47:01 crc kubenswrapper[4631]: I1128 13:47:01.107232 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-8a8c-account-create-update-lskqg"] Nov 28 13:47:01 crc kubenswrapper[4631]: I1128 13:47:01.115570 4631 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-99b5-account-create-update-67zx8"] Nov 28 13:47:01 crc kubenswrapper[4631]: I1128 13:47:01.123823 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-wrd26"] Nov 28 13:47:01 crc kubenswrapper[4631]: I1128 13:47:01.131717 4631 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-8a8c-account-create-update-lskqg"] Nov 28 13:47:01 crc kubenswrapper[4631]: I1128 13:47:01.140011 4631 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-m4zmz"] Nov 28 13:47:01 crc kubenswrapper[4631]: I1128 13:47:01.147599 4631 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-wrd26"] Nov 28 13:47:01 crc kubenswrapper[4631]: I1128 13:47:01.155087 4631 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-c506-account-create-update-nr5s5"] Nov 28 13:47:01 crc kubenswrapper[4631]: I1128 13:47:01.528658 4631 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="127100c3-8f48-4ddf-a289-f08f3cdf0d8a" path="/var/lib/kubelet/pods/127100c3-8f48-4ddf-a289-f08f3cdf0d8a/volumes" Nov 28 13:47:01 crc kubenswrapper[4631]: I1128 13:47:01.530208 4631 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="95d067ec-b8ae-4286-a35f-eb431d025ac9" path="/var/lib/kubelet/pods/95d067ec-b8ae-4286-a35f-eb431d025ac9/volumes" Nov 28 13:47:01 crc kubenswrapper[4631]: I1128 13:47:01.531732 4631 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="98564bcf-8a3f-4324-af81-bafdc2ed2934" path="/var/lib/kubelet/pods/98564bcf-8a3f-4324-af81-bafdc2ed2934/volumes" Nov 28 13:47:01 crc kubenswrapper[4631]: I1128 13:47:01.533224 4631 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cee396c9-93ab-45dd-82d0-9641e985644a" path="/var/lib/kubelet/pods/cee396c9-93ab-45dd-82d0-9641e985644a/volumes" Nov 28 13:47:01 crc kubenswrapper[4631]: I1128 13:47:01.535091 4631 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e8fd4db1-5308-414f-9eac-ddef2c41867a" path="/var/lib/kubelet/pods/e8fd4db1-5308-414f-9eac-ddef2c41867a/volumes" Nov 28 13:47:01 crc kubenswrapper[4631]: I1128 13:47:01.536063 4631 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ecb43d0a-167c-42a9-9ea9-bc303f5890b8" path="/var/lib/kubelet/pods/ecb43d0a-167c-42a9-9ea9-bc303f5890b8/volumes" Nov 28 13:47:02 crc kubenswrapper[4631]: I1128 13:47:02.889707 4631 scope.go:117] "RemoveContainer" containerID="b03a3cbbf0292dae3632dfc0115e24e09e6ef8227d086f5bfbb211d9d09c3046" Nov 28 13:47:02 crc kubenswrapper[4631]: I1128 13:47:02.934578 4631 scope.go:117] "RemoveContainer" containerID="bf7fdb2038d321e6b67de7b7a577f31a3431fe807da206dd011898c5be98301b" Nov 28 13:47:03 crc kubenswrapper[4631]: I1128 13:47:03.001240 4631 scope.go:117] "RemoveContainer" containerID="459e0a9725c8f0753466ab2f643456b6345e1e52cf742bce27b3e8c960230948" Nov 28 13:47:03 crc kubenswrapper[4631]: I1128 13:47:03.044416 4631 scope.go:117] "RemoveContainer" containerID="9d45d82ad9a1cab47bbcabf7ac8ec4a6f275787e84b79703b999c96b65a918c2" Nov 28 13:47:03 crc kubenswrapper[4631]: I1128 13:47:03.085324 4631 scope.go:117] "RemoveContainer" containerID="b55a0c4fae79fdf4bb59d9a51cb2dc17f21eefaee959939aee4a11adfc266b4e" Nov 28 13:47:03 crc kubenswrapper[4631]: I1128 13:47:03.114253 4631 scope.go:117] "RemoveContainer" containerID="a3f65d289623bf199c36739c4c862db70137638bad3d4278065fec2e5206cf1d" Nov 28 13:47:03 crc kubenswrapper[4631]: I1128 13:47:03.172781 4631 scope.go:117] "RemoveContainer" containerID="1cac3f2d3153a738c4382184487f445a93cf68bcfaf941c66086e3a38cf0dcf8" Nov 28 13:47:03 crc kubenswrapper[4631]: I1128 13:47:03.231346 4631 scope.go:117] "RemoveContainer" containerID="354420ec5de1c2a4fa0ce766623d0472250845914103d054c4b82bac7368edb8" Nov 28 13:47:03 crc kubenswrapper[4631]: I1128 13:47:03.258485 4631 scope.go:117] "RemoveContainer" containerID="fc8711f77c49ff38e295ea35860dbf498a2f8e950091ebc9ba461288c475526d" Nov 28 13:47:03 crc kubenswrapper[4631]: I1128 13:47:03.281428 4631 scope.go:117] "RemoveContainer" containerID="f1cab3e61e6936a06f7682175f6c6e2f4d0dc741337beca101fa628f19cce22d" Nov 28 13:47:03 crc kubenswrapper[4631]: I1128 13:47:03.305945 4631 scope.go:117] "RemoveContainer" containerID="0176fdd3cef121c949afaf075e55b66120e88ef916c4a388add9afe002dcac8d" Nov 28 13:47:03 crc kubenswrapper[4631]: I1128 13:47:03.346891 4631 scope.go:117] "RemoveContainer" containerID="94079e09b20e24fe68c2194a6d9eeb2494fdc13708f300c19079b06192e7eda9" Nov 28 13:47:03 crc kubenswrapper[4631]: I1128 13:47:03.377500 4631 scope.go:117] "RemoveContainer" containerID="a1646b6221d6723f5f5f2cc3c750881618fc6f98c65326ef8d092470b7a48482" Nov 28 13:47:03 crc kubenswrapper[4631]: I1128 13:47:03.440634 4631 scope.go:117] "RemoveContainer" containerID="dc82724e514dbd2f2f7f5b8d2e7c30e436f62a081faae75f298fdffee1ee9332" Nov 28 13:47:03 crc kubenswrapper[4631]: I1128 13:47:03.463520 4631 scope.go:117] "RemoveContainer" containerID="482ce3c6826106b00b097fc029b492fb977fb23e7ade8a08dfaa1ece091c188e" Nov 28 13:47:13 crc kubenswrapper[4631]: I1128 13:47:13.044413 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-xs457"] Nov 28 13:47:13 crc kubenswrapper[4631]: I1128 13:47:13.057458 4631 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-xs457"] Nov 28 13:47:13 crc kubenswrapper[4631]: I1128 13:47:13.528782 4631 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6844d09-9bb8-44a7-bf9a-eca18ce5c1c8" path="/var/lib/kubelet/pods/b6844d09-9bb8-44a7-bf9a-eca18ce5c1c8/volumes" Nov 28 13:47:15 crc kubenswrapper[4631]: I1128 13:47:15.519750 4631 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/heat-operator-controller-manager-5b77f656f-8pvq9" podUID="9cc92c5c-285e-4604-bc60-7637938a78c8" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.71:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 28 13:47:17 crc kubenswrapper[4631]: I1128 13:47:17.227429 4631 generic.go:334] "Generic (PLEG): container finished" podID="44ae011d-6420-4400-9c7c-21463e3c82b1" containerID="812c7c40987259fd36752aa82b116fec7562779d3e5df6f7f438155ff00d0820" exitCode=0 Nov 28 13:47:17 crc kubenswrapper[4631]: I1128 13:47:17.227600 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-xz7jm" event={"ID":"44ae011d-6420-4400-9c7c-21463e3c82b1","Type":"ContainerDied","Data":"812c7c40987259fd36752aa82b116fec7562779d3e5df6f7f438155ff00d0820"} Nov 28 13:47:18 crc kubenswrapper[4631]: I1128 13:47:18.652123 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-xz7jm" Nov 28 13:47:18 crc kubenswrapper[4631]: I1128 13:47:18.779132 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/44ae011d-6420-4400-9c7c-21463e3c82b1-bootstrap-combined-ca-bundle\") pod \"44ae011d-6420-4400-9c7c-21463e3c82b1\" (UID: \"44ae011d-6420-4400-9c7c-21463e3c82b1\") " Nov 28 13:47:18 crc kubenswrapper[4631]: I1128 13:47:18.779214 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/44ae011d-6420-4400-9c7c-21463e3c82b1-ssh-key\") pod \"44ae011d-6420-4400-9c7c-21463e3c82b1\" (UID: \"44ae011d-6420-4400-9c7c-21463e3c82b1\") " Nov 28 13:47:18 crc kubenswrapper[4631]: I1128 13:47:18.779382 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/44ae011d-6420-4400-9c7c-21463e3c82b1-inventory\") pod \"44ae011d-6420-4400-9c7c-21463e3c82b1\" (UID: \"44ae011d-6420-4400-9c7c-21463e3c82b1\") " Nov 28 13:47:18 crc kubenswrapper[4631]: I1128 13:47:18.779425 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n5l9z\" (UniqueName: \"kubernetes.io/projected/44ae011d-6420-4400-9c7c-21463e3c82b1-kube-api-access-n5l9z\") pod \"44ae011d-6420-4400-9c7c-21463e3c82b1\" (UID: \"44ae011d-6420-4400-9c7c-21463e3c82b1\") " Nov 28 13:47:18 crc kubenswrapper[4631]: I1128 13:47:18.785833 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44ae011d-6420-4400-9c7c-21463e3c82b1-kube-api-access-n5l9z" (OuterVolumeSpecName: "kube-api-access-n5l9z") pod "44ae011d-6420-4400-9c7c-21463e3c82b1" (UID: "44ae011d-6420-4400-9c7c-21463e3c82b1"). InnerVolumeSpecName "kube-api-access-n5l9z". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:47:18 crc kubenswrapper[4631]: I1128 13:47:18.786378 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/44ae011d-6420-4400-9c7c-21463e3c82b1-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "44ae011d-6420-4400-9c7c-21463e3c82b1" (UID: "44ae011d-6420-4400-9c7c-21463e3c82b1"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:47:18 crc kubenswrapper[4631]: I1128 13:47:18.812450 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/44ae011d-6420-4400-9c7c-21463e3c82b1-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "44ae011d-6420-4400-9c7c-21463e3c82b1" (UID: "44ae011d-6420-4400-9c7c-21463e3c82b1"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:47:18 crc kubenswrapper[4631]: I1128 13:47:18.812548 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/44ae011d-6420-4400-9c7c-21463e3c82b1-inventory" (OuterVolumeSpecName: "inventory") pod "44ae011d-6420-4400-9c7c-21463e3c82b1" (UID: "44ae011d-6420-4400-9c7c-21463e3c82b1"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:47:18 crc kubenswrapper[4631]: I1128 13:47:18.881669 4631 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/44ae011d-6420-4400-9c7c-21463e3c82b1-inventory\") on node \"crc\" DevicePath \"\"" Nov 28 13:47:18 crc kubenswrapper[4631]: I1128 13:47:18.881708 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n5l9z\" (UniqueName: \"kubernetes.io/projected/44ae011d-6420-4400-9c7c-21463e3c82b1-kube-api-access-n5l9z\") on node \"crc\" DevicePath \"\"" Nov 28 13:47:18 crc kubenswrapper[4631]: I1128 13:47:18.881720 4631 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/44ae011d-6420-4400-9c7c-21463e3c82b1-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 13:47:18 crc kubenswrapper[4631]: I1128 13:47:18.881729 4631 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/44ae011d-6420-4400-9c7c-21463e3c82b1-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 28 13:47:19 crc kubenswrapper[4631]: I1128 13:47:19.246918 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-xz7jm" event={"ID":"44ae011d-6420-4400-9c7c-21463e3c82b1","Type":"ContainerDied","Data":"3bafea57cfda07b8698ecbc6b31bcd09cb79c7abc3c7216606d86e4d223c3053"} Nov 28 13:47:19 crc kubenswrapper[4631]: I1128 13:47:19.247234 4631 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3bafea57cfda07b8698ecbc6b31bcd09cb79c7abc3c7216606d86e4d223c3053" Nov 28 13:47:19 crc kubenswrapper[4631]: I1128 13:47:19.247437 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-xz7jm" Nov 28 13:47:19 crc kubenswrapper[4631]: I1128 13:47:19.357365 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-2j6xr"] Nov 28 13:47:19 crc kubenswrapper[4631]: E1128 13:47:19.357829 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="44ae011d-6420-4400-9c7c-21463e3c82b1" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Nov 28 13:47:19 crc kubenswrapper[4631]: I1128 13:47:19.357847 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="44ae011d-6420-4400-9c7c-21463e3c82b1" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Nov 28 13:47:19 crc kubenswrapper[4631]: E1128 13:47:19.357866 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1f7a85b3-c1a6-4d7b-af11-8ed258e20818" containerName="extract-utilities" Nov 28 13:47:19 crc kubenswrapper[4631]: I1128 13:47:19.357873 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="1f7a85b3-c1a6-4d7b-af11-8ed258e20818" containerName="extract-utilities" Nov 28 13:47:19 crc kubenswrapper[4631]: E1128 13:47:19.357888 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2007e6df-db00-4c6d-8632-07e934613379" containerName="registry-server" Nov 28 13:47:19 crc kubenswrapper[4631]: I1128 13:47:19.357895 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="2007e6df-db00-4c6d-8632-07e934613379" containerName="registry-server" Nov 28 13:47:19 crc kubenswrapper[4631]: E1128 13:47:19.357909 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1f7a85b3-c1a6-4d7b-af11-8ed258e20818" containerName="extract-content" Nov 28 13:47:19 crc kubenswrapper[4631]: I1128 13:47:19.357915 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="1f7a85b3-c1a6-4d7b-af11-8ed258e20818" containerName="extract-content" Nov 28 13:47:19 crc kubenswrapper[4631]: E1128 13:47:19.357928 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2007e6df-db00-4c6d-8632-07e934613379" containerName="extract-utilities" Nov 28 13:47:19 crc kubenswrapper[4631]: I1128 13:47:19.357934 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="2007e6df-db00-4c6d-8632-07e934613379" containerName="extract-utilities" Nov 28 13:47:19 crc kubenswrapper[4631]: E1128 13:47:19.357958 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1f7a85b3-c1a6-4d7b-af11-8ed258e20818" containerName="registry-server" Nov 28 13:47:19 crc kubenswrapper[4631]: I1128 13:47:19.357964 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="1f7a85b3-c1a6-4d7b-af11-8ed258e20818" containerName="registry-server" Nov 28 13:47:19 crc kubenswrapper[4631]: E1128 13:47:19.357977 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2007e6df-db00-4c6d-8632-07e934613379" containerName="extract-content" Nov 28 13:47:19 crc kubenswrapper[4631]: I1128 13:47:19.357982 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="2007e6df-db00-4c6d-8632-07e934613379" containerName="extract-content" Nov 28 13:47:19 crc kubenswrapper[4631]: I1128 13:47:19.358152 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="2007e6df-db00-4c6d-8632-07e934613379" containerName="registry-server" Nov 28 13:47:19 crc kubenswrapper[4631]: I1128 13:47:19.358174 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="44ae011d-6420-4400-9c7c-21463e3c82b1" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Nov 28 13:47:19 crc kubenswrapper[4631]: I1128 13:47:19.358189 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="1f7a85b3-c1a6-4d7b-af11-8ed258e20818" containerName="registry-server" Nov 28 13:47:19 crc kubenswrapper[4631]: I1128 13:47:19.359091 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-2j6xr" Nov 28 13:47:19 crc kubenswrapper[4631]: I1128 13:47:19.364427 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 28 13:47:19 crc kubenswrapper[4631]: I1128 13:47:19.364436 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 28 13:47:19 crc kubenswrapper[4631]: I1128 13:47:19.364570 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 28 13:47:19 crc kubenswrapper[4631]: I1128 13:47:19.374065 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-2j6xr"] Nov 28 13:47:19 crc kubenswrapper[4631]: I1128 13:47:19.376157 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-vp4ns" Nov 28 13:47:19 crc kubenswrapper[4631]: E1128 13:47:19.458154 4631 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod44ae011d_6420_4400_9c7c_21463e3c82b1.slice\": RecentStats: unable to find data in memory cache]" Nov 28 13:47:19 crc kubenswrapper[4631]: I1128 13:47:19.495733 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2nws7\" (UniqueName: \"kubernetes.io/projected/2a23172b-a4cc-41e3-9464-2b3e7f7a51f8-kube-api-access-2nws7\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-2j6xr\" (UID: \"2a23172b-a4cc-41e3-9464-2b3e7f7a51f8\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-2j6xr" Nov 28 13:47:19 crc kubenswrapper[4631]: I1128 13:47:19.495836 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2a23172b-a4cc-41e3-9464-2b3e7f7a51f8-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-2j6xr\" (UID: \"2a23172b-a4cc-41e3-9464-2b3e7f7a51f8\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-2j6xr" Nov 28 13:47:19 crc kubenswrapper[4631]: I1128 13:47:19.495856 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2a23172b-a4cc-41e3-9464-2b3e7f7a51f8-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-2j6xr\" (UID: \"2a23172b-a4cc-41e3-9464-2b3e7f7a51f8\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-2j6xr" Nov 28 13:47:19 crc kubenswrapper[4631]: I1128 13:47:19.597825 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2a23172b-a4cc-41e3-9464-2b3e7f7a51f8-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-2j6xr\" (UID: \"2a23172b-a4cc-41e3-9464-2b3e7f7a51f8\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-2j6xr" Nov 28 13:47:19 crc kubenswrapper[4631]: I1128 13:47:19.597869 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2a23172b-a4cc-41e3-9464-2b3e7f7a51f8-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-2j6xr\" (UID: \"2a23172b-a4cc-41e3-9464-2b3e7f7a51f8\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-2j6xr" Nov 28 13:47:19 crc kubenswrapper[4631]: I1128 13:47:19.597977 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2nws7\" (UniqueName: \"kubernetes.io/projected/2a23172b-a4cc-41e3-9464-2b3e7f7a51f8-kube-api-access-2nws7\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-2j6xr\" (UID: \"2a23172b-a4cc-41e3-9464-2b3e7f7a51f8\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-2j6xr" Nov 28 13:47:19 crc kubenswrapper[4631]: I1128 13:47:19.604678 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2a23172b-a4cc-41e3-9464-2b3e7f7a51f8-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-2j6xr\" (UID: \"2a23172b-a4cc-41e3-9464-2b3e7f7a51f8\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-2j6xr" Nov 28 13:47:19 crc kubenswrapper[4631]: I1128 13:47:19.614040 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2a23172b-a4cc-41e3-9464-2b3e7f7a51f8-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-2j6xr\" (UID: \"2a23172b-a4cc-41e3-9464-2b3e7f7a51f8\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-2j6xr" Nov 28 13:47:19 crc kubenswrapper[4631]: I1128 13:47:19.616391 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2nws7\" (UniqueName: \"kubernetes.io/projected/2a23172b-a4cc-41e3-9464-2b3e7f7a51f8-kube-api-access-2nws7\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-2j6xr\" (UID: \"2a23172b-a4cc-41e3-9464-2b3e7f7a51f8\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-2j6xr" Nov 28 13:47:19 crc kubenswrapper[4631]: I1128 13:47:19.634783 4631 patch_prober.go:28] interesting pod/machine-config-daemon-47sxw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 28 13:47:19 crc kubenswrapper[4631]: I1128 13:47:19.634849 4631 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" podUID="8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 28 13:47:19 crc kubenswrapper[4631]: I1128 13:47:19.685865 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-2j6xr" Nov 28 13:47:20 crc kubenswrapper[4631]: I1128 13:47:20.274534 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-2j6xr"] Nov 28 13:47:21 crc kubenswrapper[4631]: I1128 13:47:21.265540 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-2j6xr" event={"ID":"2a23172b-a4cc-41e3-9464-2b3e7f7a51f8","Type":"ContainerStarted","Data":"22d474501dd70646b16952d731656831447f79fa2f853ce29b6a9f23669540a7"} Nov 28 13:47:22 crc kubenswrapper[4631]: I1128 13:47:22.276745 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-2j6xr" event={"ID":"2a23172b-a4cc-41e3-9464-2b3e7f7a51f8","Type":"ContainerStarted","Data":"fdc1e876e90151475f0b0ad56e1a982556b283197b4f4c4460e39a4868ea145f"} Nov 28 13:47:22 crc kubenswrapper[4631]: I1128 13:47:22.296057 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-2j6xr" podStartSLOduration=2.554716146 podStartE2EDuration="3.296031906s" podCreationTimestamp="2025-11-28 13:47:19 +0000 UTC" firstStartedPulling="2025-11-28 13:47:20.281642201 +0000 UTC m=+1597.088945545" lastFinishedPulling="2025-11-28 13:47:21.022957961 +0000 UTC m=+1597.830261305" observedRunningTime="2025-11-28 13:47:22.295028402 +0000 UTC m=+1599.102331746" watchObservedRunningTime="2025-11-28 13:47:22.296031906 +0000 UTC m=+1599.103335250" Nov 28 13:47:49 crc kubenswrapper[4631]: I1128 13:47:49.634637 4631 patch_prober.go:28] interesting pod/machine-config-daemon-47sxw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 28 13:47:49 crc kubenswrapper[4631]: I1128 13:47:49.635389 4631 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" podUID="8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 28 13:47:49 crc kubenswrapper[4631]: I1128 13:47:49.635460 4631 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" Nov 28 13:47:49 crc kubenswrapper[4631]: I1128 13:47:49.636395 4631 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"54d38280b4f1c123c61719ed71e612b2bf974f3bf3d8ff13fe19d70fa5ca5198"} pod="openshift-machine-config-operator/machine-config-daemon-47sxw" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 28 13:47:49 crc kubenswrapper[4631]: I1128 13:47:49.636459 4631 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" podUID="8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1" containerName="machine-config-daemon" containerID="cri-o://54d38280b4f1c123c61719ed71e612b2bf974f3bf3d8ff13fe19d70fa5ca5198" gracePeriod=600 Nov 28 13:47:50 crc kubenswrapper[4631]: E1128 13:47:50.275148 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-47sxw_openshift-machine-config-operator(8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1)\"" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" podUID="8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1" Nov 28 13:47:50 crc kubenswrapper[4631]: I1128 13:47:50.545022 4631 generic.go:334] "Generic (PLEG): container finished" podID="8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1" containerID="54d38280b4f1c123c61719ed71e612b2bf974f3bf3d8ff13fe19d70fa5ca5198" exitCode=0 Nov 28 13:47:50 crc kubenswrapper[4631]: I1128 13:47:50.545064 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" event={"ID":"8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1","Type":"ContainerDied","Data":"54d38280b4f1c123c61719ed71e612b2bf974f3bf3d8ff13fe19d70fa5ca5198"} Nov 28 13:47:50 crc kubenswrapper[4631]: I1128 13:47:50.545097 4631 scope.go:117] "RemoveContainer" containerID="05e6feabd7664ef6344f7bae4443e46fb2b0ab0814dcbdcb00623d5d82759c9c" Nov 28 13:47:50 crc kubenswrapper[4631]: I1128 13:47:50.545713 4631 scope.go:117] "RemoveContainer" containerID="54d38280b4f1c123c61719ed71e612b2bf974f3bf3d8ff13fe19d70fa5ca5198" Nov 28 13:47:50 crc kubenswrapper[4631]: E1128 13:47:50.545951 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-47sxw_openshift-machine-config-operator(8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1)\"" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" podUID="8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1" Nov 28 13:47:52 crc kubenswrapper[4631]: I1128 13:47:52.048199 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-6lwcj"] Nov 28 13:47:52 crc kubenswrapper[4631]: I1128 13:47:52.056165 4631 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-6lwcj"] Nov 28 13:47:53 crc kubenswrapper[4631]: I1128 13:47:53.526414 4631 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="268feb7e-4677-4c47-b837-e04bab4b8010" path="/var/lib/kubelet/pods/268feb7e-4677-4c47-b837-e04bab4b8010/volumes" Nov 28 13:48:02 crc kubenswrapper[4631]: I1128 13:48:02.514430 4631 scope.go:117] "RemoveContainer" containerID="54d38280b4f1c123c61719ed71e612b2bf974f3bf3d8ff13fe19d70fa5ca5198" Nov 28 13:48:02 crc kubenswrapper[4631]: E1128 13:48:02.515309 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-47sxw_openshift-machine-config-operator(8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1)\"" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" podUID="8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1" Nov 28 13:48:03 crc kubenswrapper[4631]: I1128 13:48:03.719916 4631 scope.go:117] "RemoveContainer" containerID="8b91851c1160b6f6360af453ef82165da62b7b53a25c4fb8390d0871f3df8f25" Nov 28 13:48:03 crc kubenswrapper[4631]: I1128 13:48:03.748080 4631 scope.go:117] "RemoveContainer" containerID="22d98a24eacf769b7c3570fb24c8bf34f199e9c081e79a87c78dffae0d919cc9" Nov 28 13:48:14 crc kubenswrapper[4631]: I1128 13:48:14.050889 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-dcm9x"] Nov 28 13:48:14 crc kubenswrapper[4631]: I1128 13:48:14.064895 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-xhft6"] Nov 28 13:48:14 crc kubenswrapper[4631]: I1128 13:48:14.076579 4631 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-xhft6"] Nov 28 13:48:14 crc kubenswrapper[4631]: I1128 13:48:14.087471 4631 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-dcm9x"] Nov 28 13:48:14 crc kubenswrapper[4631]: I1128 13:48:14.513924 4631 scope.go:117] "RemoveContainer" containerID="54d38280b4f1c123c61719ed71e612b2bf974f3bf3d8ff13fe19d70fa5ca5198" Nov 28 13:48:14 crc kubenswrapper[4631]: E1128 13:48:14.514253 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-47sxw_openshift-machine-config-operator(8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1)\"" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" podUID="8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1" Nov 28 13:48:15 crc kubenswrapper[4631]: I1128 13:48:15.524876 4631 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="05e73b90-be19-4f14-9824-a5c27361adff" path="/var/lib/kubelet/pods/05e73b90-be19-4f14-9824-a5c27361adff/volumes" Nov 28 13:48:15 crc kubenswrapper[4631]: I1128 13:48:15.525719 4631 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="33e9838e-f616-4504-90d5-799de265be04" path="/var/lib/kubelet/pods/33e9838e-f616-4504-90d5-799de265be04/volumes" Nov 28 13:48:26 crc kubenswrapper[4631]: I1128 13:48:26.513786 4631 scope.go:117] "RemoveContainer" containerID="54d38280b4f1c123c61719ed71e612b2bf974f3bf3d8ff13fe19d70fa5ca5198" Nov 28 13:48:26 crc kubenswrapper[4631]: E1128 13:48:26.514698 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-47sxw_openshift-machine-config-operator(8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1)\"" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" podUID="8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1" Nov 28 13:48:28 crc kubenswrapper[4631]: I1128 13:48:28.033219 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-gprj7"] Nov 28 13:48:28 crc kubenswrapper[4631]: I1128 13:48:28.046888 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-z6tgb"] Nov 28 13:48:28 crc kubenswrapper[4631]: I1128 13:48:28.060557 4631 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-gprj7"] Nov 28 13:48:28 crc kubenswrapper[4631]: I1128 13:48:28.069325 4631 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-z6tgb"] Nov 28 13:48:29 crc kubenswrapper[4631]: I1128 13:48:29.526882 4631 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3050013d-2db4-4801-8626-1fb6398b714e" path="/var/lib/kubelet/pods/3050013d-2db4-4801-8626-1fb6398b714e/volumes" Nov 28 13:48:29 crc kubenswrapper[4631]: I1128 13:48:29.528549 4631 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="af357d67-237c-468a-81ac-990964e8e8b2" path="/var/lib/kubelet/pods/af357d67-237c-468a-81ac-990964e8e8b2/volumes" Nov 28 13:48:38 crc kubenswrapper[4631]: I1128 13:48:38.513276 4631 scope.go:117] "RemoveContainer" containerID="54d38280b4f1c123c61719ed71e612b2bf974f3bf3d8ff13fe19d70fa5ca5198" Nov 28 13:48:38 crc kubenswrapper[4631]: E1128 13:48:38.514249 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-47sxw_openshift-machine-config-operator(8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1)\"" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" podUID="8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1" Nov 28 13:48:50 crc kubenswrapper[4631]: I1128 13:48:50.513127 4631 scope.go:117] "RemoveContainer" containerID="54d38280b4f1c123c61719ed71e612b2bf974f3bf3d8ff13fe19d70fa5ca5198" Nov 28 13:48:50 crc kubenswrapper[4631]: E1128 13:48:50.513933 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-47sxw_openshift-machine-config-operator(8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1)\"" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" podUID="8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1" Nov 28 13:49:03 crc kubenswrapper[4631]: I1128 13:49:03.853500 4631 scope.go:117] "RemoveContainer" containerID="0aaa7f880cc9f4f5eb075f9907b370aaa7455dc78316ee92682613ee25465639" Nov 28 13:49:03 crc kubenswrapper[4631]: I1128 13:49:03.948386 4631 scope.go:117] "RemoveContainer" containerID="4cb68320336ce8ebb3c54401b0bf63ad92137da24f81bdcca4fff5bcff77299f" Nov 28 13:49:03 crc kubenswrapper[4631]: I1128 13:49:03.999641 4631 scope.go:117] "RemoveContainer" containerID="e71efaf4f5015e3c6b71bcb5af4c2d9dab05f04b0454ced34b4bd97d9f353272" Nov 28 13:49:04 crc kubenswrapper[4631]: I1128 13:49:04.040255 4631 scope.go:117] "RemoveContainer" containerID="8659b5cf744dfcf5eb7c3313fb4013f0f03ffb7680a0047f33139bfff995c446" Nov 28 13:49:04 crc kubenswrapper[4631]: I1128 13:49:04.513106 4631 scope.go:117] "RemoveContainer" containerID="54d38280b4f1c123c61719ed71e612b2bf974f3bf3d8ff13fe19d70fa5ca5198" Nov 28 13:49:04 crc kubenswrapper[4631]: E1128 13:49:04.513591 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-47sxw_openshift-machine-config-operator(8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1)\"" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" podUID="8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1" Nov 28 13:49:12 crc kubenswrapper[4631]: I1128 13:49:12.271061 4631 generic.go:334] "Generic (PLEG): container finished" podID="2a23172b-a4cc-41e3-9464-2b3e7f7a51f8" containerID="fdc1e876e90151475f0b0ad56e1a982556b283197b4f4c4460e39a4868ea145f" exitCode=0 Nov 28 13:49:12 crc kubenswrapper[4631]: I1128 13:49:12.271162 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-2j6xr" event={"ID":"2a23172b-a4cc-41e3-9464-2b3e7f7a51f8","Type":"ContainerDied","Data":"fdc1e876e90151475f0b0ad56e1a982556b283197b4f4c4460e39a4868ea145f"} Nov 28 13:49:13 crc kubenswrapper[4631]: I1128 13:49:13.694445 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-2j6xr" Nov 28 13:49:13 crc kubenswrapper[4631]: I1128 13:49:13.785781 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2a23172b-a4cc-41e3-9464-2b3e7f7a51f8-ssh-key\") pod \"2a23172b-a4cc-41e3-9464-2b3e7f7a51f8\" (UID: \"2a23172b-a4cc-41e3-9464-2b3e7f7a51f8\") " Nov 28 13:49:13 crc kubenswrapper[4631]: I1128 13:49:13.785924 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2nws7\" (UniqueName: \"kubernetes.io/projected/2a23172b-a4cc-41e3-9464-2b3e7f7a51f8-kube-api-access-2nws7\") pod \"2a23172b-a4cc-41e3-9464-2b3e7f7a51f8\" (UID: \"2a23172b-a4cc-41e3-9464-2b3e7f7a51f8\") " Nov 28 13:49:13 crc kubenswrapper[4631]: I1128 13:49:13.786082 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2a23172b-a4cc-41e3-9464-2b3e7f7a51f8-inventory\") pod \"2a23172b-a4cc-41e3-9464-2b3e7f7a51f8\" (UID: \"2a23172b-a4cc-41e3-9464-2b3e7f7a51f8\") " Nov 28 13:49:13 crc kubenswrapper[4631]: I1128 13:49:13.791586 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2a23172b-a4cc-41e3-9464-2b3e7f7a51f8-kube-api-access-2nws7" (OuterVolumeSpecName: "kube-api-access-2nws7") pod "2a23172b-a4cc-41e3-9464-2b3e7f7a51f8" (UID: "2a23172b-a4cc-41e3-9464-2b3e7f7a51f8"). InnerVolumeSpecName "kube-api-access-2nws7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:49:13 crc kubenswrapper[4631]: I1128 13:49:13.821490 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2a23172b-a4cc-41e3-9464-2b3e7f7a51f8-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "2a23172b-a4cc-41e3-9464-2b3e7f7a51f8" (UID: "2a23172b-a4cc-41e3-9464-2b3e7f7a51f8"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:49:13 crc kubenswrapper[4631]: I1128 13:49:13.831416 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2a23172b-a4cc-41e3-9464-2b3e7f7a51f8-inventory" (OuterVolumeSpecName: "inventory") pod "2a23172b-a4cc-41e3-9464-2b3e7f7a51f8" (UID: "2a23172b-a4cc-41e3-9464-2b3e7f7a51f8"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:49:13 crc kubenswrapper[4631]: I1128 13:49:13.888613 4631 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2a23172b-a4cc-41e3-9464-2b3e7f7a51f8-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 28 13:49:13 crc kubenswrapper[4631]: I1128 13:49:13.888656 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2nws7\" (UniqueName: \"kubernetes.io/projected/2a23172b-a4cc-41e3-9464-2b3e7f7a51f8-kube-api-access-2nws7\") on node \"crc\" DevicePath \"\"" Nov 28 13:49:13 crc kubenswrapper[4631]: I1128 13:49:13.888670 4631 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2a23172b-a4cc-41e3-9464-2b3e7f7a51f8-inventory\") on node \"crc\" DevicePath \"\"" Nov 28 13:49:14 crc kubenswrapper[4631]: I1128 13:49:14.288988 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-2j6xr" event={"ID":"2a23172b-a4cc-41e3-9464-2b3e7f7a51f8","Type":"ContainerDied","Data":"22d474501dd70646b16952d731656831447f79fa2f853ce29b6a9f23669540a7"} Nov 28 13:49:14 crc kubenswrapper[4631]: I1128 13:49:14.289329 4631 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="22d474501dd70646b16952d731656831447f79fa2f853ce29b6a9f23669540a7" Nov 28 13:49:14 crc kubenswrapper[4631]: I1128 13:49:14.289059 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-2j6xr" Nov 28 13:49:14 crc kubenswrapper[4631]: I1128 13:49:14.394661 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-4lxnq"] Nov 28 13:49:14 crc kubenswrapper[4631]: E1128 13:49:14.395328 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2a23172b-a4cc-41e3-9464-2b3e7f7a51f8" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Nov 28 13:49:14 crc kubenswrapper[4631]: I1128 13:49:14.395353 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="2a23172b-a4cc-41e3-9464-2b3e7f7a51f8" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Nov 28 13:49:14 crc kubenswrapper[4631]: I1128 13:49:14.395674 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="2a23172b-a4cc-41e3-9464-2b3e7f7a51f8" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Nov 28 13:49:14 crc kubenswrapper[4631]: I1128 13:49:14.396593 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-4lxnq" Nov 28 13:49:14 crc kubenswrapper[4631]: I1128 13:49:14.398460 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-vp4ns" Nov 28 13:49:14 crc kubenswrapper[4631]: I1128 13:49:14.398679 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 28 13:49:14 crc kubenswrapper[4631]: I1128 13:49:14.398854 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 28 13:49:14 crc kubenswrapper[4631]: I1128 13:49:14.399110 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 28 13:49:14 crc kubenswrapper[4631]: I1128 13:49:14.417405 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-4lxnq"] Nov 28 13:49:14 crc kubenswrapper[4631]: I1128 13:49:14.498968 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4d2450a3-e851-4bfb-83c1-115334898784-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-4lxnq\" (UID: \"4d2450a3-e851-4bfb-83c1-115334898784\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-4lxnq" Nov 28 13:49:14 crc kubenswrapper[4631]: I1128 13:49:14.499070 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4d2450a3-e851-4bfb-83c1-115334898784-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-4lxnq\" (UID: \"4d2450a3-e851-4bfb-83c1-115334898784\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-4lxnq" Nov 28 13:49:14 crc kubenswrapper[4631]: I1128 13:49:14.499114 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sr5td\" (UniqueName: \"kubernetes.io/projected/4d2450a3-e851-4bfb-83c1-115334898784-kube-api-access-sr5td\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-4lxnq\" (UID: \"4d2450a3-e851-4bfb-83c1-115334898784\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-4lxnq" Nov 28 13:49:14 crc kubenswrapper[4631]: I1128 13:49:14.600893 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4d2450a3-e851-4bfb-83c1-115334898784-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-4lxnq\" (UID: \"4d2450a3-e851-4bfb-83c1-115334898784\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-4lxnq" Nov 28 13:49:14 crc kubenswrapper[4631]: I1128 13:49:14.601022 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4d2450a3-e851-4bfb-83c1-115334898784-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-4lxnq\" (UID: \"4d2450a3-e851-4bfb-83c1-115334898784\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-4lxnq" Nov 28 13:49:14 crc kubenswrapper[4631]: I1128 13:49:14.601097 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sr5td\" (UniqueName: \"kubernetes.io/projected/4d2450a3-e851-4bfb-83c1-115334898784-kube-api-access-sr5td\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-4lxnq\" (UID: \"4d2450a3-e851-4bfb-83c1-115334898784\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-4lxnq" Nov 28 13:49:14 crc kubenswrapper[4631]: I1128 13:49:14.606221 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4d2450a3-e851-4bfb-83c1-115334898784-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-4lxnq\" (UID: \"4d2450a3-e851-4bfb-83c1-115334898784\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-4lxnq" Nov 28 13:49:14 crc kubenswrapper[4631]: I1128 13:49:14.606323 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4d2450a3-e851-4bfb-83c1-115334898784-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-4lxnq\" (UID: \"4d2450a3-e851-4bfb-83c1-115334898784\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-4lxnq" Nov 28 13:49:14 crc kubenswrapper[4631]: I1128 13:49:14.626699 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sr5td\" (UniqueName: \"kubernetes.io/projected/4d2450a3-e851-4bfb-83c1-115334898784-kube-api-access-sr5td\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-4lxnq\" (UID: \"4d2450a3-e851-4bfb-83c1-115334898784\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-4lxnq" Nov 28 13:49:14 crc kubenswrapper[4631]: I1128 13:49:14.718481 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-4lxnq" Nov 28 13:49:15 crc kubenswrapper[4631]: I1128 13:49:15.236807 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-4lxnq"] Nov 28 13:49:15 crc kubenswrapper[4631]: I1128 13:49:15.298632 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-4lxnq" event={"ID":"4d2450a3-e851-4bfb-83c1-115334898784","Type":"ContainerStarted","Data":"c344c8c0c76ec483a9cf788fabd4967e8b0a56c3b7348307bfcfe530eb7090f5"} Nov 28 13:49:16 crc kubenswrapper[4631]: I1128 13:49:16.321460 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-4lxnq" event={"ID":"4d2450a3-e851-4bfb-83c1-115334898784","Type":"ContainerStarted","Data":"2ec8c03b5e14f6c048226b11fa2b4b16ff5fb5513167b7ab41d7f27866c2df3b"} Nov 28 13:49:17 crc kubenswrapper[4631]: I1128 13:49:17.514239 4631 scope.go:117] "RemoveContainer" containerID="54d38280b4f1c123c61719ed71e612b2bf974f3bf3d8ff13fe19d70fa5ca5198" Nov 28 13:49:17 crc kubenswrapper[4631]: E1128 13:49:17.514577 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-47sxw_openshift-machine-config-operator(8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1)\"" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" podUID="8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1" Nov 28 13:49:32 crc kubenswrapper[4631]: I1128 13:49:32.513951 4631 scope.go:117] "RemoveContainer" containerID="54d38280b4f1c123c61719ed71e612b2bf974f3bf3d8ff13fe19d70fa5ca5198" Nov 28 13:49:32 crc kubenswrapper[4631]: E1128 13:49:32.514775 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-47sxw_openshift-machine-config-operator(8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1)\"" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" podUID="8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1" Nov 28 13:49:44 crc kubenswrapper[4631]: I1128 13:49:44.048127 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-4lxnq" podStartSLOduration=29.584321243 podStartE2EDuration="30.048100342s" podCreationTimestamp="2025-11-28 13:49:14 +0000 UTC" firstStartedPulling="2025-11-28 13:49:15.24233637 +0000 UTC m=+1712.049639734" lastFinishedPulling="2025-11-28 13:49:15.706115489 +0000 UTC m=+1712.513418833" observedRunningTime="2025-11-28 13:49:16.355525909 +0000 UTC m=+1713.162829253" watchObservedRunningTime="2025-11-28 13:49:44.048100342 +0000 UTC m=+1740.855403686" Nov 28 13:49:44 crc kubenswrapper[4631]: I1128 13:49:44.050827 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-sqgzw"] Nov 28 13:49:44 crc kubenswrapper[4631]: I1128 13:49:44.058368 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-5cd3-account-create-update-hrtjp"] Nov 28 13:49:44 crc kubenswrapper[4631]: I1128 13:49:44.066580 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-fc8wj"] Nov 28 13:49:44 crc kubenswrapper[4631]: I1128 13:49:44.078023 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-x4t6b"] Nov 28 13:49:44 crc kubenswrapper[4631]: I1128 13:49:44.099996 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-0720-account-create-update-fhl44"] Nov 28 13:49:44 crc kubenswrapper[4631]: I1128 13:49:44.115463 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-099f-account-create-update-f2xhr"] Nov 28 13:49:44 crc kubenswrapper[4631]: I1128 13:49:44.122268 4631 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-5cd3-account-create-update-hrtjp"] Nov 28 13:49:44 crc kubenswrapper[4631]: I1128 13:49:44.130534 4631 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-sqgzw"] Nov 28 13:49:44 crc kubenswrapper[4631]: I1128 13:49:44.140928 4631 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-099f-account-create-update-f2xhr"] Nov 28 13:49:44 crc kubenswrapper[4631]: I1128 13:49:44.149251 4631 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-fc8wj"] Nov 28 13:49:44 crc kubenswrapper[4631]: I1128 13:49:44.160344 4631 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-0720-account-create-update-fhl44"] Nov 28 13:49:44 crc kubenswrapper[4631]: I1128 13:49:44.171913 4631 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-x4t6b"] Nov 28 13:49:44 crc kubenswrapper[4631]: I1128 13:49:44.513755 4631 scope.go:117] "RemoveContainer" containerID="54d38280b4f1c123c61719ed71e612b2bf974f3bf3d8ff13fe19d70fa5ca5198" Nov 28 13:49:44 crc kubenswrapper[4631]: E1128 13:49:44.514454 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-47sxw_openshift-machine-config-operator(8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1)\"" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" podUID="8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1" Nov 28 13:49:45 crc kubenswrapper[4631]: I1128 13:49:45.530578 4631 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0f61d189-17d8-4154-ac95-df78d4042055" path="/var/lib/kubelet/pods/0f61d189-17d8-4154-ac95-df78d4042055/volumes" Nov 28 13:49:45 crc kubenswrapper[4631]: I1128 13:49:45.531347 4631 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="325fdb2e-59f8-4d94-860f-3ae7bd2e231e" path="/var/lib/kubelet/pods/325fdb2e-59f8-4d94-860f-3ae7bd2e231e/volumes" Nov 28 13:49:45 crc kubenswrapper[4631]: I1128 13:49:45.532109 4631 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5a1eb7ab-04ab-476e-aa72-2a3a3f81eb21" path="/var/lib/kubelet/pods/5a1eb7ab-04ab-476e-aa72-2a3a3f81eb21/volumes" Nov 28 13:49:45 crc kubenswrapper[4631]: I1128 13:49:45.532858 4631 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="641b1d24-f22a-4d63-8e9e-d76c3dde2135" path="/var/lib/kubelet/pods/641b1d24-f22a-4d63-8e9e-d76c3dde2135/volumes" Nov 28 13:49:45 crc kubenswrapper[4631]: I1128 13:49:45.534262 4631 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="98edb37a-2698-4c30-a397-70d83441063d" path="/var/lib/kubelet/pods/98edb37a-2698-4c30-a397-70d83441063d/volumes" Nov 28 13:49:45 crc kubenswrapper[4631]: I1128 13:49:45.534966 4631 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c5a28e08-ff16-4ecf-97ae-543c9de270f8" path="/var/lib/kubelet/pods/c5a28e08-ff16-4ecf-97ae-543c9de270f8/volumes" Nov 28 13:49:57 crc kubenswrapper[4631]: I1128 13:49:57.513589 4631 scope.go:117] "RemoveContainer" containerID="54d38280b4f1c123c61719ed71e612b2bf974f3bf3d8ff13fe19d70fa5ca5198" Nov 28 13:49:57 crc kubenswrapper[4631]: E1128 13:49:57.514477 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-47sxw_openshift-machine-config-operator(8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1)\"" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" podUID="8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1" Nov 28 13:50:04 crc kubenswrapper[4631]: I1128 13:50:04.173912 4631 scope.go:117] "RemoveContainer" containerID="ef5349a016a2aa007616d1b0bd682c055321668c090392075d244fb0a424d9e2" Nov 28 13:50:04 crc kubenswrapper[4631]: I1128 13:50:04.217763 4631 scope.go:117] "RemoveContainer" containerID="92e24cc7e65e275d64d28d4779c3e78ddb7fa944fc5c9460736e81d594bb492f" Nov 28 13:50:04 crc kubenswrapper[4631]: I1128 13:50:04.265941 4631 scope.go:117] "RemoveContainer" containerID="a08b1a5b404934f0700e2e2d8cf33b72868086a48b28b7e15d65392a9e1d62a1" Nov 28 13:50:04 crc kubenswrapper[4631]: I1128 13:50:04.312670 4631 scope.go:117] "RemoveContainer" containerID="aba4ee440f097b8a145c46ec9a1b706abef9c12de69dbf06a20556acea03afa4" Nov 28 13:50:04 crc kubenswrapper[4631]: I1128 13:50:04.368847 4631 scope.go:117] "RemoveContainer" containerID="fa715c259033c082f538da5fdbeffa2585161112c7dcee2b390f49966d22329d" Nov 28 13:50:04 crc kubenswrapper[4631]: I1128 13:50:04.419355 4631 scope.go:117] "RemoveContainer" containerID="b742ee5d45fe80719ab66c8cdecf4b2c06b42061a669d4bff8463258e674cf40" Nov 28 13:50:08 crc kubenswrapper[4631]: I1128 13:50:08.513631 4631 scope.go:117] "RemoveContainer" containerID="54d38280b4f1c123c61719ed71e612b2bf974f3bf3d8ff13fe19d70fa5ca5198" Nov 28 13:50:08 crc kubenswrapper[4631]: E1128 13:50:08.514169 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-47sxw_openshift-machine-config-operator(8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1)\"" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" podUID="8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1" Nov 28 13:50:20 crc kubenswrapper[4631]: I1128 13:50:20.514793 4631 scope.go:117] "RemoveContainer" containerID="54d38280b4f1c123c61719ed71e612b2bf974f3bf3d8ff13fe19d70fa5ca5198" Nov 28 13:50:20 crc kubenswrapper[4631]: E1128 13:50:20.516072 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-47sxw_openshift-machine-config-operator(8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1)\"" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" podUID="8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1" Nov 28 13:50:32 crc kubenswrapper[4631]: I1128 13:50:32.514005 4631 scope.go:117] "RemoveContainer" containerID="54d38280b4f1c123c61719ed71e612b2bf974f3bf3d8ff13fe19d70fa5ca5198" Nov 28 13:50:32 crc kubenswrapper[4631]: E1128 13:50:32.514776 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-47sxw_openshift-machine-config-operator(8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1)\"" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" podUID="8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1" Nov 28 13:50:33 crc kubenswrapper[4631]: I1128 13:50:33.050662 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-tqj67"] Nov 28 13:50:33 crc kubenswrapper[4631]: I1128 13:50:33.058274 4631 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-tqj67"] Nov 28 13:50:33 crc kubenswrapper[4631]: I1128 13:50:33.531975 4631 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8a4fda29-ad66-4077-b2f9-796fe6fd961a" path="/var/lib/kubelet/pods/8a4fda29-ad66-4077-b2f9-796fe6fd961a/volumes" Nov 28 13:50:38 crc kubenswrapper[4631]: I1128 13:50:38.110883 4631 generic.go:334] "Generic (PLEG): container finished" podID="4d2450a3-e851-4bfb-83c1-115334898784" containerID="2ec8c03b5e14f6c048226b11fa2b4b16ff5fb5513167b7ab41d7f27866c2df3b" exitCode=0 Nov 28 13:50:38 crc kubenswrapper[4631]: I1128 13:50:38.110941 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-4lxnq" event={"ID":"4d2450a3-e851-4bfb-83c1-115334898784","Type":"ContainerDied","Data":"2ec8c03b5e14f6c048226b11fa2b4b16ff5fb5513167b7ab41d7f27866c2df3b"} Nov 28 13:50:39 crc kubenswrapper[4631]: I1128 13:50:39.721125 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-4lxnq" Nov 28 13:50:39 crc kubenswrapper[4631]: I1128 13:50:39.773897 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4d2450a3-e851-4bfb-83c1-115334898784-inventory\") pod \"4d2450a3-e851-4bfb-83c1-115334898784\" (UID: \"4d2450a3-e851-4bfb-83c1-115334898784\") " Nov 28 13:50:39 crc kubenswrapper[4631]: I1128 13:50:39.774203 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sr5td\" (UniqueName: \"kubernetes.io/projected/4d2450a3-e851-4bfb-83c1-115334898784-kube-api-access-sr5td\") pod \"4d2450a3-e851-4bfb-83c1-115334898784\" (UID: \"4d2450a3-e851-4bfb-83c1-115334898784\") " Nov 28 13:50:39 crc kubenswrapper[4631]: I1128 13:50:39.774328 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4d2450a3-e851-4bfb-83c1-115334898784-ssh-key\") pod \"4d2450a3-e851-4bfb-83c1-115334898784\" (UID: \"4d2450a3-e851-4bfb-83c1-115334898784\") " Nov 28 13:50:39 crc kubenswrapper[4631]: I1128 13:50:39.780811 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4d2450a3-e851-4bfb-83c1-115334898784-kube-api-access-sr5td" (OuterVolumeSpecName: "kube-api-access-sr5td") pod "4d2450a3-e851-4bfb-83c1-115334898784" (UID: "4d2450a3-e851-4bfb-83c1-115334898784"). InnerVolumeSpecName "kube-api-access-sr5td". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:50:39 crc kubenswrapper[4631]: I1128 13:50:39.809534 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4d2450a3-e851-4bfb-83c1-115334898784-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "4d2450a3-e851-4bfb-83c1-115334898784" (UID: "4d2450a3-e851-4bfb-83c1-115334898784"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:50:39 crc kubenswrapper[4631]: I1128 13:50:39.827402 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4d2450a3-e851-4bfb-83c1-115334898784-inventory" (OuterVolumeSpecName: "inventory") pod "4d2450a3-e851-4bfb-83c1-115334898784" (UID: "4d2450a3-e851-4bfb-83c1-115334898784"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:50:39 crc kubenswrapper[4631]: I1128 13:50:39.878382 4631 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4d2450a3-e851-4bfb-83c1-115334898784-inventory\") on node \"crc\" DevicePath \"\"" Nov 28 13:50:39 crc kubenswrapper[4631]: I1128 13:50:39.878430 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sr5td\" (UniqueName: \"kubernetes.io/projected/4d2450a3-e851-4bfb-83c1-115334898784-kube-api-access-sr5td\") on node \"crc\" DevicePath \"\"" Nov 28 13:50:39 crc kubenswrapper[4631]: I1128 13:50:39.878444 4631 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4d2450a3-e851-4bfb-83c1-115334898784-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 28 13:50:40 crc kubenswrapper[4631]: I1128 13:50:40.151667 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-4lxnq" event={"ID":"4d2450a3-e851-4bfb-83c1-115334898784","Type":"ContainerDied","Data":"c344c8c0c76ec483a9cf788fabd4967e8b0a56c3b7348307bfcfe530eb7090f5"} Nov 28 13:50:40 crc kubenswrapper[4631]: I1128 13:50:40.152027 4631 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c344c8c0c76ec483a9cf788fabd4967e8b0a56c3b7348307bfcfe530eb7090f5" Nov 28 13:50:40 crc kubenswrapper[4631]: I1128 13:50:40.151812 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-4lxnq" Nov 28 13:50:40 crc kubenswrapper[4631]: I1128 13:50:40.276481 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-pwfzg"] Nov 28 13:50:40 crc kubenswrapper[4631]: E1128 13:50:40.276928 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d2450a3-e851-4bfb-83c1-115334898784" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Nov 28 13:50:40 crc kubenswrapper[4631]: I1128 13:50:40.276953 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d2450a3-e851-4bfb-83c1-115334898784" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Nov 28 13:50:40 crc kubenswrapper[4631]: I1128 13:50:40.277202 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="4d2450a3-e851-4bfb-83c1-115334898784" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Nov 28 13:50:40 crc kubenswrapper[4631]: I1128 13:50:40.278006 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-pwfzg" Nov 28 13:50:40 crc kubenswrapper[4631]: I1128 13:50:40.281198 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 28 13:50:40 crc kubenswrapper[4631]: I1128 13:50:40.281412 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-vp4ns" Nov 28 13:50:40 crc kubenswrapper[4631]: I1128 13:50:40.281606 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 28 13:50:40 crc kubenswrapper[4631]: I1128 13:50:40.281741 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 28 13:50:40 crc kubenswrapper[4631]: I1128 13:50:40.294920 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-pwfzg"] Nov 28 13:50:40 crc kubenswrapper[4631]: I1128 13:50:40.389736 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/53cec30e-969e-4ccf-8ae2-76852297457a-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-pwfzg\" (UID: \"53cec30e-969e-4ccf-8ae2-76852297457a\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-pwfzg" Nov 28 13:50:40 crc kubenswrapper[4631]: I1128 13:50:40.390749 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/53cec30e-969e-4ccf-8ae2-76852297457a-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-pwfzg\" (UID: \"53cec30e-969e-4ccf-8ae2-76852297457a\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-pwfzg" Nov 28 13:50:40 crc kubenswrapper[4631]: I1128 13:50:40.391092 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t49vv\" (UniqueName: \"kubernetes.io/projected/53cec30e-969e-4ccf-8ae2-76852297457a-kube-api-access-t49vv\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-pwfzg\" (UID: \"53cec30e-969e-4ccf-8ae2-76852297457a\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-pwfzg" Nov 28 13:50:40 crc kubenswrapper[4631]: I1128 13:50:40.493044 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/53cec30e-969e-4ccf-8ae2-76852297457a-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-pwfzg\" (UID: \"53cec30e-969e-4ccf-8ae2-76852297457a\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-pwfzg" Nov 28 13:50:40 crc kubenswrapper[4631]: I1128 13:50:40.493548 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t49vv\" (UniqueName: \"kubernetes.io/projected/53cec30e-969e-4ccf-8ae2-76852297457a-kube-api-access-t49vv\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-pwfzg\" (UID: \"53cec30e-969e-4ccf-8ae2-76852297457a\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-pwfzg" Nov 28 13:50:40 crc kubenswrapper[4631]: I1128 13:50:40.493796 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/53cec30e-969e-4ccf-8ae2-76852297457a-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-pwfzg\" (UID: \"53cec30e-969e-4ccf-8ae2-76852297457a\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-pwfzg" Nov 28 13:50:40 crc kubenswrapper[4631]: I1128 13:50:40.498974 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/53cec30e-969e-4ccf-8ae2-76852297457a-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-pwfzg\" (UID: \"53cec30e-969e-4ccf-8ae2-76852297457a\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-pwfzg" Nov 28 13:50:40 crc kubenswrapper[4631]: I1128 13:50:40.499044 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/53cec30e-969e-4ccf-8ae2-76852297457a-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-pwfzg\" (UID: \"53cec30e-969e-4ccf-8ae2-76852297457a\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-pwfzg" Nov 28 13:50:40 crc kubenswrapper[4631]: I1128 13:50:40.515907 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t49vv\" (UniqueName: \"kubernetes.io/projected/53cec30e-969e-4ccf-8ae2-76852297457a-kube-api-access-t49vv\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-pwfzg\" (UID: \"53cec30e-969e-4ccf-8ae2-76852297457a\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-pwfzg" Nov 28 13:50:40 crc kubenswrapper[4631]: I1128 13:50:40.604059 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-pwfzg" Nov 28 13:50:41 crc kubenswrapper[4631]: I1128 13:50:41.192544 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-pwfzg"] Nov 28 13:50:41 crc kubenswrapper[4631]: I1128 13:50:41.202143 4631 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 28 13:50:42 crc kubenswrapper[4631]: I1128 13:50:42.174908 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-pwfzg" event={"ID":"53cec30e-969e-4ccf-8ae2-76852297457a","Type":"ContainerStarted","Data":"6b57753d924ae15a5636be91821755bc3b0c195d0100593bc2afdc09bf0b39d1"} Nov 28 13:50:42 crc kubenswrapper[4631]: I1128 13:50:42.175225 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-pwfzg" event={"ID":"53cec30e-969e-4ccf-8ae2-76852297457a","Type":"ContainerStarted","Data":"97a1ab5eaca8228e3669dce6ac1408a7435c200bb73ce6c693db1b28fc66c48a"} Nov 28 13:50:42 crc kubenswrapper[4631]: I1128 13:50:42.199073 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-pwfzg" podStartSLOduration=1.572628762 podStartE2EDuration="2.199050261s" podCreationTimestamp="2025-11-28 13:50:40 +0000 UTC" firstStartedPulling="2025-11-28 13:50:41.201808411 +0000 UTC m=+1798.009111755" lastFinishedPulling="2025-11-28 13:50:41.82822991 +0000 UTC m=+1798.635533254" observedRunningTime="2025-11-28 13:50:42.196848818 +0000 UTC m=+1799.004152182" watchObservedRunningTime="2025-11-28 13:50:42.199050261 +0000 UTC m=+1799.006353605" Nov 28 13:50:46 crc kubenswrapper[4631]: I1128 13:50:46.514584 4631 scope.go:117] "RemoveContainer" containerID="54d38280b4f1c123c61719ed71e612b2bf974f3bf3d8ff13fe19d70fa5ca5198" Nov 28 13:50:46 crc kubenswrapper[4631]: E1128 13:50:46.515482 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-47sxw_openshift-machine-config-operator(8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1)\"" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" podUID="8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1" Nov 28 13:50:48 crc kubenswrapper[4631]: I1128 13:50:48.236840 4631 generic.go:334] "Generic (PLEG): container finished" podID="53cec30e-969e-4ccf-8ae2-76852297457a" containerID="6b57753d924ae15a5636be91821755bc3b0c195d0100593bc2afdc09bf0b39d1" exitCode=0 Nov 28 13:50:48 crc kubenswrapper[4631]: I1128 13:50:48.237686 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-pwfzg" event={"ID":"53cec30e-969e-4ccf-8ae2-76852297457a","Type":"ContainerDied","Data":"6b57753d924ae15a5636be91821755bc3b0c195d0100593bc2afdc09bf0b39d1"} Nov 28 13:50:49 crc kubenswrapper[4631]: I1128 13:50:49.685111 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-pwfzg" Nov 28 13:50:49 crc kubenswrapper[4631]: I1128 13:50:49.831363 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/53cec30e-969e-4ccf-8ae2-76852297457a-ssh-key\") pod \"53cec30e-969e-4ccf-8ae2-76852297457a\" (UID: \"53cec30e-969e-4ccf-8ae2-76852297457a\") " Nov 28 13:50:49 crc kubenswrapper[4631]: I1128 13:50:49.831482 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/53cec30e-969e-4ccf-8ae2-76852297457a-inventory\") pod \"53cec30e-969e-4ccf-8ae2-76852297457a\" (UID: \"53cec30e-969e-4ccf-8ae2-76852297457a\") " Nov 28 13:50:49 crc kubenswrapper[4631]: I1128 13:50:49.831592 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t49vv\" (UniqueName: \"kubernetes.io/projected/53cec30e-969e-4ccf-8ae2-76852297457a-kube-api-access-t49vv\") pod \"53cec30e-969e-4ccf-8ae2-76852297457a\" (UID: \"53cec30e-969e-4ccf-8ae2-76852297457a\") " Nov 28 13:50:49 crc kubenswrapper[4631]: I1128 13:50:49.837528 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/53cec30e-969e-4ccf-8ae2-76852297457a-kube-api-access-t49vv" (OuterVolumeSpecName: "kube-api-access-t49vv") pod "53cec30e-969e-4ccf-8ae2-76852297457a" (UID: "53cec30e-969e-4ccf-8ae2-76852297457a"). InnerVolumeSpecName "kube-api-access-t49vv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:50:49 crc kubenswrapper[4631]: I1128 13:50:49.862481 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/53cec30e-969e-4ccf-8ae2-76852297457a-inventory" (OuterVolumeSpecName: "inventory") pod "53cec30e-969e-4ccf-8ae2-76852297457a" (UID: "53cec30e-969e-4ccf-8ae2-76852297457a"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:50:49 crc kubenswrapper[4631]: I1128 13:50:49.862790 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/53cec30e-969e-4ccf-8ae2-76852297457a-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "53cec30e-969e-4ccf-8ae2-76852297457a" (UID: "53cec30e-969e-4ccf-8ae2-76852297457a"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:50:49 crc kubenswrapper[4631]: I1128 13:50:49.935553 4631 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/53cec30e-969e-4ccf-8ae2-76852297457a-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 28 13:50:49 crc kubenswrapper[4631]: I1128 13:50:49.935602 4631 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/53cec30e-969e-4ccf-8ae2-76852297457a-inventory\") on node \"crc\" DevicePath \"\"" Nov 28 13:50:49 crc kubenswrapper[4631]: I1128 13:50:49.935631 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t49vv\" (UniqueName: \"kubernetes.io/projected/53cec30e-969e-4ccf-8ae2-76852297457a-kube-api-access-t49vv\") on node \"crc\" DevicePath \"\"" Nov 28 13:50:50 crc kubenswrapper[4631]: I1128 13:50:50.255853 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-pwfzg" event={"ID":"53cec30e-969e-4ccf-8ae2-76852297457a","Type":"ContainerDied","Data":"97a1ab5eaca8228e3669dce6ac1408a7435c200bb73ce6c693db1b28fc66c48a"} Nov 28 13:50:50 crc kubenswrapper[4631]: I1128 13:50:50.255904 4631 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="97a1ab5eaca8228e3669dce6ac1408a7435c200bb73ce6c693db1b28fc66c48a" Nov 28 13:50:50 crc kubenswrapper[4631]: I1128 13:50:50.255971 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-pwfzg" Nov 28 13:50:50 crc kubenswrapper[4631]: I1128 13:50:50.337478 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-gsp5r"] Nov 28 13:50:50 crc kubenswrapper[4631]: E1128 13:50:50.337972 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="53cec30e-969e-4ccf-8ae2-76852297457a" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Nov 28 13:50:50 crc kubenswrapper[4631]: I1128 13:50:50.337996 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="53cec30e-969e-4ccf-8ae2-76852297457a" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Nov 28 13:50:50 crc kubenswrapper[4631]: I1128 13:50:50.338259 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="53cec30e-969e-4ccf-8ae2-76852297457a" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Nov 28 13:50:50 crc kubenswrapper[4631]: I1128 13:50:50.339094 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-gsp5r" Nov 28 13:50:50 crc kubenswrapper[4631]: I1128 13:50:50.342589 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 28 13:50:50 crc kubenswrapper[4631]: I1128 13:50:50.342620 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-vp4ns" Nov 28 13:50:50 crc kubenswrapper[4631]: I1128 13:50:50.343428 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 28 13:50:50 crc kubenswrapper[4631]: I1128 13:50:50.343808 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 28 13:50:50 crc kubenswrapper[4631]: I1128 13:50:50.357446 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-gsp5r"] Nov 28 13:50:50 crc kubenswrapper[4631]: I1128 13:50:50.446850 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x59t6\" (UniqueName: \"kubernetes.io/projected/5f971cb1-7c00-4dbe-abb1-e4eeb825fcee-kube-api-access-x59t6\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-gsp5r\" (UID: \"5f971cb1-7c00-4dbe-abb1-e4eeb825fcee\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-gsp5r" Nov 28 13:50:50 crc kubenswrapper[4631]: I1128 13:50:50.446945 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5f971cb1-7c00-4dbe-abb1-e4eeb825fcee-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-gsp5r\" (UID: \"5f971cb1-7c00-4dbe-abb1-e4eeb825fcee\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-gsp5r" Nov 28 13:50:50 crc kubenswrapper[4631]: I1128 13:50:50.447091 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5f971cb1-7c00-4dbe-abb1-e4eeb825fcee-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-gsp5r\" (UID: \"5f971cb1-7c00-4dbe-abb1-e4eeb825fcee\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-gsp5r" Nov 28 13:50:50 crc kubenswrapper[4631]: I1128 13:50:50.549188 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x59t6\" (UniqueName: \"kubernetes.io/projected/5f971cb1-7c00-4dbe-abb1-e4eeb825fcee-kube-api-access-x59t6\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-gsp5r\" (UID: \"5f971cb1-7c00-4dbe-abb1-e4eeb825fcee\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-gsp5r" Nov 28 13:50:50 crc kubenswrapper[4631]: I1128 13:50:50.549322 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5f971cb1-7c00-4dbe-abb1-e4eeb825fcee-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-gsp5r\" (UID: \"5f971cb1-7c00-4dbe-abb1-e4eeb825fcee\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-gsp5r" Nov 28 13:50:50 crc kubenswrapper[4631]: I1128 13:50:50.549432 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5f971cb1-7c00-4dbe-abb1-e4eeb825fcee-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-gsp5r\" (UID: \"5f971cb1-7c00-4dbe-abb1-e4eeb825fcee\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-gsp5r" Nov 28 13:50:50 crc kubenswrapper[4631]: I1128 13:50:50.553843 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5f971cb1-7c00-4dbe-abb1-e4eeb825fcee-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-gsp5r\" (UID: \"5f971cb1-7c00-4dbe-abb1-e4eeb825fcee\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-gsp5r" Nov 28 13:50:50 crc kubenswrapper[4631]: I1128 13:50:50.554349 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5f971cb1-7c00-4dbe-abb1-e4eeb825fcee-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-gsp5r\" (UID: \"5f971cb1-7c00-4dbe-abb1-e4eeb825fcee\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-gsp5r" Nov 28 13:50:50 crc kubenswrapper[4631]: I1128 13:50:50.569087 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x59t6\" (UniqueName: \"kubernetes.io/projected/5f971cb1-7c00-4dbe-abb1-e4eeb825fcee-kube-api-access-x59t6\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-gsp5r\" (UID: \"5f971cb1-7c00-4dbe-abb1-e4eeb825fcee\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-gsp5r" Nov 28 13:50:50 crc kubenswrapper[4631]: I1128 13:50:50.661040 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-gsp5r" Nov 28 13:50:51 crc kubenswrapper[4631]: I1128 13:50:51.194410 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-gsp5r"] Nov 28 13:50:51 crc kubenswrapper[4631]: I1128 13:50:51.269209 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-gsp5r" event={"ID":"5f971cb1-7c00-4dbe-abb1-e4eeb825fcee","Type":"ContainerStarted","Data":"7da3e2d1a01c91481791636cdd61caa2f06be09b008d52bfbf03ea7e1b1800b3"} Nov 28 13:50:52 crc kubenswrapper[4631]: I1128 13:50:52.278417 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-gsp5r" event={"ID":"5f971cb1-7c00-4dbe-abb1-e4eeb825fcee","Type":"ContainerStarted","Data":"6a54244b941bd62ab27f01b036e98d742d4b6a1e8802a4daffc54f2fcdd78710"} Nov 28 13:50:52 crc kubenswrapper[4631]: I1128 13:50:52.298804 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-gsp5r" podStartSLOduration=1.7086947590000001 podStartE2EDuration="2.298778927s" podCreationTimestamp="2025-11-28 13:50:50 +0000 UTC" firstStartedPulling="2025-11-28 13:50:51.2028243 +0000 UTC m=+1808.010127644" lastFinishedPulling="2025-11-28 13:50:51.792908468 +0000 UTC m=+1808.600211812" observedRunningTime="2025-11-28 13:50:52.295495218 +0000 UTC m=+1809.102798562" watchObservedRunningTime="2025-11-28 13:50:52.298778927 +0000 UTC m=+1809.106082271" Nov 28 13:50:59 crc kubenswrapper[4631]: I1128 13:50:59.513775 4631 scope.go:117] "RemoveContainer" containerID="54d38280b4f1c123c61719ed71e612b2bf974f3bf3d8ff13fe19d70fa5ca5198" Nov 28 13:50:59 crc kubenswrapper[4631]: E1128 13:50:59.516703 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-47sxw_openshift-machine-config-operator(8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1)\"" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" podUID="8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1" Nov 28 13:51:02 crc kubenswrapper[4631]: I1128 13:51:02.045802 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-tn4hn"] Nov 28 13:51:02 crc kubenswrapper[4631]: I1128 13:51:02.056272 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-cs6bm"] Nov 28 13:51:02 crc kubenswrapper[4631]: I1128 13:51:02.070268 4631 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-tn4hn"] Nov 28 13:51:02 crc kubenswrapper[4631]: I1128 13:51:02.079703 4631 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-cs6bm"] Nov 28 13:51:03 crc kubenswrapper[4631]: I1128 13:51:03.540878 4631 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="476e6d13-87a1-4155-ae6b-f6453e9c2044" path="/var/lib/kubelet/pods/476e6d13-87a1-4155-ae6b-f6453e9c2044/volumes" Nov 28 13:51:03 crc kubenswrapper[4631]: I1128 13:51:03.542119 4631 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6b7cc6f7-1a01-4c53-818d-f40548209a13" path="/var/lib/kubelet/pods/6b7cc6f7-1a01-4c53-818d-f40548209a13/volumes" Nov 28 13:51:04 crc kubenswrapper[4631]: I1128 13:51:04.588037 4631 scope.go:117] "RemoveContainer" containerID="44bdbae8c83b2de9874d7d4a4580cd7185b4f75b99e0304faaeb9f093af2f0d2" Nov 28 13:51:04 crc kubenswrapper[4631]: I1128 13:51:04.669400 4631 scope.go:117] "RemoveContainer" containerID="d9cdc53db79944ad3a916e49b0da1cb467958dc12bea86bc7c817bf6487bf6ff" Nov 28 13:51:04 crc kubenswrapper[4631]: I1128 13:51:04.713729 4631 scope.go:117] "RemoveContainer" containerID="5e760b5a84be2bd8352e4542c8d70a6aafdf38b37d9178d5a56071b495bfeab2" Nov 28 13:51:14 crc kubenswrapper[4631]: I1128 13:51:14.513844 4631 scope.go:117] "RemoveContainer" containerID="54d38280b4f1c123c61719ed71e612b2bf974f3bf3d8ff13fe19d70fa5ca5198" Nov 28 13:51:14 crc kubenswrapper[4631]: E1128 13:51:14.514728 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-47sxw_openshift-machine-config-operator(8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1)\"" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" podUID="8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1" Nov 28 13:51:25 crc kubenswrapper[4631]: I1128 13:51:25.517134 4631 scope.go:117] "RemoveContainer" containerID="54d38280b4f1c123c61719ed71e612b2bf974f3bf3d8ff13fe19d70fa5ca5198" Nov 28 13:51:25 crc kubenswrapper[4631]: E1128 13:51:25.518592 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-47sxw_openshift-machine-config-operator(8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1)\"" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" podUID="8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1" Nov 28 13:51:33 crc kubenswrapper[4631]: I1128 13:51:33.693441 4631 generic.go:334] "Generic (PLEG): container finished" podID="5f971cb1-7c00-4dbe-abb1-e4eeb825fcee" containerID="6a54244b941bd62ab27f01b036e98d742d4b6a1e8802a4daffc54f2fcdd78710" exitCode=0 Nov 28 13:51:33 crc kubenswrapper[4631]: I1128 13:51:33.693528 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-gsp5r" event={"ID":"5f971cb1-7c00-4dbe-abb1-e4eeb825fcee","Type":"ContainerDied","Data":"6a54244b941bd62ab27f01b036e98d742d4b6a1e8802a4daffc54f2fcdd78710"} Nov 28 13:51:35 crc kubenswrapper[4631]: I1128 13:51:35.137231 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-gsp5r" Nov 28 13:51:35 crc kubenswrapper[4631]: I1128 13:51:35.212729 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5f971cb1-7c00-4dbe-abb1-e4eeb825fcee-inventory\") pod \"5f971cb1-7c00-4dbe-abb1-e4eeb825fcee\" (UID: \"5f971cb1-7c00-4dbe-abb1-e4eeb825fcee\") " Nov 28 13:51:35 crc kubenswrapper[4631]: I1128 13:51:35.212800 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5f971cb1-7c00-4dbe-abb1-e4eeb825fcee-ssh-key\") pod \"5f971cb1-7c00-4dbe-abb1-e4eeb825fcee\" (UID: \"5f971cb1-7c00-4dbe-abb1-e4eeb825fcee\") " Nov 28 13:51:35 crc kubenswrapper[4631]: I1128 13:51:35.213197 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x59t6\" (UniqueName: \"kubernetes.io/projected/5f971cb1-7c00-4dbe-abb1-e4eeb825fcee-kube-api-access-x59t6\") pod \"5f971cb1-7c00-4dbe-abb1-e4eeb825fcee\" (UID: \"5f971cb1-7c00-4dbe-abb1-e4eeb825fcee\") " Nov 28 13:51:35 crc kubenswrapper[4631]: I1128 13:51:35.220721 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5f971cb1-7c00-4dbe-abb1-e4eeb825fcee-kube-api-access-x59t6" (OuterVolumeSpecName: "kube-api-access-x59t6") pod "5f971cb1-7c00-4dbe-abb1-e4eeb825fcee" (UID: "5f971cb1-7c00-4dbe-abb1-e4eeb825fcee"). InnerVolumeSpecName "kube-api-access-x59t6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:51:35 crc kubenswrapper[4631]: I1128 13:51:35.246429 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5f971cb1-7c00-4dbe-abb1-e4eeb825fcee-inventory" (OuterVolumeSpecName: "inventory") pod "5f971cb1-7c00-4dbe-abb1-e4eeb825fcee" (UID: "5f971cb1-7c00-4dbe-abb1-e4eeb825fcee"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:51:35 crc kubenswrapper[4631]: I1128 13:51:35.248883 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5f971cb1-7c00-4dbe-abb1-e4eeb825fcee-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "5f971cb1-7c00-4dbe-abb1-e4eeb825fcee" (UID: "5f971cb1-7c00-4dbe-abb1-e4eeb825fcee"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:51:35 crc kubenswrapper[4631]: I1128 13:51:35.316510 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x59t6\" (UniqueName: \"kubernetes.io/projected/5f971cb1-7c00-4dbe-abb1-e4eeb825fcee-kube-api-access-x59t6\") on node \"crc\" DevicePath \"\"" Nov 28 13:51:35 crc kubenswrapper[4631]: I1128 13:51:35.316931 4631 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5f971cb1-7c00-4dbe-abb1-e4eeb825fcee-inventory\") on node \"crc\" DevicePath \"\"" Nov 28 13:51:35 crc kubenswrapper[4631]: I1128 13:51:35.317007 4631 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5f971cb1-7c00-4dbe-abb1-e4eeb825fcee-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 28 13:51:35 crc kubenswrapper[4631]: I1128 13:51:35.730187 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-gsp5r" event={"ID":"5f971cb1-7c00-4dbe-abb1-e4eeb825fcee","Type":"ContainerDied","Data":"7da3e2d1a01c91481791636cdd61caa2f06be09b008d52bfbf03ea7e1b1800b3"} Nov 28 13:51:35 crc kubenswrapper[4631]: I1128 13:51:35.730266 4631 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7da3e2d1a01c91481791636cdd61caa2f06be09b008d52bfbf03ea7e1b1800b3" Nov 28 13:51:35 crc kubenswrapper[4631]: I1128 13:51:35.730429 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-gsp5r" Nov 28 13:51:35 crc kubenswrapper[4631]: I1128 13:51:35.831496 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-s9jts"] Nov 28 13:51:35 crc kubenswrapper[4631]: E1128 13:51:35.832511 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5f971cb1-7c00-4dbe-abb1-e4eeb825fcee" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Nov 28 13:51:35 crc kubenswrapper[4631]: I1128 13:51:35.833082 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="5f971cb1-7c00-4dbe-abb1-e4eeb825fcee" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Nov 28 13:51:35 crc kubenswrapper[4631]: I1128 13:51:35.833380 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="5f971cb1-7c00-4dbe-abb1-e4eeb825fcee" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Nov 28 13:51:35 crc kubenswrapper[4631]: I1128 13:51:35.834273 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-s9jts" Nov 28 13:51:35 crc kubenswrapper[4631]: I1128 13:51:35.838680 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 28 13:51:35 crc kubenswrapper[4631]: I1128 13:51:35.839019 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 28 13:51:35 crc kubenswrapper[4631]: I1128 13:51:35.839296 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-vp4ns" Nov 28 13:51:35 crc kubenswrapper[4631]: I1128 13:51:35.839435 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 28 13:51:35 crc kubenswrapper[4631]: I1128 13:51:35.874445 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-s9jts"] Nov 28 13:51:35 crc kubenswrapper[4631]: I1128 13:51:35.929041 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8133b1ca-bdb2-46e1-8d39-2b45ab9284c8-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-s9jts\" (UID: \"8133b1ca-bdb2-46e1-8d39-2b45ab9284c8\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-s9jts" Nov 28 13:51:35 crc kubenswrapper[4631]: I1128 13:51:35.929487 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8133b1ca-bdb2-46e1-8d39-2b45ab9284c8-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-s9jts\" (UID: \"8133b1ca-bdb2-46e1-8d39-2b45ab9284c8\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-s9jts" Nov 28 13:51:35 crc kubenswrapper[4631]: I1128 13:51:35.929621 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-prntm\" (UniqueName: \"kubernetes.io/projected/8133b1ca-bdb2-46e1-8d39-2b45ab9284c8-kube-api-access-prntm\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-s9jts\" (UID: \"8133b1ca-bdb2-46e1-8d39-2b45ab9284c8\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-s9jts" Nov 28 13:51:36 crc kubenswrapper[4631]: I1128 13:51:36.034676 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8133b1ca-bdb2-46e1-8d39-2b45ab9284c8-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-s9jts\" (UID: \"8133b1ca-bdb2-46e1-8d39-2b45ab9284c8\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-s9jts" Nov 28 13:51:36 crc kubenswrapper[4631]: I1128 13:51:36.034801 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8133b1ca-bdb2-46e1-8d39-2b45ab9284c8-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-s9jts\" (UID: \"8133b1ca-bdb2-46e1-8d39-2b45ab9284c8\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-s9jts" Nov 28 13:51:36 crc kubenswrapper[4631]: I1128 13:51:36.034866 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-prntm\" (UniqueName: \"kubernetes.io/projected/8133b1ca-bdb2-46e1-8d39-2b45ab9284c8-kube-api-access-prntm\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-s9jts\" (UID: \"8133b1ca-bdb2-46e1-8d39-2b45ab9284c8\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-s9jts" Nov 28 13:51:36 crc kubenswrapper[4631]: I1128 13:51:36.043918 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8133b1ca-bdb2-46e1-8d39-2b45ab9284c8-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-s9jts\" (UID: \"8133b1ca-bdb2-46e1-8d39-2b45ab9284c8\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-s9jts" Nov 28 13:51:36 crc kubenswrapper[4631]: I1128 13:51:36.052754 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8133b1ca-bdb2-46e1-8d39-2b45ab9284c8-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-s9jts\" (UID: \"8133b1ca-bdb2-46e1-8d39-2b45ab9284c8\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-s9jts" Nov 28 13:51:36 crc kubenswrapper[4631]: I1128 13:51:36.056231 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-prntm\" (UniqueName: \"kubernetes.io/projected/8133b1ca-bdb2-46e1-8d39-2b45ab9284c8-kube-api-access-prntm\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-s9jts\" (UID: \"8133b1ca-bdb2-46e1-8d39-2b45ab9284c8\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-s9jts" Nov 28 13:51:36 crc kubenswrapper[4631]: I1128 13:51:36.162837 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-s9jts" Nov 28 13:51:36 crc kubenswrapper[4631]: I1128 13:51:36.668067 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-s9jts"] Nov 28 13:51:36 crc kubenswrapper[4631]: I1128 13:51:36.746426 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-s9jts" event={"ID":"8133b1ca-bdb2-46e1-8d39-2b45ab9284c8","Type":"ContainerStarted","Data":"e59da126c6291a8d7166156168f16437c880a3f76df0e7e5971b933fc5025c98"} Nov 28 13:51:37 crc kubenswrapper[4631]: I1128 13:51:37.759161 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-s9jts" event={"ID":"8133b1ca-bdb2-46e1-8d39-2b45ab9284c8","Type":"ContainerStarted","Data":"1efefda6e68c1f9a93c36d3a3fa45258721effb5de78f9be1c03649c19db643c"} Nov 28 13:51:37 crc kubenswrapper[4631]: I1128 13:51:37.788645 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-s9jts" podStartSLOduration=2.3157175309999998 podStartE2EDuration="2.788620109s" podCreationTimestamp="2025-11-28 13:51:35 +0000 UTC" firstStartedPulling="2025-11-28 13:51:36.672270143 +0000 UTC m=+1853.479573487" lastFinishedPulling="2025-11-28 13:51:37.145172721 +0000 UTC m=+1853.952476065" observedRunningTime="2025-11-28 13:51:37.78450592 +0000 UTC m=+1854.591809264" watchObservedRunningTime="2025-11-28 13:51:37.788620109 +0000 UTC m=+1854.595923443" Nov 28 13:51:40 crc kubenswrapper[4631]: I1128 13:51:40.513583 4631 scope.go:117] "RemoveContainer" containerID="54d38280b4f1c123c61719ed71e612b2bf974f3bf3d8ff13fe19d70fa5ca5198" Nov 28 13:51:40 crc kubenswrapper[4631]: E1128 13:51:40.514117 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-47sxw_openshift-machine-config-operator(8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1)\"" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" podUID="8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1" Nov 28 13:51:46 crc kubenswrapper[4631]: I1128 13:51:46.048629 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-24v2l"] Nov 28 13:51:46 crc kubenswrapper[4631]: I1128 13:51:46.063396 4631 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-24v2l"] Nov 28 13:51:47 crc kubenswrapper[4631]: I1128 13:51:47.531106 4631 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0339440e-8081-4f68-9c6f-b62e5be5fc27" path="/var/lib/kubelet/pods/0339440e-8081-4f68-9c6f-b62e5be5fc27/volumes" Nov 28 13:51:52 crc kubenswrapper[4631]: I1128 13:51:52.513217 4631 scope.go:117] "RemoveContainer" containerID="54d38280b4f1c123c61719ed71e612b2bf974f3bf3d8ff13fe19d70fa5ca5198" Nov 28 13:51:52 crc kubenswrapper[4631]: E1128 13:51:52.514072 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-47sxw_openshift-machine-config-operator(8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1)\"" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" podUID="8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1" Nov 28 13:52:04 crc kubenswrapper[4631]: I1128 13:52:04.825221 4631 scope.go:117] "RemoveContainer" containerID="5bb93c4d68648bdadf3355937b106f393adf48407dd01273d6aa3baebd029d62" Nov 28 13:52:06 crc kubenswrapper[4631]: I1128 13:52:06.513934 4631 scope.go:117] "RemoveContainer" containerID="54d38280b4f1c123c61719ed71e612b2bf974f3bf3d8ff13fe19d70fa5ca5198" Nov 28 13:52:06 crc kubenswrapper[4631]: E1128 13:52:06.514688 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-47sxw_openshift-machine-config-operator(8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1)\"" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" podUID="8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1" Nov 28 13:52:17 crc kubenswrapper[4631]: I1128 13:52:17.513496 4631 scope.go:117] "RemoveContainer" containerID="54d38280b4f1c123c61719ed71e612b2bf974f3bf3d8ff13fe19d70fa5ca5198" Nov 28 13:52:17 crc kubenswrapper[4631]: E1128 13:52:17.514477 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-47sxw_openshift-machine-config-operator(8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1)\"" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" podUID="8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1" Nov 28 13:52:28 crc kubenswrapper[4631]: I1128 13:52:28.514063 4631 scope.go:117] "RemoveContainer" containerID="54d38280b4f1c123c61719ed71e612b2bf974f3bf3d8ff13fe19d70fa5ca5198" Nov 28 13:52:28 crc kubenswrapper[4631]: E1128 13:52:28.514896 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-47sxw_openshift-machine-config-operator(8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1)\"" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" podUID="8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1" Nov 28 13:52:33 crc kubenswrapper[4631]: I1128 13:52:33.287194 4631 generic.go:334] "Generic (PLEG): container finished" podID="8133b1ca-bdb2-46e1-8d39-2b45ab9284c8" containerID="1efefda6e68c1f9a93c36d3a3fa45258721effb5de78f9be1c03649c19db643c" exitCode=0 Nov 28 13:52:33 crc kubenswrapper[4631]: I1128 13:52:33.287387 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-s9jts" event={"ID":"8133b1ca-bdb2-46e1-8d39-2b45ab9284c8","Type":"ContainerDied","Data":"1efefda6e68c1f9a93c36d3a3fa45258721effb5de78f9be1c03649c19db643c"} Nov 28 13:52:34 crc kubenswrapper[4631]: I1128 13:52:34.687156 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-s9jts" Nov 28 13:52:34 crc kubenswrapper[4631]: I1128 13:52:34.718497 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8133b1ca-bdb2-46e1-8d39-2b45ab9284c8-inventory\") pod \"8133b1ca-bdb2-46e1-8d39-2b45ab9284c8\" (UID: \"8133b1ca-bdb2-46e1-8d39-2b45ab9284c8\") " Nov 28 13:52:34 crc kubenswrapper[4631]: I1128 13:52:34.719453 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8133b1ca-bdb2-46e1-8d39-2b45ab9284c8-ssh-key\") pod \"8133b1ca-bdb2-46e1-8d39-2b45ab9284c8\" (UID: \"8133b1ca-bdb2-46e1-8d39-2b45ab9284c8\") " Nov 28 13:52:34 crc kubenswrapper[4631]: I1128 13:52:34.719783 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-prntm\" (UniqueName: \"kubernetes.io/projected/8133b1ca-bdb2-46e1-8d39-2b45ab9284c8-kube-api-access-prntm\") pod \"8133b1ca-bdb2-46e1-8d39-2b45ab9284c8\" (UID: \"8133b1ca-bdb2-46e1-8d39-2b45ab9284c8\") " Nov 28 13:52:34 crc kubenswrapper[4631]: I1128 13:52:34.726173 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8133b1ca-bdb2-46e1-8d39-2b45ab9284c8-kube-api-access-prntm" (OuterVolumeSpecName: "kube-api-access-prntm") pod "8133b1ca-bdb2-46e1-8d39-2b45ab9284c8" (UID: "8133b1ca-bdb2-46e1-8d39-2b45ab9284c8"). InnerVolumeSpecName "kube-api-access-prntm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:52:34 crc kubenswrapper[4631]: I1128 13:52:34.748425 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8133b1ca-bdb2-46e1-8d39-2b45ab9284c8-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "8133b1ca-bdb2-46e1-8d39-2b45ab9284c8" (UID: "8133b1ca-bdb2-46e1-8d39-2b45ab9284c8"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:52:34 crc kubenswrapper[4631]: I1128 13:52:34.754249 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8133b1ca-bdb2-46e1-8d39-2b45ab9284c8-inventory" (OuterVolumeSpecName: "inventory") pod "8133b1ca-bdb2-46e1-8d39-2b45ab9284c8" (UID: "8133b1ca-bdb2-46e1-8d39-2b45ab9284c8"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:52:34 crc kubenswrapper[4631]: I1128 13:52:34.821479 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-prntm\" (UniqueName: \"kubernetes.io/projected/8133b1ca-bdb2-46e1-8d39-2b45ab9284c8-kube-api-access-prntm\") on node \"crc\" DevicePath \"\"" Nov 28 13:52:34 crc kubenswrapper[4631]: I1128 13:52:34.821507 4631 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8133b1ca-bdb2-46e1-8d39-2b45ab9284c8-inventory\") on node \"crc\" DevicePath \"\"" Nov 28 13:52:34 crc kubenswrapper[4631]: I1128 13:52:34.821516 4631 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8133b1ca-bdb2-46e1-8d39-2b45ab9284c8-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 28 13:52:35 crc kubenswrapper[4631]: I1128 13:52:35.306687 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-s9jts" event={"ID":"8133b1ca-bdb2-46e1-8d39-2b45ab9284c8","Type":"ContainerDied","Data":"e59da126c6291a8d7166156168f16437c880a3f76df0e7e5971b933fc5025c98"} Nov 28 13:52:35 crc kubenswrapper[4631]: I1128 13:52:35.306734 4631 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e59da126c6291a8d7166156168f16437c880a3f76df0e7e5971b933fc5025c98" Nov 28 13:52:35 crc kubenswrapper[4631]: I1128 13:52:35.306735 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-s9jts" Nov 28 13:52:35 crc kubenswrapper[4631]: I1128 13:52:35.412442 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-m2k99"] Nov 28 13:52:35 crc kubenswrapper[4631]: E1128 13:52:35.412829 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8133b1ca-bdb2-46e1-8d39-2b45ab9284c8" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Nov 28 13:52:35 crc kubenswrapper[4631]: I1128 13:52:35.412844 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="8133b1ca-bdb2-46e1-8d39-2b45ab9284c8" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Nov 28 13:52:35 crc kubenswrapper[4631]: I1128 13:52:35.413031 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="8133b1ca-bdb2-46e1-8d39-2b45ab9284c8" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Nov 28 13:52:35 crc kubenswrapper[4631]: I1128 13:52:35.413670 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-m2k99" Nov 28 13:52:35 crc kubenswrapper[4631]: I1128 13:52:35.415946 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 28 13:52:35 crc kubenswrapper[4631]: I1128 13:52:35.416958 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-vp4ns" Nov 28 13:52:35 crc kubenswrapper[4631]: I1128 13:52:35.416986 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 28 13:52:35 crc kubenswrapper[4631]: I1128 13:52:35.417534 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 28 13:52:35 crc kubenswrapper[4631]: I1128 13:52:35.448551 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-m2k99"] Nov 28 13:52:35 crc kubenswrapper[4631]: I1128 13:52:35.537670 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/b1c516b6-6c37-4bfd-a0ee-f28200e19170-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-m2k99\" (UID: \"b1c516b6-6c37-4bfd-a0ee-f28200e19170\") " pod="openstack/ssh-known-hosts-edpm-deployment-m2k99" Nov 28 13:52:35 crc kubenswrapper[4631]: I1128 13:52:35.537789 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/b1c516b6-6c37-4bfd-a0ee-f28200e19170-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-m2k99\" (UID: \"b1c516b6-6c37-4bfd-a0ee-f28200e19170\") " pod="openstack/ssh-known-hosts-edpm-deployment-m2k99" Nov 28 13:52:35 crc kubenswrapper[4631]: I1128 13:52:35.537845 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h84ch\" (UniqueName: \"kubernetes.io/projected/b1c516b6-6c37-4bfd-a0ee-f28200e19170-kube-api-access-h84ch\") pod \"ssh-known-hosts-edpm-deployment-m2k99\" (UID: \"b1c516b6-6c37-4bfd-a0ee-f28200e19170\") " pod="openstack/ssh-known-hosts-edpm-deployment-m2k99" Nov 28 13:52:35 crc kubenswrapper[4631]: I1128 13:52:35.639374 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/b1c516b6-6c37-4bfd-a0ee-f28200e19170-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-m2k99\" (UID: \"b1c516b6-6c37-4bfd-a0ee-f28200e19170\") " pod="openstack/ssh-known-hosts-edpm-deployment-m2k99" Nov 28 13:52:35 crc kubenswrapper[4631]: I1128 13:52:35.639471 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/b1c516b6-6c37-4bfd-a0ee-f28200e19170-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-m2k99\" (UID: \"b1c516b6-6c37-4bfd-a0ee-f28200e19170\") " pod="openstack/ssh-known-hosts-edpm-deployment-m2k99" Nov 28 13:52:35 crc kubenswrapper[4631]: I1128 13:52:35.639541 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h84ch\" (UniqueName: \"kubernetes.io/projected/b1c516b6-6c37-4bfd-a0ee-f28200e19170-kube-api-access-h84ch\") pod \"ssh-known-hosts-edpm-deployment-m2k99\" (UID: \"b1c516b6-6c37-4bfd-a0ee-f28200e19170\") " pod="openstack/ssh-known-hosts-edpm-deployment-m2k99" Nov 28 13:52:35 crc kubenswrapper[4631]: I1128 13:52:35.650674 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/b1c516b6-6c37-4bfd-a0ee-f28200e19170-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-m2k99\" (UID: \"b1c516b6-6c37-4bfd-a0ee-f28200e19170\") " pod="openstack/ssh-known-hosts-edpm-deployment-m2k99" Nov 28 13:52:35 crc kubenswrapper[4631]: I1128 13:52:35.651320 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/b1c516b6-6c37-4bfd-a0ee-f28200e19170-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-m2k99\" (UID: \"b1c516b6-6c37-4bfd-a0ee-f28200e19170\") " pod="openstack/ssh-known-hosts-edpm-deployment-m2k99" Nov 28 13:52:35 crc kubenswrapper[4631]: I1128 13:52:35.659995 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h84ch\" (UniqueName: \"kubernetes.io/projected/b1c516b6-6c37-4bfd-a0ee-f28200e19170-kube-api-access-h84ch\") pod \"ssh-known-hosts-edpm-deployment-m2k99\" (UID: \"b1c516b6-6c37-4bfd-a0ee-f28200e19170\") " pod="openstack/ssh-known-hosts-edpm-deployment-m2k99" Nov 28 13:52:35 crc kubenswrapper[4631]: I1128 13:52:35.737144 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-m2k99" Nov 28 13:52:36 crc kubenswrapper[4631]: I1128 13:52:36.310824 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-m2k99"] Nov 28 13:52:37 crc kubenswrapper[4631]: I1128 13:52:37.322557 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-m2k99" event={"ID":"b1c516b6-6c37-4bfd-a0ee-f28200e19170","Type":"ContainerStarted","Data":"82a0a3081e55ed0959538774fd8580a40d0c65c5b1cfc12dfff02fcf97fef812"} Nov 28 13:52:39 crc kubenswrapper[4631]: I1128 13:52:39.342570 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-m2k99" event={"ID":"b1c516b6-6c37-4bfd-a0ee-f28200e19170","Type":"ContainerStarted","Data":"1cd7e4dfbca8d84f7ddb3d41631c52eae261f99c271d781387c2c98a79823487"} Nov 28 13:52:39 crc kubenswrapper[4631]: I1128 13:52:39.370467 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ssh-known-hosts-edpm-deployment-m2k99" podStartSLOduration=2.901288963 podStartE2EDuration="4.370446444s" podCreationTimestamp="2025-11-28 13:52:35 +0000 UTC" firstStartedPulling="2025-11-28 13:52:36.336414298 +0000 UTC m=+1913.143717642" lastFinishedPulling="2025-11-28 13:52:37.805571769 +0000 UTC m=+1914.612875123" observedRunningTime="2025-11-28 13:52:39.360854131 +0000 UTC m=+1916.168157505" watchObservedRunningTime="2025-11-28 13:52:39.370446444 +0000 UTC m=+1916.177749788" Nov 28 13:52:41 crc kubenswrapper[4631]: I1128 13:52:41.513082 4631 scope.go:117] "RemoveContainer" containerID="54d38280b4f1c123c61719ed71e612b2bf974f3bf3d8ff13fe19d70fa5ca5198" Nov 28 13:52:41 crc kubenswrapper[4631]: E1128 13:52:41.513435 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-47sxw_openshift-machine-config-operator(8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1)\"" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" podUID="8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1" Nov 28 13:52:46 crc kubenswrapper[4631]: I1128 13:52:46.409908 4631 generic.go:334] "Generic (PLEG): container finished" podID="b1c516b6-6c37-4bfd-a0ee-f28200e19170" containerID="1cd7e4dfbca8d84f7ddb3d41631c52eae261f99c271d781387c2c98a79823487" exitCode=0 Nov 28 13:52:46 crc kubenswrapper[4631]: I1128 13:52:46.409963 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-m2k99" event={"ID":"b1c516b6-6c37-4bfd-a0ee-f28200e19170","Type":"ContainerDied","Data":"1cd7e4dfbca8d84f7ddb3d41631c52eae261f99c271d781387c2c98a79823487"} Nov 28 13:52:47 crc kubenswrapper[4631]: I1128 13:52:47.895731 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-m2k99" Nov 28 13:52:48 crc kubenswrapper[4631]: I1128 13:52:48.053047 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/b1c516b6-6c37-4bfd-a0ee-f28200e19170-inventory-0\") pod \"b1c516b6-6c37-4bfd-a0ee-f28200e19170\" (UID: \"b1c516b6-6c37-4bfd-a0ee-f28200e19170\") " Nov 28 13:52:48 crc kubenswrapper[4631]: I1128 13:52:48.053217 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/b1c516b6-6c37-4bfd-a0ee-f28200e19170-ssh-key-openstack-edpm-ipam\") pod \"b1c516b6-6c37-4bfd-a0ee-f28200e19170\" (UID: \"b1c516b6-6c37-4bfd-a0ee-f28200e19170\") " Nov 28 13:52:48 crc kubenswrapper[4631]: I1128 13:52:48.053247 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h84ch\" (UniqueName: \"kubernetes.io/projected/b1c516b6-6c37-4bfd-a0ee-f28200e19170-kube-api-access-h84ch\") pod \"b1c516b6-6c37-4bfd-a0ee-f28200e19170\" (UID: \"b1c516b6-6c37-4bfd-a0ee-f28200e19170\") " Nov 28 13:52:48 crc kubenswrapper[4631]: I1128 13:52:48.063893 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b1c516b6-6c37-4bfd-a0ee-f28200e19170-kube-api-access-h84ch" (OuterVolumeSpecName: "kube-api-access-h84ch") pod "b1c516b6-6c37-4bfd-a0ee-f28200e19170" (UID: "b1c516b6-6c37-4bfd-a0ee-f28200e19170"). InnerVolumeSpecName "kube-api-access-h84ch". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:52:48 crc kubenswrapper[4631]: I1128 13:52:48.085280 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b1c516b6-6c37-4bfd-a0ee-f28200e19170-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "b1c516b6-6c37-4bfd-a0ee-f28200e19170" (UID: "b1c516b6-6c37-4bfd-a0ee-f28200e19170"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:52:48 crc kubenswrapper[4631]: I1128 13:52:48.095471 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b1c516b6-6c37-4bfd-a0ee-f28200e19170-inventory-0" (OuterVolumeSpecName: "inventory-0") pod "b1c516b6-6c37-4bfd-a0ee-f28200e19170" (UID: "b1c516b6-6c37-4bfd-a0ee-f28200e19170"). InnerVolumeSpecName "inventory-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:52:48 crc kubenswrapper[4631]: I1128 13:52:48.156013 4631 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/b1c516b6-6c37-4bfd-a0ee-f28200e19170-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Nov 28 13:52:48 crc kubenswrapper[4631]: I1128 13:52:48.156060 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h84ch\" (UniqueName: \"kubernetes.io/projected/b1c516b6-6c37-4bfd-a0ee-f28200e19170-kube-api-access-h84ch\") on node \"crc\" DevicePath \"\"" Nov 28 13:52:48 crc kubenswrapper[4631]: I1128 13:52:48.156073 4631 reconciler_common.go:293] "Volume detached for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/b1c516b6-6c37-4bfd-a0ee-f28200e19170-inventory-0\") on node \"crc\" DevicePath \"\"" Nov 28 13:52:48 crc kubenswrapper[4631]: I1128 13:52:48.435175 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-m2k99" event={"ID":"b1c516b6-6c37-4bfd-a0ee-f28200e19170","Type":"ContainerDied","Data":"82a0a3081e55ed0959538774fd8580a40d0c65c5b1cfc12dfff02fcf97fef812"} Nov 28 13:52:48 crc kubenswrapper[4631]: I1128 13:52:48.435233 4631 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="82a0a3081e55ed0959538774fd8580a40d0c65c5b1cfc12dfff02fcf97fef812" Nov 28 13:52:48 crc kubenswrapper[4631]: I1128 13:52:48.435393 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-m2k99" Nov 28 13:52:48 crc kubenswrapper[4631]: I1128 13:52:48.546519 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-zmp56"] Nov 28 13:52:48 crc kubenswrapper[4631]: E1128 13:52:48.547163 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b1c516b6-6c37-4bfd-a0ee-f28200e19170" containerName="ssh-known-hosts-edpm-deployment" Nov 28 13:52:48 crc kubenswrapper[4631]: I1128 13:52:48.547193 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="b1c516b6-6c37-4bfd-a0ee-f28200e19170" containerName="ssh-known-hosts-edpm-deployment" Nov 28 13:52:48 crc kubenswrapper[4631]: I1128 13:52:48.547443 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="b1c516b6-6c37-4bfd-a0ee-f28200e19170" containerName="ssh-known-hosts-edpm-deployment" Nov 28 13:52:48 crc kubenswrapper[4631]: I1128 13:52:48.548924 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-zmp56" Nov 28 13:52:48 crc kubenswrapper[4631]: I1128 13:52:48.554743 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-vp4ns" Nov 28 13:52:48 crc kubenswrapper[4631]: I1128 13:52:48.555052 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 28 13:52:48 crc kubenswrapper[4631]: I1128 13:52:48.555215 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 28 13:52:48 crc kubenswrapper[4631]: I1128 13:52:48.555400 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 28 13:52:48 crc kubenswrapper[4631]: I1128 13:52:48.561716 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-zmp56"] Nov 28 13:52:48 crc kubenswrapper[4631]: I1128 13:52:48.665124 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fwchf\" (UniqueName: \"kubernetes.io/projected/8bdc8b41-a7c5-4c60-86ce-6ac5eb65b82f-kube-api-access-fwchf\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-zmp56\" (UID: \"8bdc8b41-a7c5-4c60-86ce-6ac5eb65b82f\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-zmp56" Nov 28 13:52:48 crc kubenswrapper[4631]: I1128 13:52:48.665433 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8bdc8b41-a7c5-4c60-86ce-6ac5eb65b82f-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-zmp56\" (UID: \"8bdc8b41-a7c5-4c60-86ce-6ac5eb65b82f\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-zmp56" Nov 28 13:52:48 crc kubenswrapper[4631]: I1128 13:52:48.665518 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8bdc8b41-a7c5-4c60-86ce-6ac5eb65b82f-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-zmp56\" (UID: \"8bdc8b41-a7c5-4c60-86ce-6ac5eb65b82f\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-zmp56" Nov 28 13:52:48 crc kubenswrapper[4631]: I1128 13:52:48.768170 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8bdc8b41-a7c5-4c60-86ce-6ac5eb65b82f-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-zmp56\" (UID: \"8bdc8b41-a7c5-4c60-86ce-6ac5eb65b82f\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-zmp56" Nov 28 13:52:48 crc kubenswrapper[4631]: I1128 13:52:48.768335 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8bdc8b41-a7c5-4c60-86ce-6ac5eb65b82f-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-zmp56\" (UID: \"8bdc8b41-a7c5-4c60-86ce-6ac5eb65b82f\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-zmp56" Nov 28 13:52:48 crc kubenswrapper[4631]: I1128 13:52:48.768427 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fwchf\" (UniqueName: \"kubernetes.io/projected/8bdc8b41-a7c5-4c60-86ce-6ac5eb65b82f-kube-api-access-fwchf\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-zmp56\" (UID: \"8bdc8b41-a7c5-4c60-86ce-6ac5eb65b82f\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-zmp56" Nov 28 13:52:48 crc kubenswrapper[4631]: I1128 13:52:48.774670 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8bdc8b41-a7c5-4c60-86ce-6ac5eb65b82f-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-zmp56\" (UID: \"8bdc8b41-a7c5-4c60-86ce-6ac5eb65b82f\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-zmp56" Nov 28 13:52:48 crc kubenswrapper[4631]: I1128 13:52:48.779105 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8bdc8b41-a7c5-4c60-86ce-6ac5eb65b82f-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-zmp56\" (UID: \"8bdc8b41-a7c5-4c60-86ce-6ac5eb65b82f\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-zmp56" Nov 28 13:52:48 crc kubenswrapper[4631]: I1128 13:52:48.786931 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fwchf\" (UniqueName: \"kubernetes.io/projected/8bdc8b41-a7c5-4c60-86ce-6ac5eb65b82f-kube-api-access-fwchf\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-zmp56\" (UID: \"8bdc8b41-a7c5-4c60-86ce-6ac5eb65b82f\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-zmp56" Nov 28 13:52:48 crc kubenswrapper[4631]: I1128 13:52:48.877388 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-zmp56" Nov 28 13:52:49 crc kubenswrapper[4631]: I1128 13:52:49.286488 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-zmp56"] Nov 28 13:52:49 crc kubenswrapper[4631]: W1128 13:52:49.308774 4631 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8bdc8b41_a7c5_4c60_86ce_6ac5eb65b82f.slice/crio-33d26370efc0738400af7478bfb6d2332ce492a9cfad5f139c4d56511e914d66 WatchSource:0}: Error finding container 33d26370efc0738400af7478bfb6d2332ce492a9cfad5f139c4d56511e914d66: Status 404 returned error can't find the container with id 33d26370efc0738400af7478bfb6d2332ce492a9cfad5f139c4d56511e914d66 Nov 28 13:52:49 crc kubenswrapper[4631]: I1128 13:52:49.445380 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-zmp56" event={"ID":"8bdc8b41-a7c5-4c60-86ce-6ac5eb65b82f","Type":"ContainerStarted","Data":"33d26370efc0738400af7478bfb6d2332ce492a9cfad5f139c4d56511e914d66"} Nov 28 13:52:52 crc kubenswrapper[4631]: I1128 13:52:52.476015 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-zmp56" event={"ID":"8bdc8b41-a7c5-4c60-86ce-6ac5eb65b82f","Type":"ContainerStarted","Data":"5229420606c41e1c601fe44e5c93e362136bf9e2e44b241900464ebf14655819"} Nov 28 13:52:52 crc kubenswrapper[4631]: I1128 13:52:52.502075 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-zmp56" podStartSLOduration=2.458422582 podStartE2EDuration="4.502046713s" podCreationTimestamp="2025-11-28 13:52:48 +0000 UTC" firstStartedPulling="2025-11-28 13:52:49.312162286 +0000 UTC m=+1926.119465630" lastFinishedPulling="2025-11-28 13:52:51.355786417 +0000 UTC m=+1928.163089761" observedRunningTime="2025-11-28 13:52:52.4940856 +0000 UTC m=+1929.301388944" watchObservedRunningTime="2025-11-28 13:52:52.502046713 +0000 UTC m=+1929.309350057" Nov 28 13:52:53 crc kubenswrapper[4631]: I1128 13:52:53.520258 4631 scope.go:117] "RemoveContainer" containerID="54d38280b4f1c123c61719ed71e612b2bf974f3bf3d8ff13fe19d70fa5ca5198" Nov 28 13:52:54 crc kubenswrapper[4631]: I1128 13:52:54.501034 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" event={"ID":"8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1","Type":"ContainerStarted","Data":"f6f69c42ec6f8171f119b80c48682e57400641797951b8d603755d1b6a845e27"} Nov 28 13:53:00 crc kubenswrapper[4631]: I1128 13:53:00.577271 4631 generic.go:334] "Generic (PLEG): container finished" podID="8bdc8b41-a7c5-4c60-86ce-6ac5eb65b82f" containerID="5229420606c41e1c601fe44e5c93e362136bf9e2e44b241900464ebf14655819" exitCode=0 Nov 28 13:53:00 crc kubenswrapper[4631]: I1128 13:53:00.577794 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-zmp56" event={"ID":"8bdc8b41-a7c5-4c60-86ce-6ac5eb65b82f","Type":"ContainerDied","Data":"5229420606c41e1c601fe44e5c93e362136bf9e2e44b241900464ebf14655819"} Nov 28 13:53:02 crc kubenswrapper[4631]: I1128 13:53:02.015459 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-zmp56" Nov 28 13:53:02 crc kubenswrapper[4631]: I1128 13:53:02.146577 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8bdc8b41-a7c5-4c60-86ce-6ac5eb65b82f-ssh-key\") pod \"8bdc8b41-a7c5-4c60-86ce-6ac5eb65b82f\" (UID: \"8bdc8b41-a7c5-4c60-86ce-6ac5eb65b82f\") " Nov 28 13:53:02 crc kubenswrapper[4631]: I1128 13:53:02.146925 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8bdc8b41-a7c5-4c60-86ce-6ac5eb65b82f-inventory\") pod \"8bdc8b41-a7c5-4c60-86ce-6ac5eb65b82f\" (UID: \"8bdc8b41-a7c5-4c60-86ce-6ac5eb65b82f\") " Nov 28 13:53:02 crc kubenswrapper[4631]: I1128 13:53:02.147154 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fwchf\" (UniqueName: \"kubernetes.io/projected/8bdc8b41-a7c5-4c60-86ce-6ac5eb65b82f-kube-api-access-fwchf\") pod \"8bdc8b41-a7c5-4c60-86ce-6ac5eb65b82f\" (UID: \"8bdc8b41-a7c5-4c60-86ce-6ac5eb65b82f\") " Nov 28 13:53:02 crc kubenswrapper[4631]: I1128 13:53:02.157524 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8bdc8b41-a7c5-4c60-86ce-6ac5eb65b82f-kube-api-access-fwchf" (OuterVolumeSpecName: "kube-api-access-fwchf") pod "8bdc8b41-a7c5-4c60-86ce-6ac5eb65b82f" (UID: "8bdc8b41-a7c5-4c60-86ce-6ac5eb65b82f"). InnerVolumeSpecName "kube-api-access-fwchf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:53:02 crc kubenswrapper[4631]: I1128 13:53:02.178969 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8bdc8b41-a7c5-4c60-86ce-6ac5eb65b82f-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "8bdc8b41-a7c5-4c60-86ce-6ac5eb65b82f" (UID: "8bdc8b41-a7c5-4c60-86ce-6ac5eb65b82f"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:53:02 crc kubenswrapper[4631]: I1128 13:53:02.179530 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8bdc8b41-a7c5-4c60-86ce-6ac5eb65b82f-inventory" (OuterVolumeSpecName: "inventory") pod "8bdc8b41-a7c5-4c60-86ce-6ac5eb65b82f" (UID: "8bdc8b41-a7c5-4c60-86ce-6ac5eb65b82f"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:53:02 crc kubenswrapper[4631]: I1128 13:53:02.250000 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fwchf\" (UniqueName: \"kubernetes.io/projected/8bdc8b41-a7c5-4c60-86ce-6ac5eb65b82f-kube-api-access-fwchf\") on node \"crc\" DevicePath \"\"" Nov 28 13:53:02 crc kubenswrapper[4631]: I1128 13:53:02.250042 4631 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8bdc8b41-a7c5-4c60-86ce-6ac5eb65b82f-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 28 13:53:02 crc kubenswrapper[4631]: I1128 13:53:02.250052 4631 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8bdc8b41-a7c5-4c60-86ce-6ac5eb65b82f-inventory\") on node \"crc\" DevicePath \"\"" Nov 28 13:53:02 crc kubenswrapper[4631]: I1128 13:53:02.597095 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-zmp56" event={"ID":"8bdc8b41-a7c5-4c60-86ce-6ac5eb65b82f","Type":"ContainerDied","Data":"33d26370efc0738400af7478bfb6d2332ce492a9cfad5f139c4d56511e914d66"} Nov 28 13:53:02 crc kubenswrapper[4631]: I1128 13:53:02.597142 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-zmp56" Nov 28 13:53:02 crc kubenswrapper[4631]: I1128 13:53:02.597146 4631 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="33d26370efc0738400af7478bfb6d2332ce492a9cfad5f139c4d56511e914d66" Nov 28 13:53:02 crc kubenswrapper[4631]: I1128 13:53:02.753627 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-r2pq6"] Nov 28 13:53:02 crc kubenswrapper[4631]: E1128 13:53:02.754383 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8bdc8b41-a7c5-4c60-86ce-6ac5eb65b82f" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Nov 28 13:53:02 crc kubenswrapper[4631]: I1128 13:53:02.754412 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="8bdc8b41-a7c5-4c60-86ce-6ac5eb65b82f" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Nov 28 13:53:02 crc kubenswrapper[4631]: I1128 13:53:02.754642 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="8bdc8b41-a7c5-4c60-86ce-6ac5eb65b82f" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Nov 28 13:53:02 crc kubenswrapper[4631]: I1128 13:53:02.755594 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-r2pq6" Nov 28 13:53:02 crc kubenswrapper[4631]: I1128 13:53:02.757661 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 28 13:53:02 crc kubenswrapper[4631]: I1128 13:53:02.760912 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 28 13:53:02 crc kubenswrapper[4631]: I1128 13:53:02.765042 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-r2pq6"] Nov 28 13:53:02 crc kubenswrapper[4631]: I1128 13:53:02.768428 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 28 13:53:02 crc kubenswrapper[4631]: I1128 13:53:02.768628 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-vp4ns" Nov 28 13:53:02 crc kubenswrapper[4631]: I1128 13:53:02.860080 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ec45d53a-f54c-422a-8a43-5b9d4b4ee0c6-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-r2pq6\" (UID: \"ec45d53a-f54c-422a-8a43-5b9d4b4ee0c6\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-r2pq6" Nov 28 13:53:02 crc kubenswrapper[4631]: I1128 13:53:02.860132 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6chdq\" (UniqueName: \"kubernetes.io/projected/ec45d53a-f54c-422a-8a43-5b9d4b4ee0c6-kube-api-access-6chdq\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-r2pq6\" (UID: \"ec45d53a-f54c-422a-8a43-5b9d4b4ee0c6\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-r2pq6" Nov 28 13:53:02 crc kubenswrapper[4631]: I1128 13:53:02.860223 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ec45d53a-f54c-422a-8a43-5b9d4b4ee0c6-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-r2pq6\" (UID: \"ec45d53a-f54c-422a-8a43-5b9d4b4ee0c6\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-r2pq6" Nov 28 13:53:02 crc kubenswrapper[4631]: I1128 13:53:02.962230 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ec45d53a-f54c-422a-8a43-5b9d4b4ee0c6-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-r2pq6\" (UID: \"ec45d53a-f54c-422a-8a43-5b9d4b4ee0c6\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-r2pq6" Nov 28 13:53:02 crc kubenswrapper[4631]: I1128 13:53:02.962337 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6chdq\" (UniqueName: \"kubernetes.io/projected/ec45d53a-f54c-422a-8a43-5b9d4b4ee0c6-kube-api-access-6chdq\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-r2pq6\" (UID: \"ec45d53a-f54c-422a-8a43-5b9d4b4ee0c6\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-r2pq6" Nov 28 13:53:02 crc kubenswrapper[4631]: I1128 13:53:02.962461 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ec45d53a-f54c-422a-8a43-5b9d4b4ee0c6-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-r2pq6\" (UID: \"ec45d53a-f54c-422a-8a43-5b9d4b4ee0c6\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-r2pq6" Nov 28 13:53:02 crc kubenswrapper[4631]: I1128 13:53:02.970350 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ec45d53a-f54c-422a-8a43-5b9d4b4ee0c6-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-r2pq6\" (UID: \"ec45d53a-f54c-422a-8a43-5b9d4b4ee0c6\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-r2pq6" Nov 28 13:53:02 crc kubenswrapper[4631]: I1128 13:53:02.971144 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ec45d53a-f54c-422a-8a43-5b9d4b4ee0c6-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-r2pq6\" (UID: \"ec45d53a-f54c-422a-8a43-5b9d4b4ee0c6\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-r2pq6" Nov 28 13:53:02 crc kubenswrapper[4631]: I1128 13:53:02.987204 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6chdq\" (UniqueName: \"kubernetes.io/projected/ec45d53a-f54c-422a-8a43-5b9d4b4ee0c6-kube-api-access-6chdq\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-r2pq6\" (UID: \"ec45d53a-f54c-422a-8a43-5b9d4b4ee0c6\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-r2pq6" Nov 28 13:53:03 crc kubenswrapper[4631]: I1128 13:53:03.071732 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-r2pq6" Nov 28 13:53:03 crc kubenswrapper[4631]: I1128 13:53:03.651669 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-r2pq6"] Nov 28 13:53:04 crc kubenswrapper[4631]: I1128 13:53:04.618962 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-r2pq6" event={"ID":"ec45d53a-f54c-422a-8a43-5b9d4b4ee0c6","Type":"ContainerStarted","Data":"ce9e395c727a7194bb0ec19b47aad7cf4f46b01df0334d0a4d2d6e4846b2d636"} Nov 28 13:53:04 crc kubenswrapper[4631]: I1128 13:53:04.619514 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-r2pq6" event={"ID":"ec45d53a-f54c-422a-8a43-5b9d4b4ee0c6","Type":"ContainerStarted","Data":"c468aea7433f38f34d9b879cb57e2d746e6f0ce2aec3e2beb97bd2e03ab883b2"} Nov 28 13:53:04 crc kubenswrapper[4631]: I1128 13:53:04.650124 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-r2pq6" podStartSLOduration=2.005172944 podStartE2EDuration="2.650095434s" podCreationTimestamp="2025-11-28 13:53:02 +0000 UTC" firstStartedPulling="2025-11-28 13:53:03.641817147 +0000 UTC m=+1940.449120491" lastFinishedPulling="2025-11-28 13:53:04.286739637 +0000 UTC m=+1941.094042981" observedRunningTime="2025-11-28 13:53:04.639009055 +0000 UTC m=+1941.446312409" watchObservedRunningTime="2025-11-28 13:53:04.650095434 +0000 UTC m=+1941.457398778" Nov 28 13:53:15 crc kubenswrapper[4631]: I1128 13:53:15.723207 4631 generic.go:334] "Generic (PLEG): container finished" podID="ec45d53a-f54c-422a-8a43-5b9d4b4ee0c6" containerID="ce9e395c727a7194bb0ec19b47aad7cf4f46b01df0334d0a4d2d6e4846b2d636" exitCode=0 Nov 28 13:53:15 crc kubenswrapper[4631]: I1128 13:53:15.723811 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-r2pq6" event={"ID":"ec45d53a-f54c-422a-8a43-5b9d4b4ee0c6","Type":"ContainerDied","Data":"ce9e395c727a7194bb0ec19b47aad7cf4f46b01df0334d0a4d2d6e4846b2d636"} Nov 28 13:53:17 crc kubenswrapper[4631]: I1128 13:53:17.183022 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-r2pq6" Nov 28 13:53:17 crc kubenswrapper[4631]: I1128 13:53:17.271598 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6chdq\" (UniqueName: \"kubernetes.io/projected/ec45d53a-f54c-422a-8a43-5b9d4b4ee0c6-kube-api-access-6chdq\") pod \"ec45d53a-f54c-422a-8a43-5b9d4b4ee0c6\" (UID: \"ec45d53a-f54c-422a-8a43-5b9d4b4ee0c6\") " Nov 28 13:53:17 crc kubenswrapper[4631]: I1128 13:53:17.272212 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ec45d53a-f54c-422a-8a43-5b9d4b4ee0c6-inventory\") pod \"ec45d53a-f54c-422a-8a43-5b9d4b4ee0c6\" (UID: \"ec45d53a-f54c-422a-8a43-5b9d4b4ee0c6\") " Nov 28 13:53:17 crc kubenswrapper[4631]: I1128 13:53:17.272365 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ec45d53a-f54c-422a-8a43-5b9d4b4ee0c6-ssh-key\") pod \"ec45d53a-f54c-422a-8a43-5b9d4b4ee0c6\" (UID: \"ec45d53a-f54c-422a-8a43-5b9d4b4ee0c6\") " Nov 28 13:53:17 crc kubenswrapper[4631]: I1128 13:53:17.277633 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ec45d53a-f54c-422a-8a43-5b9d4b4ee0c6-kube-api-access-6chdq" (OuterVolumeSpecName: "kube-api-access-6chdq") pod "ec45d53a-f54c-422a-8a43-5b9d4b4ee0c6" (UID: "ec45d53a-f54c-422a-8a43-5b9d4b4ee0c6"). InnerVolumeSpecName "kube-api-access-6chdq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:53:17 crc kubenswrapper[4631]: I1128 13:53:17.301158 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ec45d53a-f54c-422a-8a43-5b9d4b4ee0c6-inventory" (OuterVolumeSpecName: "inventory") pod "ec45d53a-f54c-422a-8a43-5b9d4b4ee0c6" (UID: "ec45d53a-f54c-422a-8a43-5b9d4b4ee0c6"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:53:17 crc kubenswrapper[4631]: I1128 13:53:17.310375 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ec45d53a-f54c-422a-8a43-5b9d4b4ee0c6-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "ec45d53a-f54c-422a-8a43-5b9d4b4ee0c6" (UID: "ec45d53a-f54c-422a-8a43-5b9d4b4ee0c6"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:53:17 crc kubenswrapper[4631]: I1128 13:53:17.374313 4631 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ec45d53a-f54c-422a-8a43-5b9d4b4ee0c6-inventory\") on node \"crc\" DevicePath \"\"" Nov 28 13:53:17 crc kubenswrapper[4631]: I1128 13:53:17.374611 4631 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ec45d53a-f54c-422a-8a43-5b9d4b4ee0c6-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 28 13:53:17 crc kubenswrapper[4631]: I1128 13:53:17.374681 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6chdq\" (UniqueName: \"kubernetes.io/projected/ec45d53a-f54c-422a-8a43-5b9d4b4ee0c6-kube-api-access-6chdq\") on node \"crc\" DevicePath \"\"" Nov 28 13:53:17 crc kubenswrapper[4631]: I1128 13:53:17.741252 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-r2pq6" event={"ID":"ec45d53a-f54c-422a-8a43-5b9d4b4ee0c6","Type":"ContainerDied","Data":"c468aea7433f38f34d9b879cb57e2d746e6f0ce2aec3e2beb97bd2e03ab883b2"} Nov 28 13:53:17 crc kubenswrapper[4631]: I1128 13:53:17.741313 4631 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c468aea7433f38f34d9b879cb57e2d746e6f0ce2aec3e2beb97bd2e03ab883b2" Nov 28 13:53:17 crc kubenswrapper[4631]: I1128 13:53:17.741349 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-r2pq6" Nov 28 13:53:17 crc kubenswrapper[4631]: I1128 13:53:17.883328 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-c6hq5"] Nov 28 13:53:17 crc kubenswrapper[4631]: E1128 13:53:17.884087 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ec45d53a-f54c-422a-8a43-5b9d4b4ee0c6" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Nov 28 13:53:17 crc kubenswrapper[4631]: I1128 13:53:17.884117 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="ec45d53a-f54c-422a-8a43-5b9d4b4ee0c6" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Nov 28 13:53:17 crc kubenswrapper[4631]: I1128 13:53:17.884405 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="ec45d53a-f54c-422a-8a43-5b9d4b4ee0c6" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Nov 28 13:53:17 crc kubenswrapper[4631]: I1128 13:53:17.885219 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-c6hq5" Nov 28 13:53:17 crc kubenswrapper[4631]: I1128 13:53:17.890838 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 28 13:53:17 crc kubenswrapper[4631]: I1128 13:53:17.891072 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 28 13:53:17 crc kubenswrapper[4631]: I1128 13:53:17.891111 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-neutron-metadata-default-certs-0" Nov 28 13:53:17 crc kubenswrapper[4631]: I1128 13:53:17.891719 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-ovn-default-certs-0" Nov 28 13:53:17 crc kubenswrapper[4631]: I1128 13:53:17.892161 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-vp4ns" Nov 28 13:53:17 crc kubenswrapper[4631]: I1128 13:53:17.892241 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-telemetry-default-certs-0" Nov 28 13:53:17 crc kubenswrapper[4631]: I1128 13:53:17.892324 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-libvirt-default-certs-0" Nov 28 13:53:17 crc kubenswrapper[4631]: I1128 13:53:17.892732 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 28 13:53:17 crc kubenswrapper[4631]: I1128 13:53:17.914304 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-c6hq5"] Nov 28 13:53:17 crc kubenswrapper[4631]: I1128 13:53:17.986977 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f7fda9a-fd1f-4485-a838-f5fee5a83b6e-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-c6hq5\" (UID: \"4f7fda9a-fd1f-4485-a838-f5fee5a83b6e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-c6hq5" Nov 28 13:53:17 crc kubenswrapper[4631]: I1128 13:53:17.987021 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/4f7fda9a-fd1f-4485-a838-f5fee5a83b6e-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-c6hq5\" (UID: \"4f7fda9a-fd1f-4485-a838-f5fee5a83b6e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-c6hq5" Nov 28 13:53:17 crc kubenswrapper[4631]: I1128 13:53:17.987048 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4f7fda9a-fd1f-4485-a838-f5fee5a83b6e-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-c6hq5\" (UID: \"4f7fda9a-fd1f-4485-a838-f5fee5a83b6e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-c6hq5" Nov 28 13:53:17 crc kubenswrapper[4631]: I1128 13:53:17.987078 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f7fda9a-fd1f-4485-a838-f5fee5a83b6e-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-c6hq5\" (UID: \"4f7fda9a-fd1f-4485-a838-f5fee5a83b6e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-c6hq5" Nov 28 13:53:17 crc kubenswrapper[4631]: I1128 13:53:17.987255 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f7fda9a-fd1f-4485-a838-f5fee5a83b6e-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-c6hq5\" (UID: \"4f7fda9a-fd1f-4485-a838-f5fee5a83b6e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-c6hq5" Nov 28 13:53:17 crc kubenswrapper[4631]: I1128 13:53:17.987383 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/4f7fda9a-fd1f-4485-a838-f5fee5a83b6e-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-c6hq5\" (UID: \"4f7fda9a-fd1f-4485-a838-f5fee5a83b6e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-c6hq5" Nov 28 13:53:17 crc kubenswrapper[4631]: I1128 13:53:17.987445 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/4f7fda9a-fd1f-4485-a838-f5fee5a83b6e-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-c6hq5\" (UID: \"4f7fda9a-fd1f-4485-a838-f5fee5a83b6e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-c6hq5" Nov 28 13:53:17 crc kubenswrapper[4631]: I1128 13:53:17.987498 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4f7fda9a-fd1f-4485-a838-f5fee5a83b6e-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-c6hq5\" (UID: \"4f7fda9a-fd1f-4485-a838-f5fee5a83b6e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-c6hq5" Nov 28 13:53:17 crc kubenswrapper[4631]: I1128 13:53:17.987568 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f7fda9a-fd1f-4485-a838-f5fee5a83b6e-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-c6hq5\" (UID: \"4f7fda9a-fd1f-4485-a838-f5fee5a83b6e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-c6hq5" Nov 28 13:53:17 crc kubenswrapper[4631]: I1128 13:53:17.987673 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f7fda9a-fd1f-4485-a838-f5fee5a83b6e-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-c6hq5\" (UID: \"4f7fda9a-fd1f-4485-a838-f5fee5a83b6e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-c6hq5" Nov 28 13:53:17 crc kubenswrapper[4631]: I1128 13:53:17.987736 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/4f7fda9a-fd1f-4485-a838-f5fee5a83b6e-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-c6hq5\" (UID: \"4f7fda9a-fd1f-4485-a838-f5fee5a83b6e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-c6hq5" Nov 28 13:53:17 crc kubenswrapper[4631]: I1128 13:53:17.987841 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f7fda9a-fd1f-4485-a838-f5fee5a83b6e-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-c6hq5\" (UID: \"4f7fda9a-fd1f-4485-a838-f5fee5a83b6e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-c6hq5" Nov 28 13:53:17 crc kubenswrapper[4631]: I1128 13:53:17.987891 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f7fda9a-fd1f-4485-a838-f5fee5a83b6e-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-c6hq5\" (UID: \"4f7fda9a-fd1f-4485-a838-f5fee5a83b6e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-c6hq5" Nov 28 13:53:17 crc kubenswrapper[4631]: I1128 13:53:17.987931 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n8n4d\" (UniqueName: \"kubernetes.io/projected/4f7fda9a-fd1f-4485-a838-f5fee5a83b6e-kube-api-access-n8n4d\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-c6hq5\" (UID: \"4f7fda9a-fd1f-4485-a838-f5fee5a83b6e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-c6hq5" Nov 28 13:53:18 crc kubenswrapper[4631]: I1128 13:53:18.090107 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n8n4d\" (UniqueName: \"kubernetes.io/projected/4f7fda9a-fd1f-4485-a838-f5fee5a83b6e-kube-api-access-n8n4d\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-c6hq5\" (UID: \"4f7fda9a-fd1f-4485-a838-f5fee5a83b6e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-c6hq5" Nov 28 13:53:18 crc kubenswrapper[4631]: I1128 13:53:18.090243 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f7fda9a-fd1f-4485-a838-f5fee5a83b6e-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-c6hq5\" (UID: \"4f7fda9a-fd1f-4485-a838-f5fee5a83b6e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-c6hq5" Nov 28 13:53:18 crc kubenswrapper[4631]: I1128 13:53:18.090268 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/4f7fda9a-fd1f-4485-a838-f5fee5a83b6e-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-c6hq5\" (UID: \"4f7fda9a-fd1f-4485-a838-f5fee5a83b6e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-c6hq5" Nov 28 13:53:18 crc kubenswrapper[4631]: I1128 13:53:18.090321 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4f7fda9a-fd1f-4485-a838-f5fee5a83b6e-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-c6hq5\" (UID: \"4f7fda9a-fd1f-4485-a838-f5fee5a83b6e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-c6hq5" Nov 28 13:53:18 crc kubenswrapper[4631]: I1128 13:53:18.090354 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f7fda9a-fd1f-4485-a838-f5fee5a83b6e-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-c6hq5\" (UID: \"4f7fda9a-fd1f-4485-a838-f5fee5a83b6e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-c6hq5" Nov 28 13:53:18 crc kubenswrapper[4631]: I1128 13:53:18.090384 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f7fda9a-fd1f-4485-a838-f5fee5a83b6e-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-c6hq5\" (UID: \"4f7fda9a-fd1f-4485-a838-f5fee5a83b6e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-c6hq5" Nov 28 13:53:18 crc kubenswrapper[4631]: I1128 13:53:18.090423 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/4f7fda9a-fd1f-4485-a838-f5fee5a83b6e-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-c6hq5\" (UID: \"4f7fda9a-fd1f-4485-a838-f5fee5a83b6e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-c6hq5" Nov 28 13:53:18 crc kubenswrapper[4631]: I1128 13:53:18.090450 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/4f7fda9a-fd1f-4485-a838-f5fee5a83b6e-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-c6hq5\" (UID: \"4f7fda9a-fd1f-4485-a838-f5fee5a83b6e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-c6hq5" Nov 28 13:53:18 crc kubenswrapper[4631]: I1128 13:53:18.090479 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4f7fda9a-fd1f-4485-a838-f5fee5a83b6e-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-c6hq5\" (UID: \"4f7fda9a-fd1f-4485-a838-f5fee5a83b6e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-c6hq5" Nov 28 13:53:18 crc kubenswrapper[4631]: I1128 13:53:18.090506 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f7fda9a-fd1f-4485-a838-f5fee5a83b6e-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-c6hq5\" (UID: \"4f7fda9a-fd1f-4485-a838-f5fee5a83b6e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-c6hq5" Nov 28 13:53:18 crc kubenswrapper[4631]: I1128 13:53:18.090552 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f7fda9a-fd1f-4485-a838-f5fee5a83b6e-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-c6hq5\" (UID: \"4f7fda9a-fd1f-4485-a838-f5fee5a83b6e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-c6hq5" Nov 28 13:53:18 crc kubenswrapper[4631]: I1128 13:53:18.090574 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/4f7fda9a-fd1f-4485-a838-f5fee5a83b6e-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-c6hq5\" (UID: \"4f7fda9a-fd1f-4485-a838-f5fee5a83b6e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-c6hq5" Nov 28 13:53:18 crc kubenswrapper[4631]: I1128 13:53:18.090605 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f7fda9a-fd1f-4485-a838-f5fee5a83b6e-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-c6hq5\" (UID: \"4f7fda9a-fd1f-4485-a838-f5fee5a83b6e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-c6hq5" Nov 28 13:53:18 crc kubenswrapper[4631]: I1128 13:53:18.090628 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f7fda9a-fd1f-4485-a838-f5fee5a83b6e-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-c6hq5\" (UID: \"4f7fda9a-fd1f-4485-a838-f5fee5a83b6e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-c6hq5" Nov 28 13:53:18 crc kubenswrapper[4631]: I1128 13:53:18.097326 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4f7fda9a-fd1f-4485-a838-f5fee5a83b6e-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-c6hq5\" (UID: \"4f7fda9a-fd1f-4485-a838-f5fee5a83b6e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-c6hq5" Nov 28 13:53:18 crc kubenswrapper[4631]: I1128 13:53:18.097895 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/4f7fda9a-fd1f-4485-a838-f5fee5a83b6e-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-c6hq5\" (UID: \"4f7fda9a-fd1f-4485-a838-f5fee5a83b6e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-c6hq5" Nov 28 13:53:18 crc kubenswrapper[4631]: I1128 13:53:18.098415 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f7fda9a-fd1f-4485-a838-f5fee5a83b6e-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-c6hq5\" (UID: \"4f7fda9a-fd1f-4485-a838-f5fee5a83b6e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-c6hq5" Nov 28 13:53:18 crc kubenswrapper[4631]: I1128 13:53:18.098595 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f7fda9a-fd1f-4485-a838-f5fee5a83b6e-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-c6hq5\" (UID: \"4f7fda9a-fd1f-4485-a838-f5fee5a83b6e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-c6hq5" Nov 28 13:53:18 crc kubenswrapper[4631]: I1128 13:53:18.098770 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/4f7fda9a-fd1f-4485-a838-f5fee5a83b6e-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-c6hq5\" (UID: \"4f7fda9a-fd1f-4485-a838-f5fee5a83b6e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-c6hq5" Nov 28 13:53:18 crc kubenswrapper[4631]: I1128 13:53:18.097972 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/4f7fda9a-fd1f-4485-a838-f5fee5a83b6e-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-c6hq5\" (UID: \"4f7fda9a-fd1f-4485-a838-f5fee5a83b6e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-c6hq5" Nov 28 13:53:18 crc kubenswrapper[4631]: I1128 13:53:18.102000 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f7fda9a-fd1f-4485-a838-f5fee5a83b6e-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-c6hq5\" (UID: \"4f7fda9a-fd1f-4485-a838-f5fee5a83b6e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-c6hq5" Nov 28 13:53:18 crc kubenswrapper[4631]: I1128 13:53:18.102849 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f7fda9a-fd1f-4485-a838-f5fee5a83b6e-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-c6hq5\" (UID: \"4f7fda9a-fd1f-4485-a838-f5fee5a83b6e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-c6hq5" Nov 28 13:53:18 crc kubenswrapper[4631]: I1128 13:53:18.106447 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f7fda9a-fd1f-4485-a838-f5fee5a83b6e-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-c6hq5\" (UID: \"4f7fda9a-fd1f-4485-a838-f5fee5a83b6e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-c6hq5" Nov 28 13:53:18 crc kubenswrapper[4631]: I1128 13:53:18.107814 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f7fda9a-fd1f-4485-a838-f5fee5a83b6e-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-c6hq5\" (UID: \"4f7fda9a-fd1f-4485-a838-f5fee5a83b6e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-c6hq5" Nov 28 13:53:18 crc kubenswrapper[4631]: I1128 13:53:18.115232 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f7fda9a-fd1f-4485-a838-f5fee5a83b6e-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-c6hq5\" (UID: \"4f7fda9a-fd1f-4485-a838-f5fee5a83b6e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-c6hq5" Nov 28 13:53:18 crc kubenswrapper[4631]: I1128 13:53:18.123052 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4f7fda9a-fd1f-4485-a838-f5fee5a83b6e-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-c6hq5\" (UID: \"4f7fda9a-fd1f-4485-a838-f5fee5a83b6e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-c6hq5" Nov 28 13:53:18 crc kubenswrapper[4631]: I1128 13:53:18.123431 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/4f7fda9a-fd1f-4485-a838-f5fee5a83b6e-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-c6hq5\" (UID: \"4f7fda9a-fd1f-4485-a838-f5fee5a83b6e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-c6hq5" Nov 28 13:53:18 crc kubenswrapper[4631]: I1128 13:53:18.123809 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n8n4d\" (UniqueName: \"kubernetes.io/projected/4f7fda9a-fd1f-4485-a838-f5fee5a83b6e-kube-api-access-n8n4d\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-c6hq5\" (UID: \"4f7fda9a-fd1f-4485-a838-f5fee5a83b6e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-c6hq5" Nov 28 13:53:18 crc kubenswrapper[4631]: I1128 13:53:18.204519 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-c6hq5" Nov 28 13:53:18 crc kubenswrapper[4631]: I1128 13:53:18.713961 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-c6hq5"] Nov 28 13:53:18 crc kubenswrapper[4631]: I1128 13:53:18.751252 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-c6hq5" event={"ID":"4f7fda9a-fd1f-4485-a838-f5fee5a83b6e","Type":"ContainerStarted","Data":"0a2d184ce996de7423208eb0d3001a83c20fde01c1b0d7876d1d7d6a5a39dff9"} Nov 28 13:53:19 crc kubenswrapper[4631]: I1128 13:53:19.763716 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-c6hq5" event={"ID":"4f7fda9a-fd1f-4485-a838-f5fee5a83b6e","Type":"ContainerStarted","Data":"c01b3a41d3dc956637eb574fa941b1b5e0d9ebbced94ff53aa18d1bf55636321"} Nov 28 13:53:19 crc kubenswrapper[4631]: I1128 13:53:19.794083 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-c6hq5" podStartSLOduration=2.20991934 podStartE2EDuration="2.794056815s" podCreationTimestamp="2025-11-28 13:53:17 +0000 UTC" firstStartedPulling="2025-11-28 13:53:18.716985518 +0000 UTC m=+1955.524288862" lastFinishedPulling="2025-11-28 13:53:19.301122993 +0000 UTC m=+1956.108426337" observedRunningTime="2025-11-28 13:53:19.786951972 +0000 UTC m=+1956.594255316" watchObservedRunningTime="2025-11-28 13:53:19.794056815 +0000 UTC m=+1956.601360159" Nov 28 13:54:00 crc kubenswrapper[4631]: I1128 13:54:00.134437 4631 generic.go:334] "Generic (PLEG): container finished" podID="4f7fda9a-fd1f-4485-a838-f5fee5a83b6e" containerID="c01b3a41d3dc956637eb574fa941b1b5e0d9ebbced94ff53aa18d1bf55636321" exitCode=0 Nov 28 13:54:00 crc kubenswrapper[4631]: I1128 13:54:00.134627 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-c6hq5" event={"ID":"4f7fda9a-fd1f-4485-a838-f5fee5a83b6e","Type":"ContainerDied","Data":"c01b3a41d3dc956637eb574fa941b1b5e0d9ebbced94ff53aa18d1bf55636321"} Nov 28 13:54:01 crc kubenswrapper[4631]: I1128 13:54:01.579825 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-c6hq5" Nov 28 13:54:01 crc kubenswrapper[4631]: I1128 13:54:01.715772 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/4f7fda9a-fd1f-4485-a838-f5fee5a83b6e-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"4f7fda9a-fd1f-4485-a838-f5fee5a83b6e\" (UID: \"4f7fda9a-fd1f-4485-a838-f5fee5a83b6e\") " Nov 28 13:54:01 crc kubenswrapper[4631]: I1128 13:54:01.715819 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f7fda9a-fd1f-4485-a838-f5fee5a83b6e-telemetry-combined-ca-bundle\") pod \"4f7fda9a-fd1f-4485-a838-f5fee5a83b6e\" (UID: \"4f7fda9a-fd1f-4485-a838-f5fee5a83b6e\") " Nov 28 13:54:01 crc kubenswrapper[4631]: I1128 13:54:01.715931 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f7fda9a-fd1f-4485-a838-f5fee5a83b6e-bootstrap-combined-ca-bundle\") pod \"4f7fda9a-fd1f-4485-a838-f5fee5a83b6e\" (UID: \"4f7fda9a-fd1f-4485-a838-f5fee5a83b6e\") " Nov 28 13:54:01 crc kubenswrapper[4631]: I1128 13:54:01.715984 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/4f7fda9a-fd1f-4485-a838-f5fee5a83b6e-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"4f7fda9a-fd1f-4485-a838-f5fee5a83b6e\" (UID: \"4f7fda9a-fd1f-4485-a838-f5fee5a83b6e\") " Nov 28 13:54:01 crc kubenswrapper[4631]: I1128 13:54:01.716019 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f7fda9a-fd1f-4485-a838-f5fee5a83b6e-libvirt-combined-ca-bundle\") pod \"4f7fda9a-fd1f-4485-a838-f5fee5a83b6e\" (UID: \"4f7fda9a-fd1f-4485-a838-f5fee5a83b6e\") " Nov 28 13:54:01 crc kubenswrapper[4631]: I1128 13:54:01.716046 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4f7fda9a-fd1f-4485-a838-f5fee5a83b6e-inventory\") pod \"4f7fda9a-fd1f-4485-a838-f5fee5a83b6e\" (UID: \"4f7fda9a-fd1f-4485-a838-f5fee5a83b6e\") " Nov 28 13:54:01 crc kubenswrapper[4631]: I1128 13:54:01.716094 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n8n4d\" (UniqueName: \"kubernetes.io/projected/4f7fda9a-fd1f-4485-a838-f5fee5a83b6e-kube-api-access-n8n4d\") pod \"4f7fda9a-fd1f-4485-a838-f5fee5a83b6e\" (UID: \"4f7fda9a-fd1f-4485-a838-f5fee5a83b6e\") " Nov 28 13:54:01 crc kubenswrapper[4631]: I1128 13:54:01.716121 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f7fda9a-fd1f-4485-a838-f5fee5a83b6e-neutron-metadata-combined-ca-bundle\") pod \"4f7fda9a-fd1f-4485-a838-f5fee5a83b6e\" (UID: \"4f7fda9a-fd1f-4485-a838-f5fee5a83b6e\") " Nov 28 13:54:01 crc kubenswrapper[4631]: I1128 13:54:01.716159 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4f7fda9a-fd1f-4485-a838-f5fee5a83b6e-ssh-key\") pod \"4f7fda9a-fd1f-4485-a838-f5fee5a83b6e\" (UID: \"4f7fda9a-fd1f-4485-a838-f5fee5a83b6e\") " Nov 28 13:54:01 crc kubenswrapper[4631]: I1128 13:54:01.716190 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/4f7fda9a-fd1f-4485-a838-f5fee5a83b6e-openstack-edpm-ipam-ovn-default-certs-0\") pod \"4f7fda9a-fd1f-4485-a838-f5fee5a83b6e\" (UID: \"4f7fda9a-fd1f-4485-a838-f5fee5a83b6e\") " Nov 28 13:54:01 crc kubenswrapper[4631]: I1128 13:54:01.716250 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f7fda9a-fd1f-4485-a838-f5fee5a83b6e-repo-setup-combined-ca-bundle\") pod \"4f7fda9a-fd1f-4485-a838-f5fee5a83b6e\" (UID: \"4f7fda9a-fd1f-4485-a838-f5fee5a83b6e\") " Nov 28 13:54:01 crc kubenswrapper[4631]: I1128 13:54:01.716355 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/4f7fda9a-fd1f-4485-a838-f5fee5a83b6e-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"4f7fda9a-fd1f-4485-a838-f5fee5a83b6e\" (UID: \"4f7fda9a-fd1f-4485-a838-f5fee5a83b6e\") " Nov 28 13:54:01 crc kubenswrapper[4631]: I1128 13:54:01.716403 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f7fda9a-fd1f-4485-a838-f5fee5a83b6e-nova-combined-ca-bundle\") pod \"4f7fda9a-fd1f-4485-a838-f5fee5a83b6e\" (UID: \"4f7fda9a-fd1f-4485-a838-f5fee5a83b6e\") " Nov 28 13:54:01 crc kubenswrapper[4631]: I1128 13:54:01.716481 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f7fda9a-fd1f-4485-a838-f5fee5a83b6e-ovn-combined-ca-bundle\") pod \"4f7fda9a-fd1f-4485-a838-f5fee5a83b6e\" (UID: \"4f7fda9a-fd1f-4485-a838-f5fee5a83b6e\") " Nov 28 13:54:01 crc kubenswrapper[4631]: I1128 13:54:01.722378 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4f7fda9a-fd1f-4485-a838-f5fee5a83b6e-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "4f7fda9a-fd1f-4485-a838-f5fee5a83b6e" (UID: "4f7fda9a-fd1f-4485-a838-f5fee5a83b6e"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:54:01 crc kubenswrapper[4631]: I1128 13:54:01.723891 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4f7fda9a-fd1f-4485-a838-f5fee5a83b6e-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "4f7fda9a-fd1f-4485-a838-f5fee5a83b6e" (UID: "4f7fda9a-fd1f-4485-a838-f5fee5a83b6e"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:54:01 crc kubenswrapper[4631]: I1128 13:54:01.723986 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4f7fda9a-fd1f-4485-a838-f5fee5a83b6e-kube-api-access-n8n4d" (OuterVolumeSpecName: "kube-api-access-n8n4d") pod "4f7fda9a-fd1f-4485-a838-f5fee5a83b6e" (UID: "4f7fda9a-fd1f-4485-a838-f5fee5a83b6e"). InnerVolumeSpecName "kube-api-access-n8n4d". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:54:01 crc kubenswrapper[4631]: I1128 13:54:01.725270 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4f7fda9a-fd1f-4485-a838-f5fee5a83b6e-openstack-edpm-ipam-telemetry-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-telemetry-default-certs-0") pod "4f7fda9a-fd1f-4485-a838-f5fee5a83b6e" (UID: "4f7fda9a-fd1f-4485-a838-f5fee5a83b6e"). InnerVolumeSpecName "openstack-edpm-ipam-telemetry-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:54:01 crc kubenswrapper[4631]: I1128 13:54:01.725417 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4f7fda9a-fd1f-4485-a838-f5fee5a83b6e-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "4f7fda9a-fd1f-4485-a838-f5fee5a83b6e" (UID: "4f7fda9a-fd1f-4485-a838-f5fee5a83b6e"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:54:01 crc kubenswrapper[4631]: I1128 13:54:01.727161 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4f7fda9a-fd1f-4485-a838-f5fee5a83b6e-openstack-edpm-ipam-ovn-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-ovn-default-certs-0") pod "4f7fda9a-fd1f-4485-a838-f5fee5a83b6e" (UID: "4f7fda9a-fd1f-4485-a838-f5fee5a83b6e"). InnerVolumeSpecName "openstack-edpm-ipam-ovn-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:54:01 crc kubenswrapper[4631]: I1128 13:54:01.727454 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4f7fda9a-fd1f-4485-a838-f5fee5a83b6e-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "4f7fda9a-fd1f-4485-a838-f5fee5a83b6e" (UID: "4f7fda9a-fd1f-4485-a838-f5fee5a83b6e"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:54:01 crc kubenswrapper[4631]: I1128 13:54:01.728086 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4f7fda9a-fd1f-4485-a838-f5fee5a83b6e-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "4f7fda9a-fd1f-4485-a838-f5fee5a83b6e" (UID: "4f7fda9a-fd1f-4485-a838-f5fee5a83b6e"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:54:01 crc kubenswrapper[4631]: I1128 13:54:01.731140 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4f7fda9a-fd1f-4485-a838-f5fee5a83b6e-openstack-edpm-ipam-neutron-metadata-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-neutron-metadata-default-certs-0") pod "4f7fda9a-fd1f-4485-a838-f5fee5a83b6e" (UID: "4f7fda9a-fd1f-4485-a838-f5fee5a83b6e"). InnerVolumeSpecName "openstack-edpm-ipam-neutron-metadata-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:54:01 crc kubenswrapper[4631]: I1128 13:54:01.731704 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4f7fda9a-fd1f-4485-a838-f5fee5a83b6e-openstack-edpm-ipam-libvirt-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-libvirt-default-certs-0") pod "4f7fda9a-fd1f-4485-a838-f5fee5a83b6e" (UID: "4f7fda9a-fd1f-4485-a838-f5fee5a83b6e"). InnerVolumeSpecName "openstack-edpm-ipam-libvirt-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:54:01 crc kubenswrapper[4631]: I1128 13:54:01.732211 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4f7fda9a-fd1f-4485-a838-f5fee5a83b6e-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "4f7fda9a-fd1f-4485-a838-f5fee5a83b6e" (UID: "4f7fda9a-fd1f-4485-a838-f5fee5a83b6e"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:54:01 crc kubenswrapper[4631]: I1128 13:54:01.732458 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4f7fda9a-fd1f-4485-a838-f5fee5a83b6e-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "4f7fda9a-fd1f-4485-a838-f5fee5a83b6e" (UID: "4f7fda9a-fd1f-4485-a838-f5fee5a83b6e"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:54:01 crc kubenswrapper[4631]: I1128 13:54:01.758679 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4f7fda9a-fd1f-4485-a838-f5fee5a83b6e-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "4f7fda9a-fd1f-4485-a838-f5fee5a83b6e" (UID: "4f7fda9a-fd1f-4485-a838-f5fee5a83b6e"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:54:01 crc kubenswrapper[4631]: I1128 13:54:01.762740 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4f7fda9a-fd1f-4485-a838-f5fee5a83b6e-inventory" (OuterVolumeSpecName: "inventory") pod "4f7fda9a-fd1f-4485-a838-f5fee5a83b6e" (UID: "4f7fda9a-fd1f-4485-a838-f5fee5a83b6e"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:54:01 crc kubenswrapper[4631]: I1128 13:54:01.819248 4631 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f7fda9a-fd1f-4485-a838-f5fee5a83b6e-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 13:54:01 crc kubenswrapper[4631]: I1128 13:54:01.819452 4631 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/4f7fda9a-fd1f-4485-a838-f5fee5a83b6e-openstack-edpm-ipam-telemetry-default-certs-0\") on node \"crc\" DevicePath \"\"" Nov 28 13:54:01 crc kubenswrapper[4631]: I1128 13:54:01.819470 4631 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f7fda9a-fd1f-4485-a838-f5fee5a83b6e-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 13:54:01 crc kubenswrapper[4631]: I1128 13:54:01.819486 4631 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4f7fda9a-fd1f-4485-a838-f5fee5a83b6e-inventory\") on node \"crc\" DevicePath \"\"" Nov 28 13:54:01 crc kubenswrapper[4631]: I1128 13:54:01.819498 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n8n4d\" (UniqueName: \"kubernetes.io/projected/4f7fda9a-fd1f-4485-a838-f5fee5a83b6e-kube-api-access-n8n4d\") on node \"crc\" DevicePath \"\"" Nov 28 13:54:01 crc kubenswrapper[4631]: I1128 13:54:01.819509 4631 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f7fda9a-fd1f-4485-a838-f5fee5a83b6e-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 13:54:01 crc kubenswrapper[4631]: I1128 13:54:01.819522 4631 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4f7fda9a-fd1f-4485-a838-f5fee5a83b6e-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 28 13:54:01 crc kubenswrapper[4631]: I1128 13:54:01.819534 4631 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/4f7fda9a-fd1f-4485-a838-f5fee5a83b6e-openstack-edpm-ipam-ovn-default-certs-0\") on node \"crc\" DevicePath \"\"" Nov 28 13:54:01 crc kubenswrapper[4631]: I1128 13:54:01.819553 4631 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f7fda9a-fd1f-4485-a838-f5fee5a83b6e-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 13:54:01 crc kubenswrapper[4631]: I1128 13:54:01.819574 4631 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/4f7fda9a-fd1f-4485-a838-f5fee5a83b6e-openstack-edpm-ipam-libvirt-default-certs-0\") on node \"crc\" DevicePath \"\"" Nov 28 13:54:01 crc kubenswrapper[4631]: I1128 13:54:01.819587 4631 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f7fda9a-fd1f-4485-a838-f5fee5a83b6e-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 13:54:01 crc kubenswrapper[4631]: I1128 13:54:01.819600 4631 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f7fda9a-fd1f-4485-a838-f5fee5a83b6e-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 13:54:01 crc kubenswrapper[4631]: I1128 13:54:01.819611 4631 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/4f7fda9a-fd1f-4485-a838-f5fee5a83b6e-openstack-edpm-ipam-neutron-metadata-default-certs-0\") on node \"crc\" DevicePath \"\"" Nov 28 13:54:01 crc kubenswrapper[4631]: I1128 13:54:01.819641 4631 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f7fda9a-fd1f-4485-a838-f5fee5a83b6e-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 13:54:02 crc kubenswrapper[4631]: I1128 13:54:02.155218 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-c6hq5" event={"ID":"4f7fda9a-fd1f-4485-a838-f5fee5a83b6e","Type":"ContainerDied","Data":"0a2d184ce996de7423208eb0d3001a83c20fde01c1b0d7876d1d7d6a5a39dff9"} Nov 28 13:54:02 crc kubenswrapper[4631]: I1128 13:54:02.155307 4631 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0a2d184ce996de7423208eb0d3001a83c20fde01c1b0d7876d1d7d6a5a39dff9" Nov 28 13:54:02 crc kubenswrapper[4631]: I1128 13:54:02.155322 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-c6hq5" Nov 28 13:54:02 crc kubenswrapper[4631]: I1128 13:54:02.276046 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-2sls9"] Nov 28 13:54:02 crc kubenswrapper[4631]: E1128 13:54:02.276553 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4f7fda9a-fd1f-4485-a838-f5fee5a83b6e" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Nov 28 13:54:02 crc kubenswrapper[4631]: I1128 13:54:02.276578 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="4f7fda9a-fd1f-4485-a838-f5fee5a83b6e" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Nov 28 13:54:02 crc kubenswrapper[4631]: I1128 13:54:02.276798 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="4f7fda9a-fd1f-4485-a838-f5fee5a83b6e" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Nov 28 13:54:02 crc kubenswrapper[4631]: I1128 13:54:02.277808 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-2sls9" Nov 28 13:54:02 crc kubenswrapper[4631]: I1128 13:54:02.283531 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 28 13:54:02 crc kubenswrapper[4631]: I1128 13:54:02.283775 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-vp4ns" Nov 28 13:54:02 crc kubenswrapper[4631]: I1128 13:54:02.283931 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 28 13:54:02 crc kubenswrapper[4631]: I1128 13:54:02.284057 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-config" Nov 28 13:54:02 crc kubenswrapper[4631]: I1128 13:54:02.284222 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 28 13:54:02 crc kubenswrapper[4631]: I1128 13:54:02.294410 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-2sls9"] Nov 28 13:54:02 crc kubenswrapper[4631]: I1128 13:54:02.430745 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c09a85dc-b780-41c9-8122-653d4703aa7e-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-2sls9\" (UID: \"c09a85dc-b780-41c9-8122-653d4703aa7e\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-2sls9" Nov 28 13:54:02 crc kubenswrapper[4631]: I1128 13:54:02.430833 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/c09a85dc-b780-41c9-8122-653d4703aa7e-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-2sls9\" (UID: \"c09a85dc-b780-41c9-8122-653d4703aa7e\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-2sls9" Nov 28 13:54:02 crc kubenswrapper[4631]: I1128 13:54:02.431234 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c09a85dc-b780-41c9-8122-653d4703aa7e-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-2sls9\" (UID: \"c09a85dc-b780-41c9-8122-653d4703aa7e\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-2sls9" Nov 28 13:54:02 crc kubenswrapper[4631]: I1128 13:54:02.431397 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9rlbm\" (UniqueName: \"kubernetes.io/projected/c09a85dc-b780-41c9-8122-653d4703aa7e-kube-api-access-9rlbm\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-2sls9\" (UID: \"c09a85dc-b780-41c9-8122-653d4703aa7e\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-2sls9" Nov 28 13:54:02 crc kubenswrapper[4631]: I1128 13:54:02.431623 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c09a85dc-b780-41c9-8122-653d4703aa7e-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-2sls9\" (UID: \"c09a85dc-b780-41c9-8122-653d4703aa7e\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-2sls9" Nov 28 13:54:02 crc kubenswrapper[4631]: I1128 13:54:02.533755 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c09a85dc-b780-41c9-8122-653d4703aa7e-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-2sls9\" (UID: \"c09a85dc-b780-41c9-8122-653d4703aa7e\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-2sls9" Nov 28 13:54:02 crc kubenswrapper[4631]: I1128 13:54:02.533825 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9rlbm\" (UniqueName: \"kubernetes.io/projected/c09a85dc-b780-41c9-8122-653d4703aa7e-kube-api-access-9rlbm\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-2sls9\" (UID: \"c09a85dc-b780-41c9-8122-653d4703aa7e\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-2sls9" Nov 28 13:54:02 crc kubenswrapper[4631]: I1128 13:54:02.533892 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c09a85dc-b780-41c9-8122-653d4703aa7e-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-2sls9\" (UID: \"c09a85dc-b780-41c9-8122-653d4703aa7e\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-2sls9" Nov 28 13:54:02 crc kubenswrapper[4631]: I1128 13:54:02.533915 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c09a85dc-b780-41c9-8122-653d4703aa7e-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-2sls9\" (UID: \"c09a85dc-b780-41c9-8122-653d4703aa7e\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-2sls9" Nov 28 13:54:02 crc kubenswrapper[4631]: I1128 13:54:02.533947 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/c09a85dc-b780-41c9-8122-653d4703aa7e-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-2sls9\" (UID: \"c09a85dc-b780-41c9-8122-653d4703aa7e\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-2sls9" Nov 28 13:54:02 crc kubenswrapper[4631]: I1128 13:54:02.535303 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/c09a85dc-b780-41c9-8122-653d4703aa7e-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-2sls9\" (UID: \"c09a85dc-b780-41c9-8122-653d4703aa7e\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-2sls9" Nov 28 13:54:02 crc kubenswrapper[4631]: I1128 13:54:02.538054 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c09a85dc-b780-41c9-8122-653d4703aa7e-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-2sls9\" (UID: \"c09a85dc-b780-41c9-8122-653d4703aa7e\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-2sls9" Nov 28 13:54:02 crc kubenswrapper[4631]: I1128 13:54:02.538142 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c09a85dc-b780-41c9-8122-653d4703aa7e-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-2sls9\" (UID: \"c09a85dc-b780-41c9-8122-653d4703aa7e\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-2sls9" Nov 28 13:54:02 crc kubenswrapper[4631]: I1128 13:54:02.547833 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c09a85dc-b780-41c9-8122-653d4703aa7e-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-2sls9\" (UID: \"c09a85dc-b780-41c9-8122-653d4703aa7e\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-2sls9" Nov 28 13:54:02 crc kubenswrapper[4631]: I1128 13:54:02.553551 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9rlbm\" (UniqueName: \"kubernetes.io/projected/c09a85dc-b780-41c9-8122-653d4703aa7e-kube-api-access-9rlbm\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-2sls9\" (UID: \"c09a85dc-b780-41c9-8122-653d4703aa7e\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-2sls9" Nov 28 13:54:02 crc kubenswrapper[4631]: I1128 13:54:02.600349 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-2sls9" Nov 28 13:54:03 crc kubenswrapper[4631]: I1128 13:54:03.152457 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-2sls9"] Nov 28 13:54:04 crc kubenswrapper[4631]: I1128 13:54:04.184353 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-2sls9" event={"ID":"c09a85dc-b780-41c9-8122-653d4703aa7e","Type":"ContainerStarted","Data":"639dacf5cffcf75693e5eac449744cff4ce4a0729c1f60d3961fd670a60f79bd"} Nov 28 13:54:04 crc kubenswrapper[4631]: I1128 13:54:04.184738 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-2sls9" event={"ID":"c09a85dc-b780-41c9-8122-653d4703aa7e","Type":"ContainerStarted","Data":"bd8114646555df58b772ae35891365b0e848a55f2e996bddf3e0b5e684939c38"} Nov 28 13:54:04 crc kubenswrapper[4631]: I1128 13:54:04.205802 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-2sls9" podStartSLOduration=1.652140315 podStartE2EDuration="2.2057853s" podCreationTimestamp="2025-11-28 13:54:02 +0000 UTC" firstStartedPulling="2025-11-28 13:54:03.168889888 +0000 UTC m=+1999.976193232" lastFinishedPulling="2025-11-28 13:54:03.722534873 +0000 UTC m=+2000.529838217" observedRunningTime="2025-11-28 13:54:04.202973162 +0000 UTC m=+2001.010276506" watchObservedRunningTime="2025-11-28 13:54:04.2057853 +0000 UTC m=+2001.013088644" Nov 28 13:54:46 crc kubenswrapper[4631]: I1128 13:54:46.054473 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-7l8hn"] Nov 28 13:54:46 crc kubenswrapper[4631]: I1128 13:54:46.057751 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7l8hn" Nov 28 13:54:46 crc kubenswrapper[4631]: I1128 13:54:46.077587 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-7l8hn"] Nov 28 13:54:46 crc kubenswrapper[4631]: I1128 13:54:46.250775 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5fecbfbb-c1df-4171-bd01-d9d9a036cbc2-utilities\") pod \"redhat-operators-7l8hn\" (UID: \"5fecbfbb-c1df-4171-bd01-d9d9a036cbc2\") " pod="openshift-marketplace/redhat-operators-7l8hn" Nov 28 13:54:46 crc kubenswrapper[4631]: I1128 13:54:46.250828 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5fecbfbb-c1df-4171-bd01-d9d9a036cbc2-catalog-content\") pod \"redhat-operators-7l8hn\" (UID: \"5fecbfbb-c1df-4171-bd01-d9d9a036cbc2\") " pod="openshift-marketplace/redhat-operators-7l8hn" Nov 28 13:54:46 crc kubenswrapper[4631]: I1128 13:54:46.250855 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b449m\" (UniqueName: \"kubernetes.io/projected/5fecbfbb-c1df-4171-bd01-d9d9a036cbc2-kube-api-access-b449m\") pod \"redhat-operators-7l8hn\" (UID: \"5fecbfbb-c1df-4171-bd01-d9d9a036cbc2\") " pod="openshift-marketplace/redhat-operators-7l8hn" Nov 28 13:54:46 crc kubenswrapper[4631]: I1128 13:54:46.352920 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5fecbfbb-c1df-4171-bd01-d9d9a036cbc2-utilities\") pod \"redhat-operators-7l8hn\" (UID: \"5fecbfbb-c1df-4171-bd01-d9d9a036cbc2\") " pod="openshift-marketplace/redhat-operators-7l8hn" Nov 28 13:54:46 crc kubenswrapper[4631]: I1128 13:54:46.352989 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5fecbfbb-c1df-4171-bd01-d9d9a036cbc2-catalog-content\") pod \"redhat-operators-7l8hn\" (UID: \"5fecbfbb-c1df-4171-bd01-d9d9a036cbc2\") " pod="openshift-marketplace/redhat-operators-7l8hn" Nov 28 13:54:46 crc kubenswrapper[4631]: I1128 13:54:46.353029 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b449m\" (UniqueName: \"kubernetes.io/projected/5fecbfbb-c1df-4171-bd01-d9d9a036cbc2-kube-api-access-b449m\") pod \"redhat-operators-7l8hn\" (UID: \"5fecbfbb-c1df-4171-bd01-d9d9a036cbc2\") " pod="openshift-marketplace/redhat-operators-7l8hn" Nov 28 13:54:46 crc kubenswrapper[4631]: I1128 13:54:46.353548 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5fecbfbb-c1df-4171-bd01-d9d9a036cbc2-utilities\") pod \"redhat-operators-7l8hn\" (UID: \"5fecbfbb-c1df-4171-bd01-d9d9a036cbc2\") " pod="openshift-marketplace/redhat-operators-7l8hn" Nov 28 13:54:46 crc kubenswrapper[4631]: I1128 13:54:46.353606 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5fecbfbb-c1df-4171-bd01-d9d9a036cbc2-catalog-content\") pod \"redhat-operators-7l8hn\" (UID: \"5fecbfbb-c1df-4171-bd01-d9d9a036cbc2\") " pod="openshift-marketplace/redhat-operators-7l8hn" Nov 28 13:54:46 crc kubenswrapper[4631]: I1128 13:54:46.386461 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b449m\" (UniqueName: \"kubernetes.io/projected/5fecbfbb-c1df-4171-bd01-d9d9a036cbc2-kube-api-access-b449m\") pod \"redhat-operators-7l8hn\" (UID: \"5fecbfbb-c1df-4171-bd01-d9d9a036cbc2\") " pod="openshift-marketplace/redhat-operators-7l8hn" Nov 28 13:54:46 crc kubenswrapper[4631]: I1128 13:54:46.409034 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7l8hn" Nov 28 13:54:46 crc kubenswrapper[4631]: I1128 13:54:46.963508 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-7l8hn"] Nov 28 13:54:47 crc kubenswrapper[4631]: I1128 13:54:47.605828 4631 generic.go:334] "Generic (PLEG): container finished" podID="5fecbfbb-c1df-4171-bd01-d9d9a036cbc2" containerID="1efbe6c968409c8c724e3a17d4004f032287eb8479bd4e63eb1986709d329f01" exitCode=0 Nov 28 13:54:47 crc kubenswrapper[4631]: I1128 13:54:47.605947 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7l8hn" event={"ID":"5fecbfbb-c1df-4171-bd01-d9d9a036cbc2","Type":"ContainerDied","Data":"1efbe6c968409c8c724e3a17d4004f032287eb8479bd4e63eb1986709d329f01"} Nov 28 13:54:47 crc kubenswrapper[4631]: I1128 13:54:47.607467 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7l8hn" event={"ID":"5fecbfbb-c1df-4171-bd01-d9d9a036cbc2","Type":"ContainerStarted","Data":"343ade472f7fe6613f0b3829cfac2a3392107dfb5ad3f306828240ee98dd946c"} Nov 28 13:54:49 crc kubenswrapper[4631]: I1128 13:54:49.628014 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7l8hn" event={"ID":"5fecbfbb-c1df-4171-bd01-d9d9a036cbc2","Type":"ContainerStarted","Data":"29a664d7081c7b39096a13b7db7aae8c3ccef8b9796dc96c9584f6b1c49015e8"} Nov 28 13:54:53 crc kubenswrapper[4631]: I1128 13:54:53.670713 4631 generic.go:334] "Generic (PLEG): container finished" podID="5fecbfbb-c1df-4171-bd01-d9d9a036cbc2" containerID="29a664d7081c7b39096a13b7db7aae8c3ccef8b9796dc96c9584f6b1c49015e8" exitCode=0 Nov 28 13:54:53 crc kubenswrapper[4631]: I1128 13:54:53.670795 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7l8hn" event={"ID":"5fecbfbb-c1df-4171-bd01-d9d9a036cbc2","Type":"ContainerDied","Data":"29a664d7081c7b39096a13b7db7aae8c3ccef8b9796dc96c9584f6b1c49015e8"} Nov 28 13:54:54 crc kubenswrapper[4631]: I1128 13:54:54.683328 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7l8hn" event={"ID":"5fecbfbb-c1df-4171-bd01-d9d9a036cbc2","Type":"ContainerStarted","Data":"e0629c20ee700776f96f2382d216d8a1a284e954ab54741e5b94111e720056d1"} Nov 28 13:54:54 crc kubenswrapper[4631]: I1128 13:54:54.712546 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-7l8hn" podStartSLOduration=2.203552032 podStartE2EDuration="8.712522221s" podCreationTimestamp="2025-11-28 13:54:46 +0000 UTC" firstStartedPulling="2025-11-28 13:54:47.607919989 +0000 UTC m=+2044.415223343" lastFinishedPulling="2025-11-28 13:54:54.116890188 +0000 UTC m=+2050.924193532" observedRunningTime="2025-11-28 13:54:54.704890436 +0000 UTC m=+2051.512193780" watchObservedRunningTime="2025-11-28 13:54:54.712522221 +0000 UTC m=+2051.519825565" Nov 28 13:54:56 crc kubenswrapper[4631]: I1128 13:54:56.410314 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-7l8hn" Nov 28 13:54:56 crc kubenswrapper[4631]: I1128 13:54:56.410627 4631 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-7l8hn" Nov 28 13:54:57 crc kubenswrapper[4631]: I1128 13:54:57.473195 4631 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-7l8hn" podUID="5fecbfbb-c1df-4171-bd01-d9d9a036cbc2" containerName="registry-server" probeResult="failure" output=< Nov 28 13:54:57 crc kubenswrapper[4631]: timeout: failed to connect service ":50051" within 1s Nov 28 13:54:57 crc kubenswrapper[4631]: > Nov 28 13:55:06 crc kubenswrapper[4631]: I1128 13:55:06.461401 4631 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-7l8hn" Nov 28 13:55:06 crc kubenswrapper[4631]: I1128 13:55:06.517699 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-7l8hn" Nov 28 13:55:06 crc kubenswrapper[4631]: I1128 13:55:06.702754 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-7l8hn"] Nov 28 13:55:07 crc kubenswrapper[4631]: I1128 13:55:07.809360 4631 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-7l8hn" podUID="5fecbfbb-c1df-4171-bd01-d9d9a036cbc2" containerName="registry-server" containerID="cri-o://e0629c20ee700776f96f2382d216d8a1a284e954ab54741e5b94111e720056d1" gracePeriod=2 Nov 28 13:55:08 crc kubenswrapper[4631]: I1128 13:55:08.302831 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7l8hn" Nov 28 13:55:08 crc kubenswrapper[4631]: I1128 13:55:08.414606 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b449m\" (UniqueName: \"kubernetes.io/projected/5fecbfbb-c1df-4171-bd01-d9d9a036cbc2-kube-api-access-b449m\") pod \"5fecbfbb-c1df-4171-bd01-d9d9a036cbc2\" (UID: \"5fecbfbb-c1df-4171-bd01-d9d9a036cbc2\") " Nov 28 13:55:08 crc kubenswrapper[4631]: I1128 13:55:08.414730 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5fecbfbb-c1df-4171-bd01-d9d9a036cbc2-catalog-content\") pod \"5fecbfbb-c1df-4171-bd01-d9d9a036cbc2\" (UID: \"5fecbfbb-c1df-4171-bd01-d9d9a036cbc2\") " Nov 28 13:55:08 crc kubenswrapper[4631]: I1128 13:55:08.414777 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5fecbfbb-c1df-4171-bd01-d9d9a036cbc2-utilities\") pod \"5fecbfbb-c1df-4171-bd01-d9d9a036cbc2\" (UID: \"5fecbfbb-c1df-4171-bd01-d9d9a036cbc2\") " Nov 28 13:55:08 crc kubenswrapper[4631]: I1128 13:55:08.415696 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5fecbfbb-c1df-4171-bd01-d9d9a036cbc2-utilities" (OuterVolumeSpecName: "utilities") pod "5fecbfbb-c1df-4171-bd01-d9d9a036cbc2" (UID: "5fecbfbb-c1df-4171-bd01-d9d9a036cbc2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 13:55:08 crc kubenswrapper[4631]: I1128 13:55:08.422642 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fecbfbb-c1df-4171-bd01-d9d9a036cbc2-kube-api-access-b449m" (OuterVolumeSpecName: "kube-api-access-b449m") pod "5fecbfbb-c1df-4171-bd01-d9d9a036cbc2" (UID: "5fecbfbb-c1df-4171-bd01-d9d9a036cbc2"). InnerVolumeSpecName "kube-api-access-b449m". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:55:08 crc kubenswrapper[4631]: I1128 13:55:08.517720 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b449m\" (UniqueName: \"kubernetes.io/projected/5fecbfbb-c1df-4171-bd01-d9d9a036cbc2-kube-api-access-b449m\") on node \"crc\" DevicePath \"\"" Nov 28 13:55:08 crc kubenswrapper[4631]: I1128 13:55:08.517757 4631 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5fecbfbb-c1df-4171-bd01-d9d9a036cbc2-utilities\") on node \"crc\" DevicePath \"\"" Nov 28 13:55:08 crc kubenswrapper[4631]: I1128 13:55:08.522484 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5fecbfbb-c1df-4171-bd01-d9d9a036cbc2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5fecbfbb-c1df-4171-bd01-d9d9a036cbc2" (UID: "5fecbfbb-c1df-4171-bd01-d9d9a036cbc2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 13:55:08 crc kubenswrapper[4631]: I1128 13:55:08.619733 4631 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5fecbfbb-c1df-4171-bd01-d9d9a036cbc2-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 28 13:55:08 crc kubenswrapper[4631]: I1128 13:55:08.823678 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7l8hn" Nov 28 13:55:08 crc kubenswrapper[4631]: I1128 13:55:08.823769 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7l8hn" event={"ID":"5fecbfbb-c1df-4171-bd01-d9d9a036cbc2","Type":"ContainerDied","Data":"e0629c20ee700776f96f2382d216d8a1a284e954ab54741e5b94111e720056d1"} Nov 28 13:55:08 crc kubenswrapper[4631]: I1128 13:55:08.824068 4631 scope.go:117] "RemoveContainer" containerID="e0629c20ee700776f96f2382d216d8a1a284e954ab54741e5b94111e720056d1" Nov 28 13:55:08 crc kubenswrapper[4631]: I1128 13:55:08.823583 4631 generic.go:334] "Generic (PLEG): container finished" podID="5fecbfbb-c1df-4171-bd01-d9d9a036cbc2" containerID="e0629c20ee700776f96f2382d216d8a1a284e954ab54741e5b94111e720056d1" exitCode=0 Nov 28 13:55:08 crc kubenswrapper[4631]: I1128 13:55:08.828495 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7l8hn" event={"ID":"5fecbfbb-c1df-4171-bd01-d9d9a036cbc2","Type":"ContainerDied","Data":"343ade472f7fe6613f0b3829cfac2a3392107dfb5ad3f306828240ee98dd946c"} Nov 28 13:55:08 crc kubenswrapper[4631]: I1128 13:55:08.864224 4631 scope.go:117] "RemoveContainer" containerID="29a664d7081c7b39096a13b7db7aae8c3ccef8b9796dc96c9584f6b1c49015e8" Nov 28 13:55:08 crc kubenswrapper[4631]: I1128 13:55:08.872384 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-7l8hn"] Nov 28 13:55:08 crc kubenswrapper[4631]: I1128 13:55:08.893545 4631 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-7l8hn"] Nov 28 13:55:08 crc kubenswrapper[4631]: I1128 13:55:08.904748 4631 scope.go:117] "RemoveContainer" containerID="1efbe6c968409c8c724e3a17d4004f032287eb8479bd4e63eb1986709d329f01" Nov 28 13:55:08 crc kubenswrapper[4631]: I1128 13:55:08.946437 4631 scope.go:117] "RemoveContainer" containerID="e0629c20ee700776f96f2382d216d8a1a284e954ab54741e5b94111e720056d1" Nov 28 13:55:08 crc kubenswrapper[4631]: E1128 13:55:08.947049 4631 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e0629c20ee700776f96f2382d216d8a1a284e954ab54741e5b94111e720056d1\": container with ID starting with e0629c20ee700776f96f2382d216d8a1a284e954ab54741e5b94111e720056d1 not found: ID does not exist" containerID="e0629c20ee700776f96f2382d216d8a1a284e954ab54741e5b94111e720056d1" Nov 28 13:55:08 crc kubenswrapper[4631]: I1128 13:55:08.947087 4631 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e0629c20ee700776f96f2382d216d8a1a284e954ab54741e5b94111e720056d1"} err="failed to get container status \"e0629c20ee700776f96f2382d216d8a1a284e954ab54741e5b94111e720056d1\": rpc error: code = NotFound desc = could not find container \"e0629c20ee700776f96f2382d216d8a1a284e954ab54741e5b94111e720056d1\": container with ID starting with e0629c20ee700776f96f2382d216d8a1a284e954ab54741e5b94111e720056d1 not found: ID does not exist" Nov 28 13:55:08 crc kubenswrapper[4631]: I1128 13:55:08.947119 4631 scope.go:117] "RemoveContainer" containerID="29a664d7081c7b39096a13b7db7aae8c3ccef8b9796dc96c9584f6b1c49015e8" Nov 28 13:55:08 crc kubenswrapper[4631]: E1128 13:55:08.947457 4631 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"29a664d7081c7b39096a13b7db7aae8c3ccef8b9796dc96c9584f6b1c49015e8\": container with ID starting with 29a664d7081c7b39096a13b7db7aae8c3ccef8b9796dc96c9584f6b1c49015e8 not found: ID does not exist" containerID="29a664d7081c7b39096a13b7db7aae8c3ccef8b9796dc96c9584f6b1c49015e8" Nov 28 13:55:08 crc kubenswrapper[4631]: I1128 13:55:08.947488 4631 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"29a664d7081c7b39096a13b7db7aae8c3ccef8b9796dc96c9584f6b1c49015e8"} err="failed to get container status \"29a664d7081c7b39096a13b7db7aae8c3ccef8b9796dc96c9584f6b1c49015e8\": rpc error: code = NotFound desc = could not find container \"29a664d7081c7b39096a13b7db7aae8c3ccef8b9796dc96c9584f6b1c49015e8\": container with ID starting with 29a664d7081c7b39096a13b7db7aae8c3ccef8b9796dc96c9584f6b1c49015e8 not found: ID does not exist" Nov 28 13:55:08 crc kubenswrapper[4631]: I1128 13:55:08.947509 4631 scope.go:117] "RemoveContainer" containerID="1efbe6c968409c8c724e3a17d4004f032287eb8479bd4e63eb1986709d329f01" Nov 28 13:55:08 crc kubenswrapper[4631]: E1128 13:55:08.947851 4631 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1efbe6c968409c8c724e3a17d4004f032287eb8479bd4e63eb1986709d329f01\": container with ID starting with 1efbe6c968409c8c724e3a17d4004f032287eb8479bd4e63eb1986709d329f01 not found: ID does not exist" containerID="1efbe6c968409c8c724e3a17d4004f032287eb8479bd4e63eb1986709d329f01" Nov 28 13:55:08 crc kubenswrapper[4631]: I1128 13:55:08.947877 4631 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1efbe6c968409c8c724e3a17d4004f032287eb8479bd4e63eb1986709d329f01"} err="failed to get container status \"1efbe6c968409c8c724e3a17d4004f032287eb8479bd4e63eb1986709d329f01\": rpc error: code = NotFound desc = could not find container \"1efbe6c968409c8c724e3a17d4004f032287eb8479bd4e63eb1986709d329f01\": container with ID starting with 1efbe6c968409c8c724e3a17d4004f032287eb8479bd4e63eb1986709d329f01 not found: ID does not exist" Nov 28 13:55:09 crc kubenswrapper[4631]: I1128 13:55:09.527916 4631 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fecbfbb-c1df-4171-bd01-d9d9a036cbc2" path="/var/lib/kubelet/pods/5fecbfbb-c1df-4171-bd01-d9d9a036cbc2/volumes" Nov 28 13:55:12 crc kubenswrapper[4631]: I1128 13:55:12.867102 4631 generic.go:334] "Generic (PLEG): container finished" podID="c09a85dc-b780-41c9-8122-653d4703aa7e" containerID="639dacf5cffcf75693e5eac449744cff4ce4a0729c1f60d3961fd670a60f79bd" exitCode=0 Nov 28 13:55:12 crc kubenswrapper[4631]: I1128 13:55:12.867781 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-2sls9" event={"ID":"c09a85dc-b780-41c9-8122-653d4703aa7e","Type":"ContainerDied","Data":"639dacf5cffcf75693e5eac449744cff4ce4a0729c1f60d3961fd670a60f79bd"} Nov 28 13:55:14 crc kubenswrapper[4631]: I1128 13:55:14.424353 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-2sls9" Nov 28 13:55:14 crc kubenswrapper[4631]: I1128 13:55:14.555525 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c09a85dc-b780-41c9-8122-653d4703aa7e-inventory\") pod \"c09a85dc-b780-41c9-8122-653d4703aa7e\" (UID: \"c09a85dc-b780-41c9-8122-653d4703aa7e\") " Nov 28 13:55:14 crc kubenswrapper[4631]: I1128 13:55:14.555617 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9rlbm\" (UniqueName: \"kubernetes.io/projected/c09a85dc-b780-41c9-8122-653d4703aa7e-kube-api-access-9rlbm\") pod \"c09a85dc-b780-41c9-8122-653d4703aa7e\" (UID: \"c09a85dc-b780-41c9-8122-653d4703aa7e\") " Nov 28 13:55:14 crc kubenswrapper[4631]: I1128 13:55:14.555641 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c09a85dc-b780-41c9-8122-653d4703aa7e-ovn-combined-ca-bundle\") pod \"c09a85dc-b780-41c9-8122-653d4703aa7e\" (UID: \"c09a85dc-b780-41c9-8122-653d4703aa7e\") " Nov 28 13:55:14 crc kubenswrapper[4631]: I1128 13:55:14.555802 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c09a85dc-b780-41c9-8122-653d4703aa7e-ssh-key\") pod \"c09a85dc-b780-41c9-8122-653d4703aa7e\" (UID: \"c09a85dc-b780-41c9-8122-653d4703aa7e\") " Nov 28 13:55:14 crc kubenswrapper[4631]: I1128 13:55:14.555829 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/c09a85dc-b780-41c9-8122-653d4703aa7e-ovncontroller-config-0\") pod \"c09a85dc-b780-41c9-8122-653d4703aa7e\" (UID: \"c09a85dc-b780-41c9-8122-653d4703aa7e\") " Nov 28 13:55:14 crc kubenswrapper[4631]: I1128 13:55:14.563687 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c09a85dc-b780-41c9-8122-653d4703aa7e-kube-api-access-9rlbm" (OuterVolumeSpecName: "kube-api-access-9rlbm") pod "c09a85dc-b780-41c9-8122-653d4703aa7e" (UID: "c09a85dc-b780-41c9-8122-653d4703aa7e"). InnerVolumeSpecName "kube-api-access-9rlbm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:55:14 crc kubenswrapper[4631]: I1128 13:55:14.594874 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c09a85dc-b780-41c9-8122-653d4703aa7e-ovncontroller-config-0" (OuterVolumeSpecName: "ovncontroller-config-0") pod "c09a85dc-b780-41c9-8122-653d4703aa7e" (UID: "c09a85dc-b780-41c9-8122-653d4703aa7e"). InnerVolumeSpecName "ovncontroller-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 13:55:14 crc kubenswrapper[4631]: I1128 13:55:14.601771 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c09a85dc-b780-41c9-8122-653d4703aa7e-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "c09a85dc-b780-41c9-8122-653d4703aa7e" (UID: "c09a85dc-b780-41c9-8122-653d4703aa7e"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:55:14 crc kubenswrapper[4631]: I1128 13:55:14.606518 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c09a85dc-b780-41c9-8122-653d4703aa7e-inventory" (OuterVolumeSpecName: "inventory") pod "c09a85dc-b780-41c9-8122-653d4703aa7e" (UID: "c09a85dc-b780-41c9-8122-653d4703aa7e"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:55:14 crc kubenswrapper[4631]: I1128 13:55:14.622923 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c09a85dc-b780-41c9-8122-653d4703aa7e-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "c09a85dc-b780-41c9-8122-653d4703aa7e" (UID: "c09a85dc-b780-41c9-8122-653d4703aa7e"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:55:14 crc kubenswrapper[4631]: I1128 13:55:14.658977 4631 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c09a85dc-b780-41c9-8122-653d4703aa7e-inventory\") on node \"crc\" DevicePath \"\"" Nov 28 13:55:14 crc kubenswrapper[4631]: I1128 13:55:14.659037 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9rlbm\" (UniqueName: \"kubernetes.io/projected/c09a85dc-b780-41c9-8122-653d4703aa7e-kube-api-access-9rlbm\") on node \"crc\" DevicePath \"\"" Nov 28 13:55:14 crc kubenswrapper[4631]: I1128 13:55:14.659052 4631 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c09a85dc-b780-41c9-8122-653d4703aa7e-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 13:55:14 crc kubenswrapper[4631]: I1128 13:55:14.659063 4631 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c09a85dc-b780-41c9-8122-653d4703aa7e-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 28 13:55:14 crc kubenswrapper[4631]: I1128 13:55:14.659075 4631 reconciler_common.go:293] "Volume detached for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/c09a85dc-b780-41c9-8122-653d4703aa7e-ovncontroller-config-0\") on node \"crc\" DevicePath \"\"" Nov 28 13:55:14 crc kubenswrapper[4631]: I1128 13:55:14.889064 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-2sls9" event={"ID":"c09a85dc-b780-41c9-8122-653d4703aa7e","Type":"ContainerDied","Data":"bd8114646555df58b772ae35891365b0e848a55f2e996bddf3e0b5e684939c38"} Nov 28 13:55:14 crc kubenswrapper[4631]: I1128 13:55:14.889112 4631 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bd8114646555df58b772ae35891365b0e848a55f2e996bddf3e0b5e684939c38" Nov 28 13:55:14 crc kubenswrapper[4631]: I1128 13:55:14.889173 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-2sls9" Nov 28 13:55:15 crc kubenswrapper[4631]: I1128 13:55:15.036532 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-tl8jt"] Nov 28 13:55:15 crc kubenswrapper[4631]: E1128 13:55:15.036964 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c09a85dc-b780-41c9-8122-653d4703aa7e" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Nov 28 13:55:15 crc kubenswrapper[4631]: I1128 13:55:15.036983 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="c09a85dc-b780-41c9-8122-653d4703aa7e" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Nov 28 13:55:15 crc kubenswrapper[4631]: E1128 13:55:15.037001 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5fecbfbb-c1df-4171-bd01-d9d9a036cbc2" containerName="registry-server" Nov 28 13:55:15 crc kubenswrapper[4631]: I1128 13:55:15.037011 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="5fecbfbb-c1df-4171-bd01-d9d9a036cbc2" containerName="registry-server" Nov 28 13:55:15 crc kubenswrapper[4631]: E1128 13:55:15.037025 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5fecbfbb-c1df-4171-bd01-d9d9a036cbc2" containerName="extract-utilities" Nov 28 13:55:15 crc kubenswrapper[4631]: I1128 13:55:15.037033 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="5fecbfbb-c1df-4171-bd01-d9d9a036cbc2" containerName="extract-utilities" Nov 28 13:55:15 crc kubenswrapper[4631]: E1128 13:55:15.037060 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5fecbfbb-c1df-4171-bd01-d9d9a036cbc2" containerName="extract-content" Nov 28 13:55:15 crc kubenswrapper[4631]: I1128 13:55:15.037067 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="5fecbfbb-c1df-4171-bd01-d9d9a036cbc2" containerName="extract-content" Nov 28 13:55:15 crc kubenswrapper[4631]: I1128 13:55:15.037307 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="c09a85dc-b780-41c9-8122-653d4703aa7e" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Nov 28 13:55:15 crc kubenswrapper[4631]: I1128 13:55:15.037335 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="5fecbfbb-c1df-4171-bd01-d9d9a036cbc2" containerName="registry-server" Nov 28 13:55:15 crc kubenswrapper[4631]: I1128 13:55:15.037988 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-tl8jt" Nov 28 13:55:15 crc kubenswrapper[4631]: I1128 13:55:15.043414 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-ovn-metadata-agent-neutron-config" Nov 28 13:55:15 crc kubenswrapper[4631]: I1128 13:55:15.043798 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 28 13:55:15 crc kubenswrapper[4631]: I1128 13:55:15.046359 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 28 13:55:15 crc kubenswrapper[4631]: I1128 13:55:15.046652 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-neutron-config" Nov 28 13:55:15 crc kubenswrapper[4631]: I1128 13:55:15.068660 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-vp4ns" Nov 28 13:55:15 crc kubenswrapper[4631]: I1128 13:55:15.075849 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-tl8jt"] Nov 28 13:55:15 crc kubenswrapper[4631]: I1128 13:55:15.076704 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 28 13:55:15 crc kubenswrapper[4631]: I1128 13:55:15.174558 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4c4pz\" (UniqueName: \"kubernetes.io/projected/a73f7b13-ead0-46a0-8c62-d3c1e75bc2d8-kube-api-access-4c4pz\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-tl8jt\" (UID: \"a73f7b13-ead0-46a0-8c62-d3c1e75bc2d8\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-tl8jt" Nov 28 13:55:15 crc kubenswrapper[4631]: I1128 13:55:15.174644 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/a73f7b13-ead0-46a0-8c62-d3c1e75bc2d8-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-tl8jt\" (UID: \"a73f7b13-ead0-46a0-8c62-d3c1e75bc2d8\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-tl8jt" Nov 28 13:55:15 crc kubenswrapper[4631]: I1128 13:55:15.175010 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/a73f7b13-ead0-46a0-8c62-d3c1e75bc2d8-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-tl8jt\" (UID: \"a73f7b13-ead0-46a0-8c62-d3c1e75bc2d8\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-tl8jt" Nov 28 13:55:15 crc kubenswrapper[4631]: I1128 13:55:15.175101 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a73f7b13-ead0-46a0-8c62-d3c1e75bc2d8-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-tl8jt\" (UID: \"a73f7b13-ead0-46a0-8c62-d3c1e75bc2d8\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-tl8jt" Nov 28 13:55:15 crc kubenswrapper[4631]: I1128 13:55:15.175364 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a73f7b13-ead0-46a0-8c62-d3c1e75bc2d8-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-tl8jt\" (UID: \"a73f7b13-ead0-46a0-8c62-d3c1e75bc2d8\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-tl8jt" Nov 28 13:55:15 crc kubenswrapper[4631]: I1128 13:55:15.175412 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a73f7b13-ead0-46a0-8c62-d3c1e75bc2d8-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-tl8jt\" (UID: \"a73f7b13-ead0-46a0-8c62-d3c1e75bc2d8\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-tl8jt" Nov 28 13:55:15 crc kubenswrapper[4631]: I1128 13:55:15.277227 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4c4pz\" (UniqueName: \"kubernetes.io/projected/a73f7b13-ead0-46a0-8c62-d3c1e75bc2d8-kube-api-access-4c4pz\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-tl8jt\" (UID: \"a73f7b13-ead0-46a0-8c62-d3c1e75bc2d8\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-tl8jt" Nov 28 13:55:15 crc kubenswrapper[4631]: I1128 13:55:15.277278 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/a73f7b13-ead0-46a0-8c62-d3c1e75bc2d8-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-tl8jt\" (UID: \"a73f7b13-ead0-46a0-8c62-d3c1e75bc2d8\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-tl8jt" Nov 28 13:55:15 crc kubenswrapper[4631]: I1128 13:55:15.277362 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/a73f7b13-ead0-46a0-8c62-d3c1e75bc2d8-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-tl8jt\" (UID: \"a73f7b13-ead0-46a0-8c62-d3c1e75bc2d8\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-tl8jt" Nov 28 13:55:15 crc kubenswrapper[4631]: I1128 13:55:15.277387 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a73f7b13-ead0-46a0-8c62-d3c1e75bc2d8-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-tl8jt\" (UID: \"a73f7b13-ead0-46a0-8c62-d3c1e75bc2d8\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-tl8jt" Nov 28 13:55:15 crc kubenswrapper[4631]: I1128 13:55:15.277424 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a73f7b13-ead0-46a0-8c62-d3c1e75bc2d8-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-tl8jt\" (UID: \"a73f7b13-ead0-46a0-8c62-d3c1e75bc2d8\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-tl8jt" Nov 28 13:55:15 crc kubenswrapper[4631]: I1128 13:55:15.277491 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a73f7b13-ead0-46a0-8c62-d3c1e75bc2d8-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-tl8jt\" (UID: \"a73f7b13-ead0-46a0-8c62-d3c1e75bc2d8\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-tl8jt" Nov 28 13:55:15 crc kubenswrapper[4631]: I1128 13:55:15.283493 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a73f7b13-ead0-46a0-8c62-d3c1e75bc2d8-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-tl8jt\" (UID: \"a73f7b13-ead0-46a0-8c62-d3c1e75bc2d8\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-tl8jt" Nov 28 13:55:15 crc kubenswrapper[4631]: I1128 13:55:15.283805 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/a73f7b13-ead0-46a0-8c62-d3c1e75bc2d8-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-tl8jt\" (UID: \"a73f7b13-ead0-46a0-8c62-d3c1e75bc2d8\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-tl8jt" Nov 28 13:55:15 crc kubenswrapper[4631]: I1128 13:55:15.286065 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/a73f7b13-ead0-46a0-8c62-d3c1e75bc2d8-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-tl8jt\" (UID: \"a73f7b13-ead0-46a0-8c62-d3c1e75bc2d8\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-tl8jt" Nov 28 13:55:15 crc kubenswrapper[4631]: I1128 13:55:15.287118 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a73f7b13-ead0-46a0-8c62-d3c1e75bc2d8-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-tl8jt\" (UID: \"a73f7b13-ead0-46a0-8c62-d3c1e75bc2d8\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-tl8jt" Nov 28 13:55:15 crc kubenswrapper[4631]: I1128 13:55:15.287713 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a73f7b13-ead0-46a0-8c62-d3c1e75bc2d8-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-tl8jt\" (UID: \"a73f7b13-ead0-46a0-8c62-d3c1e75bc2d8\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-tl8jt" Nov 28 13:55:15 crc kubenswrapper[4631]: I1128 13:55:15.296557 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4c4pz\" (UniqueName: \"kubernetes.io/projected/a73f7b13-ead0-46a0-8c62-d3c1e75bc2d8-kube-api-access-4c4pz\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-tl8jt\" (UID: \"a73f7b13-ead0-46a0-8c62-d3c1e75bc2d8\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-tl8jt" Nov 28 13:55:15 crc kubenswrapper[4631]: I1128 13:55:15.379861 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-tl8jt" Nov 28 13:55:16 crc kubenswrapper[4631]: I1128 13:55:16.093960 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-tl8jt"] Nov 28 13:55:16 crc kubenswrapper[4631]: I1128 13:55:16.920485 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-tl8jt" event={"ID":"a73f7b13-ead0-46a0-8c62-d3c1e75bc2d8","Type":"ContainerStarted","Data":"cdc53e23a08c8ee3166c0454e0a42412ecbc0fd4a366ad977f81e8922fb061ce"} Nov 28 13:55:16 crc kubenswrapper[4631]: I1128 13:55:16.921182 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-tl8jt" event={"ID":"a73f7b13-ead0-46a0-8c62-d3c1e75bc2d8","Type":"ContainerStarted","Data":"f32d2ba6f015d9bb0358b89b6f01c29301bd16a95fa225b83f3991cb43e970fe"} Nov 28 13:55:16 crc kubenswrapper[4631]: I1128 13:55:16.951581 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-tl8jt" podStartSLOduration=2.442940674 podStartE2EDuration="2.951270669s" podCreationTimestamp="2025-11-28 13:55:14 +0000 UTC" firstStartedPulling="2025-11-28 13:55:16.101849406 +0000 UTC m=+2072.909152750" lastFinishedPulling="2025-11-28 13:55:16.610179401 +0000 UTC m=+2073.417482745" observedRunningTime="2025-11-28 13:55:16.941631795 +0000 UTC m=+2073.748935139" watchObservedRunningTime="2025-11-28 13:55:16.951270669 +0000 UTC m=+2073.758574013" Nov 28 13:55:19 crc kubenswrapper[4631]: I1128 13:55:19.635147 4631 patch_prober.go:28] interesting pod/machine-config-daemon-47sxw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 28 13:55:19 crc kubenswrapper[4631]: I1128 13:55:19.635794 4631 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" podUID="8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 28 13:55:40 crc kubenswrapper[4631]: I1128 13:55:40.964758 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-fdd9l"] Nov 28 13:55:40 crc kubenswrapper[4631]: I1128 13:55:40.967941 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-fdd9l" Nov 28 13:55:40 crc kubenswrapper[4631]: I1128 13:55:40.989045 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-fdd9l"] Nov 28 13:55:41 crc kubenswrapper[4631]: I1128 13:55:41.009971 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1f1e9bbe-987c-4acd-85f3-e22146440472-catalog-content\") pod \"community-operators-fdd9l\" (UID: \"1f1e9bbe-987c-4acd-85f3-e22146440472\") " pod="openshift-marketplace/community-operators-fdd9l" Nov 28 13:55:41 crc kubenswrapper[4631]: I1128 13:55:41.010697 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bbftn\" (UniqueName: \"kubernetes.io/projected/1f1e9bbe-987c-4acd-85f3-e22146440472-kube-api-access-bbftn\") pod \"community-operators-fdd9l\" (UID: \"1f1e9bbe-987c-4acd-85f3-e22146440472\") " pod="openshift-marketplace/community-operators-fdd9l" Nov 28 13:55:41 crc kubenswrapper[4631]: I1128 13:55:41.010826 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1f1e9bbe-987c-4acd-85f3-e22146440472-utilities\") pod \"community-operators-fdd9l\" (UID: \"1f1e9bbe-987c-4acd-85f3-e22146440472\") " pod="openshift-marketplace/community-operators-fdd9l" Nov 28 13:55:41 crc kubenswrapper[4631]: I1128 13:55:41.112851 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1f1e9bbe-987c-4acd-85f3-e22146440472-catalog-content\") pod \"community-operators-fdd9l\" (UID: \"1f1e9bbe-987c-4acd-85f3-e22146440472\") " pod="openshift-marketplace/community-operators-fdd9l" Nov 28 13:55:41 crc kubenswrapper[4631]: I1128 13:55:41.112993 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bbftn\" (UniqueName: \"kubernetes.io/projected/1f1e9bbe-987c-4acd-85f3-e22146440472-kube-api-access-bbftn\") pod \"community-operators-fdd9l\" (UID: \"1f1e9bbe-987c-4acd-85f3-e22146440472\") " pod="openshift-marketplace/community-operators-fdd9l" Nov 28 13:55:41 crc kubenswrapper[4631]: I1128 13:55:41.113045 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1f1e9bbe-987c-4acd-85f3-e22146440472-utilities\") pod \"community-operators-fdd9l\" (UID: \"1f1e9bbe-987c-4acd-85f3-e22146440472\") " pod="openshift-marketplace/community-operators-fdd9l" Nov 28 13:55:41 crc kubenswrapper[4631]: I1128 13:55:41.113567 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1f1e9bbe-987c-4acd-85f3-e22146440472-utilities\") pod \"community-operators-fdd9l\" (UID: \"1f1e9bbe-987c-4acd-85f3-e22146440472\") " pod="openshift-marketplace/community-operators-fdd9l" Nov 28 13:55:41 crc kubenswrapper[4631]: I1128 13:55:41.113797 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1f1e9bbe-987c-4acd-85f3-e22146440472-catalog-content\") pod \"community-operators-fdd9l\" (UID: \"1f1e9bbe-987c-4acd-85f3-e22146440472\") " pod="openshift-marketplace/community-operators-fdd9l" Nov 28 13:55:41 crc kubenswrapper[4631]: I1128 13:55:41.140768 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bbftn\" (UniqueName: \"kubernetes.io/projected/1f1e9bbe-987c-4acd-85f3-e22146440472-kube-api-access-bbftn\") pod \"community-operators-fdd9l\" (UID: \"1f1e9bbe-987c-4acd-85f3-e22146440472\") " pod="openshift-marketplace/community-operators-fdd9l" Nov 28 13:55:41 crc kubenswrapper[4631]: I1128 13:55:41.290647 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-fdd9l" Nov 28 13:55:41 crc kubenswrapper[4631]: I1128 13:55:41.906534 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-fdd9l"] Nov 28 13:55:42 crc kubenswrapper[4631]: I1128 13:55:42.164037 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fdd9l" event={"ID":"1f1e9bbe-987c-4acd-85f3-e22146440472","Type":"ContainerStarted","Data":"c1c7f3e0d82db9374ad151825c72f615755d967d890050737b080c61dce9d275"} Nov 28 13:55:43 crc kubenswrapper[4631]: I1128 13:55:43.179133 4631 generic.go:334] "Generic (PLEG): container finished" podID="1f1e9bbe-987c-4acd-85f3-e22146440472" containerID="f28126a9814c3065aa169bef80450931687265878d0a111e8e61aa6621b0ba55" exitCode=0 Nov 28 13:55:43 crc kubenswrapper[4631]: I1128 13:55:43.179227 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fdd9l" event={"ID":"1f1e9bbe-987c-4acd-85f3-e22146440472","Type":"ContainerDied","Data":"f28126a9814c3065aa169bef80450931687265878d0a111e8e61aa6621b0ba55"} Nov 28 13:55:43 crc kubenswrapper[4631]: I1128 13:55:43.181507 4631 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 28 13:55:46 crc kubenswrapper[4631]: I1128 13:55:46.212451 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fdd9l" event={"ID":"1f1e9bbe-987c-4acd-85f3-e22146440472","Type":"ContainerStarted","Data":"b0b34e06b6cda5f4a3c7f921f2aa0bc8e9a0f47f153bb9b00cad5392eae9d147"} Nov 28 13:55:47 crc kubenswrapper[4631]: I1128 13:55:47.222020 4631 generic.go:334] "Generic (PLEG): container finished" podID="1f1e9bbe-987c-4acd-85f3-e22146440472" containerID="b0b34e06b6cda5f4a3c7f921f2aa0bc8e9a0f47f153bb9b00cad5392eae9d147" exitCode=0 Nov 28 13:55:47 crc kubenswrapper[4631]: I1128 13:55:47.222090 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fdd9l" event={"ID":"1f1e9bbe-987c-4acd-85f3-e22146440472","Type":"ContainerDied","Data":"b0b34e06b6cda5f4a3c7f921f2aa0bc8e9a0f47f153bb9b00cad5392eae9d147"} Nov 28 13:55:48 crc kubenswrapper[4631]: I1128 13:55:48.236375 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fdd9l" event={"ID":"1f1e9bbe-987c-4acd-85f3-e22146440472","Type":"ContainerStarted","Data":"23dc55f6d0fcabd11ae900a04ef530a2d39c73984f8a1e2f877913a066f4c745"} Nov 28 13:55:48 crc kubenswrapper[4631]: I1128 13:55:48.268789 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-fdd9l" podStartSLOduration=3.835018404 podStartE2EDuration="8.268768086s" podCreationTimestamp="2025-11-28 13:55:40 +0000 UTC" firstStartedPulling="2025-11-28 13:55:43.181128826 +0000 UTC m=+2099.988432170" lastFinishedPulling="2025-11-28 13:55:47.614878508 +0000 UTC m=+2104.422181852" observedRunningTime="2025-11-28 13:55:48.261379297 +0000 UTC m=+2105.068682661" watchObservedRunningTime="2025-11-28 13:55:48.268768086 +0000 UTC m=+2105.076071430" Nov 28 13:55:49 crc kubenswrapper[4631]: I1128 13:55:49.634641 4631 patch_prober.go:28] interesting pod/machine-config-daemon-47sxw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 28 13:55:49 crc kubenswrapper[4631]: I1128 13:55:49.634967 4631 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" podUID="8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 28 13:55:51 crc kubenswrapper[4631]: I1128 13:55:51.292499 4631 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-fdd9l" Nov 28 13:55:51 crc kubenswrapper[4631]: I1128 13:55:51.292834 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-fdd9l" Nov 28 13:55:51 crc kubenswrapper[4631]: I1128 13:55:51.346551 4631 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-fdd9l" Nov 28 13:55:52 crc kubenswrapper[4631]: I1128 13:55:52.319912 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-fdd9l" Nov 28 13:55:52 crc kubenswrapper[4631]: I1128 13:55:52.397632 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-fdd9l"] Nov 28 13:55:54 crc kubenswrapper[4631]: I1128 13:55:54.291495 4631 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-fdd9l" podUID="1f1e9bbe-987c-4acd-85f3-e22146440472" containerName="registry-server" containerID="cri-o://23dc55f6d0fcabd11ae900a04ef530a2d39c73984f8a1e2f877913a066f4c745" gracePeriod=2 Nov 28 13:55:55 crc kubenswrapper[4631]: I1128 13:55:55.277337 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-fdd9l" Nov 28 13:55:55 crc kubenswrapper[4631]: I1128 13:55:55.305555 4631 generic.go:334] "Generic (PLEG): container finished" podID="1f1e9bbe-987c-4acd-85f3-e22146440472" containerID="23dc55f6d0fcabd11ae900a04ef530a2d39c73984f8a1e2f877913a066f4c745" exitCode=0 Nov 28 13:55:55 crc kubenswrapper[4631]: I1128 13:55:55.305609 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fdd9l" event={"ID":"1f1e9bbe-987c-4acd-85f3-e22146440472","Type":"ContainerDied","Data":"23dc55f6d0fcabd11ae900a04ef530a2d39c73984f8a1e2f877913a066f4c745"} Nov 28 13:55:55 crc kubenswrapper[4631]: I1128 13:55:55.305639 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fdd9l" event={"ID":"1f1e9bbe-987c-4acd-85f3-e22146440472","Type":"ContainerDied","Data":"c1c7f3e0d82db9374ad151825c72f615755d967d890050737b080c61dce9d275"} Nov 28 13:55:55 crc kubenswrapper[4631]: I1128 13:55:55.305660 4631 scope.go:117] "RemoveContainer" containerID="23dc55f6d0fcabd11ae900a04ef530a2d39c73984f8a1e2f877913a066f4c745" Nov 28 13:55:55 crc kubenswrapper[4631]: I1128 13:55:55.305805 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-fdd9l" Nov 28 13:55:55 crc kubenswrapper[4631]: I1128 13:55:55.327702 4631 scope.go:117] "RemoveContainer" containerID="b0b34e06b6cda5f4a3c7f921f2aa0bc8e9a0f47f153bb9b00cad5392eae9d147" Nov 28 13:55:55 crc kubenswrapper[4631]: I1128 13:55:55.369393 4631 scope.go:117] "RemoveContainer" containerID="f28126a9814c3065aa169bef80450931687265878d0a111e8e61aa6621b0ba55" Nov 28 13:55:55 crc kubenswrapper[4631]: I1128 13:55:55.414052 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1f1e9bbe-987c-4acd-85f3-e22146440472-catalog-content\") pod \"1f1e9bbe-987c-4acd-85f3-e22146440472\" (UID: \"1f1e9bbe-987c-4acd-85f3-e22146440472\") " Nov 28 13:55:55 crc kubenswrapper[4631]: I1128 13:55:55.414310 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bbftn\" (UniqueName: \"kubernetes.io/projected/1f1e9bbe-987c-4acd-85f3-e22146440472-kube-api-access-bbftn\") pod \"1f1e9bbe-987c-4acd-85f3-e22146440472\" (UID: \"1f1e9bbe-987c-4acd-85f3-e22146440472\") " Nov 28 13:55:55 crc kubenswrapper[4631]: I1128 13:55:55.414403 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1f1e9bbe-987c-4acd-85f3-e22146440472-utilities\") pod \"1f1e9bbe-987c-4acd-85f3-e22146440472\" (UID: \"1f1e9bbe-987c-4acd-85f3-e22146440472\") " Nov 28 13:55:55 crc kubenswrapper[4631]: I1128 13:55:55.416165 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1f1e9bbe-987c-4acd-85f3-e22146440472-utilities" (OuterVolumeSpecName: "utilities") pod "1f1e9bbe-987c-4acd-85f3-e22146440472" (UID: "1f1e9bbe-987c-4acd-85f3-e22146440472"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 13:55:55 crc kubenswrapper[4631]: I1128 13:55:55.428245 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1f1e9bbe-987c-4acd-85f3-e22146440472-kube-api-access-bbftn" (OuterVolumeSpecName: "kube-api-access-bbftn") pod "1f1e9bbe-987c-4acd-85f3-e22146440472" (UID: "1f1e9bbe-987c-4acd-85f3-e22146440472"). InnerVolumeSpecName "kube-api-access-bbftn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:55:55 crc kubenswrapper[4631]: I1128 13:55:55.438088 4631 scope.go:117] "RemoveContainer" containerID="23dc55f6d0fcabd11ae900a04ef530a2d39c73984f8a1e2f877913a066f4c745" Nov 28 13:55:55 crc kubenswrapper[4631]: E1128 13:55:55.439857 4631 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"23dc55f6d0fcabd11ae900a04ef530a2d39c73984f8a1e2f877913a066f4c745\": container with ID starting with 23dc55f6d0fcabd11ae900a04ef530a2d39c73984f8a1e2f877913a066f4c745 not found: ID does not exist" containerID="23dc55f6d0fcabd11ae900a04ef530a2d39c73984f8a1e2f877913a066f4c745" Nov 28 13:55:55 crc kubenswrapper[4631]: I1128 13:55:55.439975 4631 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"23dc55f6d0fcabd11ae900a04ef530a2d39c73984f8a1e2f877913a066f4c745"} err="failed to get container status \"23dc55f6d0fcabd11ae900a04ef530a2d39c73984f8a1e2f877913a066f4c745\": rpc error: code = NotFound desc = could not find container \"23dc55f6d0fcabd11ae900a04ef530a2d39c73984f8a1e2f877913a066f4c745\": container with ID starting with 23dc55f6d0fcabd11ae900a04ef530a2d39c73984f8a1e2f877913a066f4c745 not found: ID does not exist" Nov 28 13:55:55 crc kubenswrapper[4631]: I1128 13:55:55.440099 4631 scope.go:117] "RemoveContainer" containerID="b0b34e06b6cda5f4a3c7f921f2aa0bc8e9a0f47f153bb9b00cad5392eae9d147" Nov 28 13:55:55 crc kubenswrapper[4631]: E1128 13:55:55.440712 4631 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b0b34e06b6cda5f4a3c7f921f2aa0bc8e9a0f47f153bb9b00cad5392eae9d147\": container with ID starting with b0b34e06b6cda5f4a3c7f921f2aa0bc8e9a0f47f153bb9b00cad5392eae9d147 not found: ID does not exist" containerID="b0b34e06b6cda5f4a3c7f921f2aa0bc8e9a0f47f153bb9b00cad5392eae9d147" Nov 28 13:55:55 crc kubenswrapper[4631]: I1128 13:55:55.440809 4631 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b0b34e06b6cda5f4a3c7f921f2aa0bc8e9a0f47f153bb9b00cad5392eae9d147"} err="failed to get container status \"b0b34e06b6cda5f4a3c7f921f2aa0bc8e9a0f47f153bb9b00cad5392eae9d147\": rpc error: code = NotFound desc = could not find container \"b0b34e06b6cda5f4a3c7f921f2aa0bc8e9a0f47f153bb9b00cad5392eae9d147\": container with ID starting with b0b34e06b6cda5f4a3c7f921f2aa0bc8e9a0f47f153bb9b00cad5392eae9d147 not found: ID does not exist" Nov 28 13:55:55 crc kubenswrapper[4631]: I1128 13:55:55.440858 4631 scope.go:117] "RemoveContainer" containerID="f28126a9814c3065aa169bef80450931687265878d0a111e8e61aa6621b0ba55" Nov 28 13:55:55 crc kubenswrapper[4631]: E1128 13:55:55.441163 4631 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f28126a9814c3065aa169bef80450931687265878d0a111e8e61aa6621b0ba55\": container with ID starting with f28126a9814c3065aa169bef80450931687265878d0a111e8e61aa6621b0ba55 not found: ID does not exist" containerID="f28126a9814c3065aa169bef80450931687265878d0a111e8e61aa6621b0ba55" Nov 28 13:55:55 crc kubenswrapper[4631]: I1128 13:55:55.441259 4631 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f28126a9814c3065aa169bef80450931687265878d0a111e8e61aa6621b0ba55"} err="failed to get container status \"f28126a9814c3065aa169bef80450931687265878d0a111e8e61aa6621b0ba55\": rpc error: code = NotFound desc = could not find container \"f28126a9814c3065aa169bef80450931687265878d0a111e8e61aa6621b0ba55\": container with ID starting with f28126a9814c3065aa169bef80450931687265878d0a111e8e61aa6621b0ba55 not found: ID does not exist" Nov 28 13:55:55 crc kubenswrapper[4631]: I1128 13:55:55.495900 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1f1e9bbe-987c-4acd-85f3-e22146440472-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1f1e9bbe-987c-4acd-85f3-e22146440472" (UID: "1f1e9bbe-987c-4acd-85f3-e22146440472"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 13:55:55 crc kubenswrapper[4631]: I1128 13:55:55.516756 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bbftn\" (UniqueName: \"kubernetes.io/projected/1f1e9bbe-987c-4acd-85f3-e22146440472-kube-api-access-bbftn\") on node \"crc\" DevicePath \"\"" Nov 28 13:55:55 crc kubenswrapper[4631]: I1128 13:55:55.516791 4631 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1f1e9bbe-987c-4acd-85f3-e22146440472-utilities\") on node \"crc\" DevicePath \"\"" Nov 28 13:55:55 crc kubenswrapper[4631]: I1128 13:55:55.516801 4631 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1f1e9bbe-987c-4acd-85f3-e22146440472-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 28 13:55:55 crc kubenswrapper[4631]: I1128 13:55:55.556035 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-42b7x"] Nov 28 13:55:55 crc kubenswrapper[4631]: E1128 13:55:55.556463 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1f1e9bbe-987c-4acd-85f3-e22146440472" containerName="extract-content" Nov 28 13:55:55 crc kubenswrapper[4631]: I1128 13:55:55.556477 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="1f1e9bbe-987c-4acd-85f3-e22146440472" containerName="extract-content" Nov 28 13:55:55 crc kubenswrapper[4631]: E1128 13:55:55.556506 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1f1e9bbe-987c-4acd-85f3-e22146440472" containerName="registry-server" Nov 28 13:55:55 crc kubenswrapper[4631]: I1128 13:55:55.556516 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="1f1e9bbe-987c-4acd-85f3-e22146440472" containerName="registry-server" Nov 28 13:55:55 crc kubenswrapper[4631]: E1128 13:55:55.556530 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1f1e9bbe-987c-4acd-85f3-e22146440472" containerName="extract-utilities" Nov 28 13:55:55 crc kubenswrapper[4631]: I1128 13:55:55.556536 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="1f1e9bbe-987c-4acd-85f3-e22146440472" containerName="extract-utilities" Nov 28 13:55:55 crc kubenswrapper[4631]: I1128 13:55:55.556720 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="1f1e9bbe-987c-4acd-85f3-e22146440472" containerName="registry-server" Nov 28 13:55:55 crc kubenswrapper[4631]: I1128 13:55:55.560246 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-42b7x"] Nov 28 13:55:55 crc kubenswrapper[4631]: I1128 13:55:55.560419 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-42b7x" Nov 28 13:55:55 crc kubenswrapper[4631]: I1128 13:55:55.632591 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-fdd9l"] Nov 28 13:55:55 crc kubenswrapper[4631]: I1128 13:55:55.640819 4631 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-fdd9l"] Nov 28 13:55:55 crc kubenswrapper[4631]: I1128 13:55:55.720718 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8a15f4ac-9cfc-4259-97f7-7083dff7012d-catalog-content\") pod \"certified-operators-42b7x\" (UID: \"8a15f4ac-9cfc-4259-97f7-7083dff7012d\") " pod="openshift-marketplace/certified-operators-42b7x" Nov 28 13:55:55 crc kubenswrapper[4631]: I1128 13:55:55.720940 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4trpw\" (UniqueName: \"kubernetes.io/projected/8a15f4ac-9cfc-4259-97f7-7083dff7012d-kube-api-access-4trpw\") pod \"certified-operators-42b7x\" (UID: \"8a15f4ac-9cfc-4259-97f7-7083dff7012d\") " pod="openshift-marketplace/certified-operators-42b7x" Nov 28 13:55:55 crc kubenswrapper[4631]: I1128 13:55:55.721221 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8a15f4ac-9cfc-4259-97f7-7083dff7012d-utilities\") pod \"certified-operators-42b7x\" (UID: \"8a15f4ac-9cfc-4259-97f7-7083dff7012d\") " pod="openshift-marketplace/certified-operators-42b7x" Nov 28 13:55:55 crc kubenswrapper[4631]: I1128 13:55:55.823873 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4trpw\" (UniqueName: \"kubernetes.io/projected/8a15f4ac-9cfc-4259-97f7-7083dff7012d-kube-api-access-4trpw\") pod \"certified-operators-42b7x\" (UID: \"8a15f4ac-9cfc-4259-97f7-7083dff7012d\") " pod="openshift-marketplace/certified-operators-42b7x" Nov 28 13:55:55 crc kubenswrapper[4631]: I1128 13:55:55.824018 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8a15f4ac-9cfc-4259-97f7-7083dff7012d-utilities\") pod \"certified-operators-42b7x\" (UID: \"8a15f4ac-9cfc-4259-97f7-7083dff7012d\") " pod="openshift-marketplace/certified-operators-42b7x" Nov 28 13:55:55 crc kubenswrapper[4631]: I1128 13:55:55.824450 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8a15f4ac-9cfc-4259-97f7-7083dff7012d-catalog-content\") pod \"certified-operators-42b7x\" (UID: \"8a15f4ac-9cfc-4259-97f7-7083dff7012d\") " pod="openshift-marketplace/certified-operators-42b7x" Nov 28 13:55:55 crc kubenswrapper[4631]: I1128 13:55:55.824800 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8a15f4ac-9cfc-4259-97f7-7083dff7012d-utilities\") pod \"certified-operators-42b7x\" (UID: \"8a15f4ac-9cfc-4259-97f7-7083dff7012d\") " pod="openshift-marketplace/certified-operators-42b7x" Nov 28 13:55:55 crc kubenswrapper[4631]: I1128 13:55:55.824959 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8a15f4ac-9cfc-4259-97f7-7083dff7012d-catalog-content\") pod \"certified-operators-42b7x\" (UID: \"8a15f4ac-9cfc-4259-97f7-7083dff7012d\") " pod="openshift-marketplace/certified-operators-42b7x" Nov 28 13:55:55 crc kubenswrapper[4631]: I1128 13:55:55.855176 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4trpw\" (UniqueName: \"kubernetes.io/projected/8a15f4ac-9cfc-4259-97f7-7083dff7012d-kube-api-access-4trpw\") pod \"certified-operators-42b7x\" (UID: \"8a15f4ac-9cfc-4259-97f7-7083dff7012d\") " pod="openshift-marketplace/certified-operators-42b7x" Nov 28 13:55:55 crc kubenswrapper[4631]: I1128 13:55:55.883765 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-42b7x" Nov 28 13:55:56 crc kubenswrapper[4631]: I1128 13:55:56.559866 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-42b7x"] Nov 28 13:55:57 crc kubenswrapper[4631]: I1128 13:55:57.358862 4631 generic.go:334] "Generic (PLEG): container finished" podID="8a15f4ac-9cfc-4259-97f7-7083dff7012d" containerID="7d427a6ccb75bb44892098ba577df18f29d869594fd3814889a246f5262ed8d6" exitCode=0 Nov 28 13:55:57 crc kubenswrapper[4631]: I1128 13:55:57.358950 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-42b7x" event={"ID":"8a15f4ac-9cfc-4259-97f7-7083dff7012d","Type":"ContainerDied","Data":"7d427a6ccb75bb44892098ba577df18f29d869594fd3814889a246f5262ed8d6"} Nov 28 13:55:57 crc kubenswrapper[4631]: I1128 13:55:57.359180 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-42b7x" event={"ID":"8a15f4ac-9cfc-4259-97f7-7083dff7012d","Type":"ContainerStarted","Data":"f4490107b93fd5ebb2d16f5e9982354ab6afa2cd150f0c5ed657436d2d88a506"} Nov 28 13:55:57 crc kubenswrapper[4631]: I1128 13:55:57.524618 4631 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1f1e9bbe-987c-4acd-85f3-e22146440472" path="/var/lib/kubelet/pods/1f1e9bbe-987c-4acd-85f3-e22146440472/volumes" Nov 28 13:55:59 crc kubenswrapper[4631]: I1128 13:55:59.381274 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-42b7x" event={"ID":"8a15f4ac-9cfc-4259-97f7-7083dff7012d","Type":"ContainerStarted","Data":"2769493690e5348d345c038306b771501d8a9dcf5fdd513f20338bf0fd219361"} Nov 28 13:56:00 crc kubenswrapper[4631]: I1128 13:56:00.390230 4631 generic.go:334] "Generic (PLEG): container finished" podID="8a15f4ac-9cfc-4259-97f7-7083dff7012d" containerID="2769493690e5348d345c038306b771501d8a9dcf5fdd513f20338bf0fd219361" exitCode=0 Nov 28 13:56:00 crc kubenswrapper[4631]: I1128 13:56:00.390326 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-42b7x" event={"ID":"8a15f4ac-9cfc-4259-97f7-7083dff7012d","Type":"ContainerDied","Data":"2769493690e5348d345c038306b771501d8a9dcf5fdd513f20338bf0fd219361"} Nov 28 13:56:01 crc kubenswrapper[4631]: I1128 13:56:01.415686 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-42b7x" event={"ID":"8a15f4ac-9cfc-4259-97f7-7083dff7012d","Type":"ContainerStarted","Data":"ef679a0a530abb296fe5e979af445c2cb8b8c53b95b95e6430c5dd1a12fd66d5"} Nov 28 13:56:01 crc kubenswrapper[4631]: I1128 13:56:01.443324 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-42b7x" podStartSLOduration=2.714633724 podStartE2EDuration="6.443272115s" podCreationTimestamp="2025-11-28 13:55:55 +0000 UTC" firstStartedPulling="2025-11-28 13:55:57.360739637 +0000 UTC m=+2114.168042981" lastFinishedPulling="2025-11-28 13:56:01.089378028 +0000 UTC m=+2117.896681372" observedRunningTime="2025-11-28 13:56:01.43976667 +0000 UTC m=+2118.247070034" watchObservedRunningTime="2025-11-28 13:56:01.443272115 +0000 UTC m=+2118.250575469" Nov 28 13:56:05 crc kubenswrapper[4631]: I1128 13:56:05.884565 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-42b7x" Nov 28 13:56:05 crc kubenswrapper[4631]: I1128 13:56:05.885302 4631 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-42b7x" Nov 28 13:56:05 crc kubenswrapper[4631]: I1128 13:56:05.934663 4631 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-42b7x" Nov 28 13:56:06 crc kubenswrapper[4631]: I1128 13:56:06.515099 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-42b7x" Nov 28 13:56:06 crc kubenswrapper[4631]: I1128 13:56:06.570511 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-42b7x"] Nov 28 13:56:08 crc kubenswrapper[4631]: I1128 13:56:08.478989 4631 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-42b7x" podUID="8a15f4ac-9cfc-4259-97f7-7083dff7012d" containerName="registry-server" containerID="cri-o://ef679a0a530abb296fe5e979af445c2cb8b8c53b95b95e6430c5dd1a12fd66d5" gracePeriod=2 Nov 28 13:56:08 crc kubenswrapper[4631]: I1128 13:56:08.945501 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-42b7x" Nov 28 13:56:09 crc kubenswrapper[4631]: I1128 13:56:09.082597 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8a15f4ac-9cfc-4259-97f7-7083dff7012d-catalog-content\") pod \"8a15f4ac-9cfc-4259-97f7-7083dff7012d\" (UID: \"8a15f4ac-9cfc-4259-97f7-7083dff7012d\") " Nov 28 13:56:09 crc kubenswrapper[4631]: I1128 13:56:09.083269 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4trpw\" (UniqueName: \"kubernetes.io/projected/8a15f4ac-9cfc-4259-97f7-7083dff7012d-kube-api-access-4trpw\") pod \"8a15f4ac-9cfc-4259-97f7-7083dff7012d\" (UID: \"8a15f4ac-9cfc-4259-97f7-7083dff7012d\") " Nov 28 13:56:09 crc kubenswrapper[4631]: I1128 13:56:09.083647 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8a15f4ac-9cfc-4259-97f7-7083dff7012d-utilities\") pod \"8a15f4ac-9cfc-4259-97f7-7083dff7012d\" (UID: \"8a15f4ac-9cfc-4259-97f7-7083dff7012d\") " Nov 28 13:56:09 crc kubenswrapper[4631]: I1128 13:56:09.086162 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8a15f4ac-9cfc-4259-97f7-7083dff7012d-utilities" (OuterVolumeSpecName: "utilities") pod "8a15f4ac-9cfc-4259-97f7-7083dff7012d" (UID: "8a15f4ac-9cfc-4259-97f7-7083dff7012d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 13:56:09 crc kubenswrapper[4631]: I1128 13:56:09.090104 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8a15f4ac-9cfc-4259-97f7-7083dff7012d-kube-api-access-4trpw" (OuterVolumeSpecName: "kube-api-access-4trpw") pod "8a15f4ac-9cfc-4259-97f7-7083dff7012d" (UID: "8a15f4ac-9cfc-4259-97f7-7083dff7012d"). InnerVolumeSpecName "kube-api-access-4trpw". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:56:09 crc kubenswrapper[4631]: I1128 13:56:09.135620 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8a15f4ac-9cfc-4259-97f7-7083dff7012d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8a15f4ac-9cfc-4259-97f7-7083dff7012d" (UID: "8a15f4ac-9cfc-4259-97f7-7083dff7012d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 13:56:09 crc kubenswrapper[4631]: I1128 13:56:09.187966 4631 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8a15f4ac-9cfc-4259-97f7-7083dff7012d-utilities\") on node \"crc\" DevicePath \"\"" Nov 28 13:56:09 crc kubenswrapper[4631]: I1128 13:56:09.188012 4631 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8a15f4ac-9cfc-4259-97f7-7083dff7012d-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 28 13:56:09 crc kubenswrapper[4631]: I1128 13:56:09.188029 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4trpw\" (UniqueName: \"kubernetes.io/projected/8a15f4ac-9cfc-4259-97f7-7083dff7012d-kube-api-access-4trpw\") on node \"crc\" DevicePath \"\"" Nov 28 13:56:09 crc kubenswrapper[4631]: I1128 13:56:09.489192 4631 generic.go:334] "Generic (PLEG): container finished" podID="8a15f4ac-9cfc-4259-97f7-7083dff7012d" containerID="ef679a0a530abb296fe5e979af445c2cb8b8c53b95b95e6430c5dd1a12fd66d5" exitCode=0 Nov 28 13:56:09 crc kubenswrapper[4631]: I1128 13:56:09.489247 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-42b7x" event={"ID":"8a15f4ac-9cfc-4259-97f7-7083dff7012d","Type":"ContainerDied","Data":"ef679a0a530abb296fe5e979af445c2cb8b8c53b95b95e6430c5dd1a12fd66d5"} Nov 28 13:56:09 crc kubenswrapper[4631]: I1128 13:56:09.489274 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-42b7x" event={"ID":"8a15f4ac-9cfc-4259-97f7-7083dff7012d","Type":"ContainerDied","Data":"f4490107b93fd5ebb2d16f5e9982354ab6afa2cd150f0c5ed657436d2d88a506"} Nov 28 13:56:09 crc kubenswrapper[4631]: I1128 13:56:09.489331 4631 scope.go:117] "RemoveContainer" containerID="ef679a0a530abb296fe5e979af445c2cb8b8c53b95b95e6430c5dd1a12fd66d5" Nov 28 13:56:09 crc kubenswrapper[4631]: I1128 13:56:09.489462 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-42b7x" Nov 28 13:56:09 crc kubenswrapper[4631]: I1128 13:56:09.522920 4631 scope.go:117] "RemoveContainer" containerID="2769493690e5348d345c038306b771501d8a9dcf5fdd513f20338bf0fd219361" Nov 28 13:56:09 crc kubenswrapper[4631]: I1128 13:56:09.567615 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-42b7x"] Nov 28 13:56:09 crc kubenswrapper[4631]: I1128 13:56:09.570127 4631 scope.go:117] "RemoveContainer" containerID="7d427a6ccb75bb44892098ba577df18f29d869594fd3814889a246f5262ed8d6" Nov 28 13:56:09 crc kubenswrapper[4631]: I1128 13:56:09.589037 4631 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-42b7x"] Nov 28 13:56:09 crc kubenswrapper[4631]: I1128 13:56:09.610099 4631 scope.go:117] "RemoveContainer" containerID="ef679a0a530abb296fe5e979af445c2cb8b8c53b95b95e6430c5dd1a12fd66d5" Nov 28 13:56:09 crc kubenswrapper[4631]: E1128 13:56:09.610909 4631 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ef679a0a530abb296fe5e979af445c2cb8b8c53b95b95e6430c5dd1a12fd66d5\": container with ID starting with ef679a0a530abb296fe5e979af445c2cb8b8c53b95b95e6430c5dd1a12fd66d5 not found: ID does not exist" containerID="ef679a0a530abb296fe5e979af445c2cb8b8c53b95b95e6430c5dd1a12fd66d5" Nov 28 13:56:09 crc kubenswrapper[4631]: I1128 13:56:09.610968 4631 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ef679a0a530abb296fe5e979af445c2cb8b8c53b95b95e6430c5dd1a12fd66d5"} err="failed to get container status \"ef679a0a530abb296fe5e979af445c2cb8b8c53b95b95e6430c5dd1a12fd66d5\": rpc error: code = NotFound desc = could not find container \"ef679a0a530abb296fe5e979af445c2cb8b8c53b95b95e6430c5dd1a12fd66d5\": container with ID starting with ef679a0a530abb296fe5e979af445c2cb8b8c53b95b95e6430c5dd1a12fd66d5 not found: ID does not exist" Nov 28 13:56:09 crc kubenswrapper[4631]: I1128 13:56:09.611142 4631 scope.go:117] "RemoveContainer" containerID="2769493690e5348d345c038306b771501d8a9dcf5fdd513f20338bf0fd219361" Nov 28 13:56:09 crc kubenswrapper[4631]: E1128 13:56:09.611535 4631 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2769493690e5348d345c038306b771501d8a9dcf5fdd513f20338bf0fd219361\": container with ID starting with 2769493690e5348d345c038306b771501d8a9dcf5fdd513f20338bf0fd219361 not found: ID does not exist" containerID="2769493690e5348d345c038306b771501d8a9dcf5fdd513f20338bf0fd219361" Nov 28 13:56:09 crc kubenswrapper[4631]: I1128 13:56:09.611636 4631 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2769493690e5348d345c038306b771501d8a9dcf5fdd513f20338bf0fd219361"} err="failed to get container status \"2769493690e5348d345c038306b771501d8a9dcf5fdd513f20338bf0fd219361\": rpc error: code = NotFound desc = could not find container \"2769493690e5348d345c038306b771501d8a9dcf5fdd513f20338bf0fd219361\": container with ID starting with 2769493690e5348d345c038306b771501d8a9dcf5fdd513f20338bf0fd219361 not found: ID does not exist" Nov 28 13:56:09 crc kubenswrapper[4631]: I1128 13:56:09.611656 4631 scope.go:117] "RemoveContainer" containerID="7d427a6ccb75bb44892098ba577df18f29d869594fd3814889a246f5262ed8d6" Nov 28 13:56:09 crc kubenswrapper[4631]: E1128 13:56:09.613463 4631 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7d427a6ccb75bb44892098ba577df18f29d869594fd3814889a246f5262ed8d6\": container with ID starting with 7d427a6ccb75bb44892098ba577df18f29d869594fd3814889a246f5262ed8d6 not found: ID does not exist" containerID="7d427a6ccb75bb44892098ba577df18f29d869594fd3814889a246f5262ed8d6" Nov 28 13:56:09 crc kubenswrapper[4631]: I1128 13:56:09.613505 4631 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7d427a6ccb75bb44892098ba577df18f29d869594fd3814889a246f5262ed8d6"} err="failed to get container status \"7d427a6ccb75bb44892098ba577df18f29d869594fd3814889a246f5262ed8d6\": rpc error: code = NotFound desc = could not find container \"7d427a6ccb75bb44892098ba577df18f29d869594fd3814889a246f5262ed8d6\": container with ID starting with 7d427a6ccb75bb44892098ba577df18f29d869594fd3814889a246f5262ed8d6 not found: ID does not exist" Nov 28 13:56:11 crc kubenswrapper[4631]: I1128 13:56:11.525181 4631 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8a15f4ac-9cfc-4259-97f7-7083dff7012d" path="/var/lib/kubelet/pods/8a15f4ac-9cfc-4259-97f7-7083dff7012d/volumes" Nov 28 13:56:13 crc kubenswrapper[4631]: I1128 13:56:13.527657 4631 generic.go:334] "Generic (PLEG): container finished" podID="a73f7b13-ead0-46a0-8c62-d3c1e75bc2d8" containerID="cdc53e23a08c8ee3166c0454e0a42412ecbc0fd4a366ad977f81e8922fb061ce" exitCode=0 Nov 28 13:56:13 crc kubenswrapper[4631]: I1128 13:56:13.527736 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-tl8jt" event={"ID":"a73f7b13-ead0-46a0-8c62-d3c1e75bc2d8","Type":"ContainerDied","Data":"cdc53e23a08c8ee3166c0454e0a42412ecbc0fd4a366ad977f81e8922fb061ce"} Nov 28 13:56:14 crc kubenswrapper[4631]: I1128 13:56:14.938452 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-tl8jt" Nov 28 13:56:14 crc kubenswrapper[4631]: I1128 13:56:14.995021 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/a73f7b13-ead0-46a0-8c62-d3c1e75bc2d8-neutron-ovn-metadata-agent-neutron-config-0\") pod \"a73f7b13-ead0-46a0-8c62-d3c1e75bc2d8\" (UID: \"a73f7b13-ead0-46a0-8c62-d3c1e75bc2d8\") " Nov 28 13:56:14 crc kubenswrapper[4631]: I1128 13:56:14.995101 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a73f7b13-ead0-46a0-8c62-d3c1e75bc2d8-neutron-metadata-combined-ca-bundle\") pod \"a73f7b13-ead0-46a0-8c62-d3c1e75bc2d8\" (UID: \"a73f7b13-ead0-46a0-8c62-d3c1e75bc2d8\") " Nov 28 13:56:14 crc kubenswrapper[4631]: I1128 13:56:14.995141 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a73f7b13-ead0-46a0-8c62-d3c1e75bc2d8-inventory\") pod \"a73f7b13-ead0-46a0-8c62-d3c1e75bc2d8\" (UID: \"a73f7b13-ead0-46a0-8c62-d3c1e75bc2d8\") " Nov 28 13:56:14 crc kubenswrapper[4631]: I1128 13:56:14.995193 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a73f7b13-ead0-46a0-8c62-d3c1e75bc2d8-ssh-key\") pod \"a73f7b13-ead0-46a0-8c62-d3c1e75bc2d8\" (UID: \"a73f7b13-ead0-46a0-8c62-d3c1e75bc2d8\") " Nov 28 13:56:14 crc kubenswrapper[4631]: I1128 13:56:14.995276 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/a73f7b13-ead0-46a0-8c62-d3c1e75bc2d8-nova-metadata-neutron-config-0\") pod \"a73f7b13-ead0-46a0-8c62-d3c1e75bc2d8\" (UID: \"a73f7b13-ead0-46a0-8c62-d3c1e75bc2d8\") " Nov 28 13:56:14 crc kubenswrapper[4631]: I1128 13:56:14.995445 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4c4pz\" (UniqueName: \"kubernetes.io/projected/a73f7b13-ead0-46a0-8c62-d3c1e75bc2d8-kube-api-access-4c4pz\") pod \"a73f7b13-ead0-46a0-8c62-d3c1e75bc2d8\" (UID: \"a73f7b13-ead0-46a0-8c62-d3c1e75bc2d8\") " Nov 28 13:56:15 crc kubenswrapper[4631]: I1128 13:56:15.002384 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a73f7b13-ead0-46a0-8c62-d3c1e75bc2d8-kube-api-access-4c4pz" (OuterVolumeSpecName: "kube-api-access-4c4pz") pod "a73f7b13-ead0-46a0-8c62-d3c1e75bc2d8" (UID: "a73f7b13-ead0-46a0-8c62-d3c1e75bc2d8"). InnerVolumeSpecName "kube-api-access-4c4pz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:56:15 crc kubenswrapper[4631]: I1128 13:56:15.014118 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a73f7b13-ead0-46a0-8c62-d3c1e75bc2d8-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "a73f7b13-ead0-46a0-8c62-d3c1e75bc2d8" (UID: "a73f7b13-ead0-46a0-8c62-d3c1e75bc2d8"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:56:15 crc kubenswrapper[4631]: I1128 13:56:15.055021 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a73f7b13-ead0-46a0-8c62-d3c1e75bc2d8-neutron-ovn-metadata-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-ovn-metadata-agent-neutron-config-0") pod "a73f7b13-ead0-46a0-8c62-d3c1e75bc2d8" (UID: "a73f7b13-ead0-46a0-8c62-d3c1e75bc2d8"). InnerVolumeSpecName "neutron-ovn-metadata-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:56:15 crc kubenswrapper[4631]: I1128 13:56:15.059018 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a73f7b13-ead0-46a0-8c62-d3c1e75bc2d8-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "a73f7b13-ead0-46a0-8c62-d3c1e75bc2d8" (UID: "a73f7b13-ead0-46a0-8c62-d3c1e75bc2d8"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:56:15 crc kubenswrapper[4631]: I1128 13:56:15.064522 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a73f7b13-ead0-46a0-8c62-d3c1e75bc2d8-inventory" (OuterVolumeSpecName: "inventory") pod "a73f7b13-ead0-46a0-8c62-d3c1e75bc2d8" (UID: "a73f7b13-ead0-46a0-8c62-d3c1e75bc2d8"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:56:15 crc kubenswrapper[4631]: I1128 13:56:15.073524 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a73f7b13-ead0-46a0-8c62-d3c1e75bc2d8-nova-metadata-neutron-config-0" (OuterVolumeSpecName: "nova-metadata-neutron-config-0") pod "a73f7b13-ead0-46a0-8c62-d3c1e75bc2d8" (UID: "a73f7b13-ead0-46a0-8c62-d3c1e75bc2d8"). InnerVolumeSpecName "nova-metadata-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 13:56:15 crc kubenswrapper[4631]: I1128 13:56:15.100182 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4c4pz\" (UniqueName: \"kubernetes.io/projected/a73f7b13-ead0-46a0-8c62-d3c1e75bc2d8-kube-api-access-4c4pz\") on node \"crc\" DevicePath \"\"" Nov 28 13:56:15 crc kubenswrapper[4631]: I1128 13:56:15.100219 4631 reconciler_common.go:293] "Volume detached for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/a73f7b13-ead0-46a0-8c62-d3c1e75bc2d8-neutron-ovn-metadata-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Nov 28 13:56:15 crc kubenswrapper[4631]: I1128 13:56:15.100277 4631 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a73f7b13-ead0-46a0-8c62-d3c1e75bc2d8-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 13:56:15 crc kubenswrapper[4631]: I1128 13:56:15.100321 4631 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a73f7b13-ead0-46a0-8c62-d3c1e75bc2d8-inventory\") on node \"crc\" DevicePath \"\"" Nov 28 13:56:15 crc kubenswrapper[4631]: I1128 13:56:15.100335 4631 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a73f7b13-ead0-46a0-8c62-d3c1e75bc2d8-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 28 13:56:15 crc kubenswrapper[4631]: I1128 13:56:15.100346 4631 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/a73f7b13-ead0-46a0-8c62-d3c1e75bc2d8-nova-metadata-neutron-config-0\") on node \"crc\" DevicePath \"\"" Nov 28 13:56:15 crc kubenswrapper[4631]: I1128 13:56:15.585552 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-tl8jt" event={"ID":"a73f7b13-ead0-46a0-8c62-d3c1e75bc2d8","Type":"ContainerDied","Data":"f32d2ba6f015d9bb0358b89b6f01c29301bd16a95fa225b83f3991cb43e970fe"} Nov 28 13:56:15 crc kubenswrapper[4631]: I1128 13:56:15.585598 4631 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f32d2ba6f015d9bb0358b89b6f01c29301bd16a95fa225b83f3991cb43e970fe" Nov 28 13:56:15 crc kubenswrapper[4631]: I1128 13:56:15.585660 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-tl8jt" Nov 28 13:56:15 crc kubenswrapper[4631]: I1128 13:56:15.676543 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-pw8wp"] Nov 28 13:56:15 crc kubenswrapper[4631]: E1128 13:56:15.677168 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a15f4ac-9cfc-4259-97f7-7083dff7012d" containerName="extract-utilities" Nov 28 13:56:15 crc kubenswrapper[4631]: I1128 13:56:15.677301 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a15f4ac-9cfc-4259-97f7-7083dff7012d" containerName="extract-utilities" Nov 28 13:56:15 crc kubenswrapper[4631]: E1128 13:56:15.677376 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a15f4ac-9cfc-4259-97f7-7083dff7012d" containerName="extract-content" Nov 28 13:56:15 crc kubenswrapper[4631]: I1128 13:56:15.677428 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a15f4ac-9cfc-4259-97f7-7083dff7012d" containerName="extract-content" Nov 28 13:56:15 crc kubenswrapper[4631]: E1128 13:56:15.677516 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a15f4ac-9cfc-4259-97f7-7083dff7012d" containerName="registry-server" Nov 28 13:56:15 crc kubenswrapper[4631]: I1128 13:56:15.678059 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a15f4ac-9cfc-4259-97f7-7083dff7012d" containerName="registry-server" Nov 28 13:56:15 crc kubenswrapper[4631]: E1128 13:56:15.678154 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a73f7b13-ead0-46a0-8c62-d3c1e75bc2d8" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Nov 28 13:56:15 crc kubenswrapper[4631]: I1128 13:56:15.678221 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="a73f7b13-ead0-46a0-8c62-d3c1e75bc2d8" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Nov 28 13:56:15 crc kubenswrapper[4631]: I1128 13:56:15.678487 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="a73f7b13-ead0-46a0-8c62-d3c1e75bc2d8" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Nov 28 13:56:15 crc kubenswrapper[4631]: I1128 13:56:15.678557 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="8a15f4ac-9cfc-4259-97f7-7083dff7012d" containerName="registry-server" Nov 28 13:56:15 crc kubenswrapper[4631]: I1128 13:56:15.679572 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-pw8wp" Nov 28 13:56:15 crc kubenswrapper[4631]: I1128 13:56:15.684086 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 28 13:56:15 crc kubenswrapper[4631]: I1128 13:56:15.684402 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 28 13:56:15 crc kubenswrapper[4631]: I1128 13:56:15.684511 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"libvirt-secret" Nov 28 13:56:15 crc kubenswrapper[4631]: I1128 13:56:15.684513 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 28 13:56:15 crc kubenswrapper[4631]: I1128 13:56:15.684758 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-vp4ns" Nov 28 13:56:15 crc kubenswrapper[4631]: I1128 13:56:15.695755 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-pw8wp"] Nov 28 13:56:15 crc kubenswrapper[4631]: I1128 13:56:15.814374 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/86dddd7e-e065-4593-8d47-4d36a2cddb22-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-pw8wp\" (UID: \"86dddd7e-e065-4593-8d47-4d36a2cddb22\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-pw8wp" Nov 28 13:56:15 crc kubenswrapper[4631]: I1128 13:56:15.814767 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/86dddd7e-e065-4593-8d47-4d36a2cddb22-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-pw8wp\" (UID: \"86dddd7e-e065-4593-8d47-4d36a2cddb22\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-pw8wp" Nov 28 13:56:15 crc kubenswrapper[4631]: I1128 13:56:15.814938 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/86dddd7e-e065-4593-8d47-4d36a2cddb22-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-pw8wp\" (UID: \"86dddd7e-e065-4593-8d47-4d36a2cddb22\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-pw8wp" Nov 28 13:56:15 crc kubenswrapper[4631]: I1128 13:56:15.815146 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lrgd6\" (UniqueName: \"kubernetes.io/projected/86dddd7e-e065-4593-8d47-4d36a2cddb22-kube-api-access-lrgd6\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-pw8wp\" (UID: \"86dddd7e-e065-4593-8d47-4d36a2cddb22\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-pw8wp" Nov 28 13:56:15 crc kubenswrapper[4631]: I1128 13:56:15.815181 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/86dddd7e-e065-4593-8d47-4d36a2cddb22-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-pw8wp\" (UID: \"86dddd7e-e065-4593-8d47-4d36a2cddb22\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-pw8wp" Nov 28 13:56:15 crc kubenswrapper[4631]: I1128 13:56:15.918098 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/86dddd7e-e065-4593-8d47-4d36a2cddb22-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-pw8wp\" (UID: \"86dddd7e-e065-4593-8d47-4d36a2cddb22\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-pw8wp" Nov 28 13:56:15 crc kubenswrapper[4631]: I1128 13:56:15.918778 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lrgd6\" (UniqueName: \"kubernetes.io/projected/86dddd7e-e065-4593-8d47-4d36a2cddb22-kube-api-access-lrgd6\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-pw8wp\" (UID: \"86dddd7e-e065-4593-8d47-4d36a2cddb22\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-pw8wp" Nov 28 13:56:15 crc kubenswrapper[4631]: I1128 13:56:15.918803 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/86dddd7e-e065-4593-8d47-4d36a2cddb22-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-pw8wp\" (UID: \"86dddd7e-e065-4593-8d47-4d36a2cddb22\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-pw8wp" Nov 28 13:56:15 crc kubenswrapper[4631]: I1128 13:56:15.918874 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/86dddd7e-e065-4593-8d47-4d36a2cddb22-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-pw8wp\" (UID: \"86dddd7e-e065-4593-8d47-4d36a2cddb22\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-pw8wp" Nov 28 13:56:15 crc kubenswrapper[4631]: I1128 13:56:15.918996 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/86dddd7e-e065-4593-8d47-4d36a2cddb22-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-pw8wp\" (UID: \"86dddd7e-e065-4593-8d47-4d36a2cddb22\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-pw8wp" Nov 28 13:56:15 crc kubenswrapper[4631]: I1128 13:56:15.923131 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/86dddd7e-e065-4593-8d47-4d36a2cddb22-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-pw8wp\" (UID: \"86dddd7e-e065-4593-8d47-4d36a2cddb22\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-pw8wp" Nov 28 13:56:15 crc kubenswrapper[4631]: I1128 13:56:15.924253 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/86dddd7e-e065-4593-8d47-4d36a2cddb22-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-pw8wp\" (UID: \"86dddd7e-e065-4593-8d47-4d36a2cddb22\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-pw8wp" Nov 28 13:56:15 crc kubenswrapper[4631]: I1128 13:56:15.924977 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/86dddd7e-e065-4593-8d47-4d36a2cddb22-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-pw8wp\" (UID: \"86dddd7e-e065-4593-8d47-4d36a2cddb22\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-pw8wp" Nov 28 13:56:15 crc kubenswrapper[4631]: I1128 13:56:15.926682 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/86dddd7e-e065-4593-8d47-4d36a2cddb22-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-pw8wp\" (UID: \"86dddd7e-e065-4593-8d47-4d36a2cddb22\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-pw8wp" Nov 28 13:56:15 crc kubenswrapper[4631]: I1128 13:56:15.938624 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lrgd6\" (UniqueName: \"kubernetes.io/projected/86dddd7e-e065-4593-8d47-4d36a2cddb22-kube-api-access-lrgd6\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-pw8wp\" (UID: \"86dddd7e-e065-4593-8d47-4d36a2cddb22\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-pw8wp" Nov 28 13:56:16 crc kubenswrapper[4631]: I1128 13:56:16.017340 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-pw8wp" Nov 28 13:56:16 crc kubenswrapper[4631]: I1128 13:56:16.450943 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-pw8wp"] Nov 28 13:56:16 crc kubenswrapper[4631]: I1128 13:56:16.595812 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-pw8wp" event={"ID":"86dddd7e-e065-4593-8d47-4d36a2cddb22","Type":"ContainerStarted","Data":"5ef1ebedaece586346c179e0edf6486a456933646eec9e4130be8acba5475d01"} Nov 28 13:56:17 crc kubenswrapper[4631]: I1128 13:56:17.657667 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-pw8wp" event={"ID":"86dddd7e-e065-4593-8d47-4d36a2cddb22","Type":"ContainerStarted","Data":"591422d62747b91572874434cfb086d5fe5ebba0af3980e91a9e5a321435801c"} Nov 28 13:56:19 crc kubenswrapper[4631]: I1128 13:56:19.635399 4631 patch_prober.go:28] interesting pod/machine-config-daemon-47sxw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 28 13:56:19 crc kubenswrapper[4631]: I1128 13:56:19.636957 4631 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" podUID="8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 28 13:56:19 crc kubenswrapper[4631]: I1128 13:56:19.637132 4631 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" Nov 28 13:56:19 crc kubenswrapper[4631]: I1128 13:56:19.638192 4631 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"f6f69c42ec6f8171f119b80c48682e57400641797951b8d603755d1b6a845e27"} pod="openshift-machine-config-operator/machine-config-daemon-47sxw" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 28 13:56:19 crc kubenswrapper[4631]: I1128 13:56:19.638364 4631 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" podUID="8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1" containerName="machine-config-daemon" containerID="cri-o://f6f69c42ec6f8171f119b80c48682e57400641797951b8d603755d1b6a845e27" gracePeriod=600 Nov 28 13:56:20 crc kubenswrapper[4631]: I1128 13:56:20.714497 4631 generic.go:334] "Generic (PLEG): container finished" podID="8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1" containerID="f6f69c42ec6f8171f119b80c48682e57400641797951b8d603755d1b6a845e27" exitCode=0 Nov 28 13:56:20 crc kubenswrapper[4631]: I1128 13:56:20.714550 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" event={"ID":"8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1","Type":"ContainerDied","Data":"f6f69c42ec6f8171f119b80c48682e57400641797951b8d603755d1b6a845e27"} Nov 28 13:56:20 crc kubenswrapper[4631]: I1128 13:56:20.714583 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" event={"ID":"8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1","Type":"ContainerStarted","Data":"b4f03963475b3c249f705cef33ea3485fc5d4cec16ae5d211e2ba99ac0014be9"} Nov 28 13:56:20 crc kubenswrapper[4631]: I1128 13:56:20.714606 4631 scope.go:117] "RemoveContainer" containerID="54d38280b4f1c123c61719ed71e612b2bf974f3bf3d8ff13fe19d70fa5ca5198" Nov 28 13:56:20 crc kubenswrapper[4631]: I1128 13:56:20.753941 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-pw8wp" podStartSLOduration=5.267929623 podStartE2EDuration="5.753907936s" podCreationTimestamp="2025-11-28 13:56:15 +0000 UTC" firstStartedPulling="2025-11-28 13:56:16.457515129 +0000 UTC m=+2133.264818473" lastFinishedPulling="2025-11-28 13:56:16.943493442 +0000 UTC m=+2133.750796786" observedRunningTime="2025-11-28 13:56:17.702648784 +0000 UTC m=+2134.509952118" watchObservedRunningTime="2025-11-28 13:56:20.753907936 +0000 UTC m=+2137.561211280" Nov 28 13:56:42 crc kubenswrapper[4631]: I1128 13:56:42.313952 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-2kpdn"] Nov 28 13:56:42 crc kubenswrapper[4631]: I1128 13:56:42.317111 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2kpdn" Nov 28 13:56:42 crc kubenswrapper[4631]: I1128 13:56:42.360325 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-2kpdn"] Nov 28 13:56:42 crc kubenswrapper[4631]: I1128 13:56:42.409889 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/69785d00-b2ed-4406-9702-8c566b10cef3-utilities\") pod \"redhat-marketplace-2kpdn\" (UID: \"69785d00-b2ed-4406-9702-8c566b10cef3\") " pod="openshift-marketplace/redhat-marketplace-2kpdn" Nov 28 13:56:42 crc kubenswrapper[4631]: I1128 13:56:42.410241 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lrq5j\" (UniqueName: \"kubernetes.io/projected/69785d00-b2ed-4406-9702-8c566b10cef3-kube-api-access-lrq5j\") pod \"redhat-marketplace-2kpdn\" (UID: \"69785d00-b2ed-4406-9702-8c566b10cef3\") " pod="openshift-marketplace/redhat-marketplace-2kpdn" Nov 28 13:56:42 crc kubenswrapper[4631]: I1128 13:56:42.410447 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/69785d00-b2ed-4406-9702-8c566b10cef3-catalog-content\") pod \"redhat-marketplace-2kpdn\" (UID: \"69785d00-b2ed-4406-9702-8c566b10cef3\") " pod="openshift-marketplace/redhat-marketplace-2kpdn" Nov 28 13:56:42 crc kubenswrapper[4631]: I1128 13:56:42.513130 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lrq5j\" (UniqueName: \"kubernetes.io/projected/69785d00-b2ed-4406-9702-8c566b10cef3-kube-api-access-lrq5j\") pod \"redhat-marketplace-2kpdn\" (UID: \"69785d00-b2ed-4406-9702-8c566b10cef3\") " pod="openshift-marketplace/redhat-marketplace-2kpdn" Nov 28 13:56:42 crc kubenswrapper[4631]: I1128 13:56:42.513246 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/69785d00-b2ed-4406-9702-8c566b10cef3-catalog-content\") pod \"redhat-marketplace-2kpdn\" (UID: \"69785d00-b2ed-4406-9702-8c566b10cef3\") " pod="openshift-marketplace/redhat-marketplace-2kpdn" Nov 28 13:56:42 crc kubenswrapper[4631]: I1128 13:56:42.513514 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/69785d00-b2ed-4406-9702-8c566b10cef3-utilities\") pod \"redhat-marketplace-2kpdn\" (UID: \"69785d00-b2ed-4406-9702-8c566b10cef3\") " pod="openshift-marketplace/redhat-marketplace-2kpdn" Nov 28 13:56:42 crc kubenswrapper[4631]: I1128 13:56:42.514077 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/69785d00-b2ed-4406-9702-8c566b10cef3-utilities\") pod \"redhat-marketplace-2kpdn\" (UID: \"69785d00-b2ed-4406-9702-8c566b10cef3\") " pod="openshift-marketplace/redhat-marketplace-2kpdn" Nov 28 13:56:42 crc kubenswrapper[4631]: I1128 13:56:42.514228 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/69785d00-b2ed-4406-9702-8c566b10cef3-catalog-content\") pod \"redhat-marketplace-2kpdn\" (UID: \"69785d00-b2ed-4406-9702-8c566b10cef3\") " pod="openshift-marketplace/redhat-marketplace-2kpdn" Nov 28 13:56:42 crc kubenswrapper[4631]: I1128 13:56:42.537259 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lrq5j\" (UniqueName: \"kubernetes.io/projected/69785d00-b2ed-4406-9702-8c566b10cef3-kube-api-access-lrq5j\") pod \"redhat-marketplace-2kpdn\" (UID: \"69785d00-b2ed-4406-9702-8c566b10cef3\") " pod="openshift-marketplace/redhat-marketplace-2kpdn" Nov 28 13:56:42 crc kubenswrapper[4631]: I1128 13:56:42.640712 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2kpdn" Nov 28 13:56:43 crc kubenswrapper[4631]: W1128 13:56:43.155077 4631 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod69785d00_b2ed_4406_9702_8c566b10cef3.slice/crio-62f1163273fee78951bd52ce986c21915227dfa6a71a185a392641daec22ae9c WatchSource:0}: Error finding container 62f1163273fee78951bd52ce986c21915227dfa6a71a185a392641daec22ae9c: Status 404 returned error can't find the container with id 62f1163273fee78951bd52ce986c21915227dfa6a71a185a392641daec22ae9c Nov 28 13:56:43 crc kubenswrapper[4631]: I1128 13:56:43.158228 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-2kpdn"] Nov 28 13:56:43 crc kubenswrapper[4631]: I1128 13:56:43.931213 4631 generic.go:334] "Generic (PLEG): container finished" podID="69785d00-b2ed-4406-9702-8c566b10cef3" containerID="3895b5ced17b594e83e2e911560b95c1e2df7533ec894b1384ec71380a62b3b9" exitCode=0 Nov 28 13:56:43 crc kubenswrapper[4631]: I1128 13:56:43.931413 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2kpdn" event={"ID":"69785d00-b2ed-4406-9702-8c566b10cef3","Type":"ContainerDied","Data":"3895b5ced17b594e83e2e911560b95c1e2df7533ec894b1384ec71380a62b3b9"} Nov 28 13:56:43 crc kubenswrapper[4631]: I1128 13:56:43.931763 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2kpdn" event={"ID":"69785d00-b2ed-4406-9702-8c566b10cef3","Type":"ContainerStarted","Data":"62f1163273fee78951bd52ce986c21915227dfa6a71a185a392641daec22ae9c"} Nov 28 13:56:45 crc kubenswrapper[4631]: I1128 13:56:45.950089 4631 generic.go:334] "Generic (PLEG): container finished" podID="69785d00-b2ed-4406-9702-8c566b10cef3" containerID="4d707031a3031512db38d88d028b43b66be9c4c77b153c72f3ad690b9b5d35a4" exitCode=0 Nov 28 13:56:45 crc kubenswrapper[4631]: I1128 13:56:45.950183 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2kpdn" event={"ID":"69785d00-b2ed-4406-9702-8c566b10cef3","Type":"ContainerDied","Data":"4d707031a3031512db38d88d028b43b66be9c4c77b153c72f3ad690b9b5d35a4"} Nov 28 13:56:46 crc kubenswrapper[4631]: I1128 13:56:46.966894 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2kpdn" event={"ID":"69785d00-b2ed-4406-9702-8c566b10cef3","Type":"ContainerStarted","Data":"f8061eae63d8fe0a0543f8276ac7aa1a95cd3b2eb956c2f916392ccaa7ccefa3"} Nov 28 13:56:46 crc kubenswrapper[4631]: I1128 13:56:46.990041 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-2kpdn" podStartSLOduration=2.47857072 podStartE2EDuration="4.990012687s" podCreationTimestamp="2025-11-28 13:56:42 +0000 UTC" firstStartedPulling="2025-11-28 13:56:43.934191767 +0000 UTC m=+2160.741495111" lastFinishedPulling="2025-11-28 13:56:46.445633734 +0000 UTC m=+2163.252937078" observedRunningTime="2025-11-28 13:56:46.98474757 +0000 UTC m=+2163.792050914" watchObservedRunningTime="2025-11-28 13:56:46.990012687 +0000 UTC m=+2163.797316041" Nov 28 13:56:52 crc kubenswrapper[4631]: I1128 13:56:52.641608 4631 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-2kpdn" Nov 28 13:56:52 crc kubenswrapper[4631]: I1128 13:56:52.642461 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-2kpdn" Nov 28 13:56:52 crc kubenswrapper[4631]: I1128 13:56:52.698183 4631 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-2kpdn" Nov 28 13:56:53 crc kubenswrapper[4631]: I1128 13:56:53.078937 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-2kpdn" Nov 28 13:56:55 crc kubenswrapper[4631]: I1128 13:56:55.183755 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-2kpdn"] Nov 28 13:56:55 crc kubenswrapper[4631]: I1128 13:56:55.184373 4631 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-2kpdn" podUID="69785d00-b2ed-4406-9702-8c566b10cef3" containerName="registry-server" containerID="cri-o://f8061eae63d8fe0a0543f8276ac7aa1a95cd3b2eb956c2f916392ccaa7ccefa3" gracePeriod=2 Nov 28 13:56:55 crc kubenswrapper[4631]: I1128 13:56:55.664834 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2kpdn" Nov 28 13:56:55 crc kubenswrapper[4631]: I1128 13:56:55.735723 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/69785d00-b2ed-4406-9702-8c566b10cef3-catalog-content\") pod \"69785d00-b2ed-4406-9702-8c566b10cef3\" (UID: \"69785d00-b2ed-4406-9702-8c566b10cef3\") " Nov 28 13:56:55 crc kubenswrapper[4631]: I1128 13:56:55.736382 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/69785d00-b2ed-4406-9702-8c566b10cef3-utilities\") pod \"69785d00-b2ed-4406-9702-8c566b10cef3\" (UID: \"69785d00-b2ed-4406-9702-8c566b10cef3\") " Nov 28 13:56:55 crc kubenswrapper[4631]: I1128 13:56:55.736479 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lrq5j\" (UniqueName: \"kubernetes.io/projected/69785d00-b2ed-4406-9702-8c566b10cef3-kube-api-access-lrq5j\") pod \"69785d00-b2ed-4406-9702-8c566b10cef3\" (UID: \"69785d00-b2ed-4406-9702-8c566b10cef3\") " Nov 28 13:56:55 crc kubenswrapper[4631]: I1128 13:56:55.743655 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/69785d00-b2ed-4406-9702-8c566b10cef3-utilities" (OuterVolumeSpecName: "utilities") pod "69785d00-b2ed-4406-9702-8c566b10cef3" (UID: "69785d00-b2ed-4406-9702-8c566b10cef3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 13:56:55 crc kubenswrapper[4631]: I1128 13:56:55.761215 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/69785d00-b2ed-4406-9702-8c566b10cef3-kube-api-access-lrq5j" (OuterVolumeSpecName: "kube-api-access-lrq5j") pod "69785d00-b2ed-4406-9702-8c566b10cef3" (UID: "69785d00-b2ed-4406-9702-8c566b10cef3"). InnerVolumeSpecName "kube-api-access-lrq5j". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 13:56:55 crc kubenswrapper[4631]: I1128 13:56:55.772374 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/69785d00-b2ed-4406-9702-8c566b10cef3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "69785d00-b2ed-4406-9702-8c566b10cef3" (UID: "69785d00-b2ed-4406-9702-8c566b10cef3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 13:56:55 crc kubenswrapper[4631]: I1128 13:56:55.838727 4631 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/69785d00-b2ed-4406-9702-8c566b10cef3-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 28 13:56:55 crc kubenswrapper[4631]: I1128 13:56:55.839022 4631 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/69785d00-b2ed-4406-9702-8c566b10cef3-utilities\") on node \"crc\" DevicePath \"\"" Nov 28 13:56:55 crc kubenswrapper[4631]: I1128 13:56:55.839094 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lrq5j\" (UniqueName: \"kubernetes.io/projected/69785d00-b2ed-4406-9702-8c566b10cef3-kube-api-access-lrq5j\") on node \"crc\" DevicePath \"\"" Nov 28 13:56:56 crc kubenswrapper[4631]: I1128 13:56:56.060875 4631 generic.go:334] "Generic (PLEG): container finished" podID="69785d00-b2ed-4406-9702-8c566b10cef3" containerID="f8061eae63d8fe0a0543f8276ac7aa1a95cd3b2eb956c2f916392ccaa7ccefa3" exitCode=0 Nov 28 13:56:56 crc kubenswrapper[4631]: I1128 13:56:56.060947 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2kpdn" Nov 28 13:56:56 crc kubenswrapper[4631]: I1128 13:56:56.060970 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2kpdn" event={"ID":"69785d00-b2ed-4406-9702-8c566b10cef3","Type":"ContainerDied","Data":"f8061eae63d8fe0a0543f8276ac7aa1a95cd3b2eb956c2f916392ccaa7ccefa3"} Nov 28 13:56:56 crc kubenswrapper[4631]: I1128 13:56:56.061510 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2kpdn" event={"ID":"69785d00-b2ed-4406-9702-8c566b10cef3","Type":"ContainerDied","Data":"62f1163273fee78951bd52ce986c21915227dfa6a71a185a392641daec22ae9c"} Nov 28 13:56:56 crc kubenswrapper[4631]: I1128 13:56:56.061585 4631 scope.go:117] "RemoveContainer" containerID="f8061eae63d8fe0a0543f8276ac7aa1a95cd3b2eb956c2f916392ccaa7ccefa3" Nov 28 13:56:56 crc kubenswrapper[4631]: I1128 13:56:56.092097 4631 scope.go:117] "RemoveContainer" containerID="4d707031a3031512db38d88d028b43b66be9c4c77b153c72f3ad690b9b5d35a4" Nov 28 13:56:56 crc kubenswrapper[4631]: I1128 13:56:56.102176 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-2kpdn"] Nov 28 13:56:56 crc kubenswrapper[4631]: I1128 13:56:56.111921 4631 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-2kpdn"] Nov 28 13:56:56 crc kubenswrapper[4631]: I1128 13:56:56.114028 4631 scope.go:117] "RemoveContainer" containerID="3895b5ced17b594e83e2e911560b95c1e2df7533ec894b1384ec71380a62b3b9" Nov 28 13:56:56 crc kubenswrapper[4631]: I1128 13:56:56.158652 4631 scope.go:117] "RemoveContainer" containerID="f8061eae63d8fe0a0543f8276ac7aa1a95cd3b2eb956c2f916392ccaa7ccefa3" Nov 28 13:56:56 crc kubenswrapper[4631]: E1128 13:56:56.159709 4631 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f8061eae63d8fe0a0543f8276ac7aa1a95cd3b2eb956c2f916392ccaa7ccefa3\": container with ID starting with f8061eae63d8fe0a0543f8276ac7aa1a95cd3b2eb956c2f916392ccaa7ccefa3 not found: ID does not exist" containerID="f8061eae63d8fe0a0543f8276ac7aa1a95cd3b2eb956c2f916392ccaa7ccefa3" Nov 28 13:56:56 crc kubenswrapper[4631]: I1128 13:56:56.159765 4631 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f8061eae63d8fe0a0543f8276ac7aa1a95cd3b2eb956c2f916392ccaa7ccefa3"} err="failed to get container status \"f8061eae63d8fe0a0543f8276ac7aa1a95cd3b2eb956c2f916392ccaa7ccefa3\": rpc error: code = NotFound desc = could not find container \"f8061eae63d8fe0a0543f8276ac7aa1a95cd3b2eb956c2f916392ccaa7ccefa3\": container with ID starting with f8061eae63d8fe0a0543f8276ac7aa1a95cd3b2eb956c2f916392ccaa7ccefa3 not found: ID does not exist" Nov 28 13:56:56 crc kubenswrapper[4631]: I1128 13:56:56.159798 4631 scope.go:117] "RemoveContainer" containerID="4d707031a3031512db38d88d028b43b66be9c4c77b153c72f3ad690b9b5d35a4" Nov 28 13:56:56 crc kubenswrapper[4631]: E1128 13:56:56.160242 4631 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4d707031a3031512db38d88d028b43b66be9c4c77b153c72f3ad690b9b5d35a4\": container with ID starting with 4d707031a3031512db38d88d028b43b66be9c4c77b153c72f3ad690b9b5d35a4 not found: ID does not exist" containerID="4d707031a3031512db38d88d028b43b66be9c4c77b153c72f3ad690b9b5d35a4" Nov 28 13:56:56 crc kubenswrapper[4631]: I1128 13:56:56.160310 4631 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4d707031a3031512db38d88d028b43b66be9c4c77b153c72f3ad690b9b5d35a4"} err="failed to get container status \"4d707031a3031512db38d88d028b43b66be9c4c77b153c72f3ad690b9b5d35a4\": rpc error: code = NotFound desc = could not find container \"4d707031a3031512db38d88d028b43b66be9c4c77b153c72f3ad690b9b5d35a4\": container with ID starting with 4d707031a3031512db38d88d028b43b66be9c4c77b153c72f3ad690b9b5d35a4 not found: ID does not exist" Nov 28 13:56:56 crc kubenswrapper[4631]: I1128 13:56:56.160325 4631 scope.go:117] "RemoveContainer" containerID="3895b5ced17b594e83e2e911560b95c1e2df7533ec894b1384ec71380a62b3b9" Nov 28 13:56:56 crc kubenswrapper[4631]: E1128 13:56:56.160711 4631 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3895b5ced17b594e83e2e911560b95c1e2df7533ec894b1384ec71380a62b3b9\": container with ID starting with 3895b5ced17b594e83e2e911560b95c1e2df7533ec894b1384ec71380a62b3b9 not found: ID does not exist" containerID="3895b5ced17b594e83e2e911560b95c1e2df7533ec894b1384ec71380a62b3b9" Nov 28 13:56:56 crc kubenswrapper[4631]: I1128 13:56:56.160732 4631 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3895b5ced17b594e83e2e911560b95c1e2df7533ec894b1384ec71380a62b3b9"} err="failed to get container status \"3895b5ced17b594e83e2e911560b95c1e2df7533ec894b1384ec71380a62b3b9\": rpc error: code = NotFound desc = could not find container \"3895b5ced17b594e83e2e911560b95c1e2df7533ec894b1384ec71380a62b3b9\": container with ID starting with 3895b5ced17b594e83e2e911560b95c1e2df7533ec894b1384ec71380a62b3b9 not found: ID does not exist" Nov 28 13:56:57 crc kubenswrapper[4631]: I1128 13:56:57.534153 4631 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="69785d00-b2ed-4406-9702-8c566b10cef3" path="/var/lib/kubelet/pods/69785d00-b2ed-4406-9702-8c566b10cef3/volumes" Nov 28 13:58:19 crc kubenswrapper[4631]: I1128 13:58:19.635211 4631 patch_prober.go:28] interesting pod/machine-config-daemon-47sxw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 28 13:58:19 crc kubenswrapper[4631]: I1128 13:58:19.635789 4631 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" podUID="8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 28 13:58:49 crc kubenswrapper[4631]: I1128 13:58:49.634931 4631 patch_prober.go:28] interesting pod/machine-config-daemon-47sxw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 28 13:58:49 crc kubenswrapper[4631]: I1128 13:58:49.635550 4631 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" podUID="8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 28 13:59:19 crc kubenswrapper[4631]: I1128 13:59:19.635412 4631 patch_prober.go:28] interesting pod/machine-config-daemon-47sxw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 28 13:59:19 crc kubenswrapper[4631]: I1128 13:59:19.635966 4631 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" podUID="8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 28 13:59:19 crc kubenswrapper[4631]: I1128 13:59:19.636020 4631 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" Nov 28 13:59:19 crc kubenswrapper[4631]: I1128 13:59:19.636852 4631 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"b4f03963475b3c249f705cef33ea3485fc5d4cec16ae5d211e2ba99ac0014be9"} pod="openshift-machine-config-operator/machine-config-daemon-47sxw" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 28 13:59:19 crc kubenswrapper[4631]: I1128 13:59:19.636906 4631 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" podUID="8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1" containerName="machine-config-daemon" containerID="cri-o://b4f03963475b3c249f705cef33ea3485fc5d4cec16ae5d211e2ba99ac0014be9" gracePeriod=600 Nov 28 13:59:19 crc kubenswrapper[4631]: E1128 13:59:19.765716 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-47sxw_openshift-machine-config-operator(8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1)\"" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" podUID="8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1" Nov 28 13:59:20 crc kubenswrapper[4631]: I1128 13:59:20.388512 4631 generic.go:334] "Generic (PLEG): container finished" podID="8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1" containerID="b4f03963475b3c249f705cef33ea3485fc5d4cec16ae5d211e2ba99ac0014be9" exitCode=0 Nov 28 13:59:20 crc kubenswrapper[4631]: I1128 13:59:20.388562 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" event={"ID":"8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1","Type":"ContainerDied","Data":"b4f03963475b3c249f705cef33ea3485fc5d4cec16ae5d211e2ba99ac0014be9"} Nov 28 13:59:20 crc kubenswrapper[4631]: I1128 13:59:20.388609 4631 scope.go:117] "RemoveContainer" containerID="f6f69c42ec6f8171f119b80c48682e57400641797951b8d603755d1b6a845e27" Nov 28 13:59:20 crc kubenswrapper[4631]: I1128 13:59:20.389265 4631 scope.go:117] "RemoveContainer" containerID="b4f03963475b3c249f705cef33ea3485fc5d4cec16ae5d211e2ba99ac0014be9" Nov 28 13:59:20 crc kubenswrapper[4631]: E1128 13:59:20.389587 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-47sxw_openshift-machine-config-operator(8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1)\"" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" podUID="8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1" Nov 28 13:59:32 crc kubenswrapper[4631]: I1128 13:59:32.513057 4631 scope.go:117] "RemoveContainer" containerID="b4f03963475b3c249f705cef33ea3485fc5d4cec16ae5d211e2ba99ac0014be9" Nov 28 13:59:32 crc kubenswrapper[4631]: E1128 13:59:32.513621 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-47sxw_openshift-machine-config-operator(8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1)\"" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" podUID="8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1" Nov 28 13:59:45 crc kubenswrapper[4631]: I1128 13:59:45.513371 4631 scope.go:117] "RemoveContainer" containerID="b4f03963475b3c249f705cef33ea3485fc5d4cec16ae5d211e2ba99ac0014be9" Nov 28 13:59:45 crc kubenswrapper[4631]: E1128 13:59:45.514013 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-47sxw_openshift-machine-config-operator(8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1)\"" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" podUID="8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1" Nov 28 13:59:56 crc kubenswrapper[4631]: I1128 13:59:56.513712 4631 scope.go:117] "RemoveContainer" containerID="b4f03963475b3c249f705cef33ea3485fc5d4cec16ae5d211e2ba99ac0014be9" Nov 28 13:59:56 crc kubenswrapper[4631]: E1128 13:59:56.514568 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-47sxw_openshift-machine-config-operator(8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1)\"" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" podUID="8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1" Nov 28 14:00:00 crc kubenswrapper[4631]: I1128 14:00:00.150688 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29405640-5bsch"] Nov 28 14:00:00 crc kubenswrapper[4631]: E1128 14:00:00.151693 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="69785d00-b2ed-4406-9702-8c566b10cef3" containerName="registry-server" Nov 28 14:00:00 crc kubenswrapper[4631]: I1128 14:00:00.151708 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="69785d00-b2ed-4406-9702-8c566b10cef3" containerName="registry-server" Nov 28 14:00:00 crc kubenswrapper[4631]: E1128 14:00:00.151718 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="69785d00-b2ed-4406-9702-8c566b10cef3" containerName="extract-content" Nov 28 14:00:00 crc kubenswrapper[4631]: I1128 14:00:00.151725 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="69785d00-b2ed-4406-9702-8c566b10cef3" containerName="extract-content" Nov 28 14:00:00 crc kubenswrapper[4631]: E1128 14:00:00.151750 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="69785d00-b2ed-4406-9702-8c566b10cef3" containerName="extract-utilities" Nov 28 14:00:00 crc kubenswrapper[4631]: I1128 14:00:00.151756 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="69785d00-b2ed-4406-9702-8c566b10cef3" containerName="extract-utilities" Nov 28 14:00:00 crc kubenswrapper[4631]: I1128 14:00:00.151950 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="69785d00-b2ed-4406-9702-8c566b10cef3" containerName="registry-server" Nov 28 14:00:00 crc kubenswrapper[4631]: I1128 14:00:00.152701 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29405640-5bsch" Nov 28 14:00:00 crc kubenswrapper[4631]: I1128 14:00:00.154621 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 28 14:00:00 crc kubenswrapper[4631]: I1128 14:00:00.155732 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 28 14:00:00 crc kubenswrapper[4631]: I1128 14:00:00.162059 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29405640-5bsch"] Nov 28 14:00:00 crc kubenswrapper[4631]: I1128 14:00:00.309689 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/33e41487-9a4a-49b9-a429-a888d7c4db38-config-volume\") pod \"collect-profiles-29405640-5bsch\" (UID: \"33e41487-9a4a-49b9-a429-a888d7c4db38\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405640-5bsch" Nov 28 14:00:00 crc kubenswrapper[4631]: I1128 14:00:00.309768 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/33e41487-9a4a-49b9-a429-a888d7c4db38-secret-volume\") pod \"collect-profiles-29405640-5bsch\" (UID: \"33e41487-9a4a-49b9-a429-a888d7c4db38\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405640-5bsch" Nov 28 14:00:00 crc kubenswrapper[4631]: I1128 14:00:00.309856 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xp5qk\" (UniqueName: \"kubernetes.io/projected/33e41487-9a4a-49b9-a429-a888d7c4db38-kube-api-access-xp5qk\") pod \"collect-profiles-29405640-5bsch\" (UID: \"33e41487-9a4a-49b9-a429-a888d7c4db38\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405640-5bsch" Nov 28 14:00:00 crc kubenswrapper[4631]: I1128 14:00:00.411912 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xp5qk\" (UniqueName: \"kubernetes.io/projected/33e41487-9a4a-49b9-a429-a888d7c4db38-kube-api-access-xp5qk\") pod \"collect-profiles-29405640-5bsch\" (UID: \"33e41487-9a4a-49b9-a429-a888d7c4db38\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405640-5bsch" Nov 28 14:00:00 crc kubenswrapper[4631]: I1128 14:00:00.412074 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/33e41487-9a4a-49b9-a429-a888d7c4db38-config-volume\") pod \"collect-profiles-29405640-5bsch\" (UID: \"33e41487-9a4a-49b9-a429-a888d7c4db38\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405640-5bsch" Nov 28 14:00:00 crc kubenswrapper[4631]: I1128 14:00:00.412122 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/33e41487-9a4a-49b9-a429-a888d7c4db38-secret-volume\") pod \"collect-profiles-29405640-5bsch\" (UID: \"33e41487-9a4a-49b9-a429-a888d7c4db38\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405640-5bsch" Nov 28 14:00:00 crc kubenswrapper[4631]: I1128 14:00:00.413546 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/33e41487-9a4a-49b9-a429-a888d7c4db38-config-volume\") pod \"collect-profiles-29405640-5bsch\" (UID: \"33e41487-9a4a-49b9-a429-a888d7c4db38\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405640-5bsch" Nov 28 14:00:00 crc kubenswrapper[4631]: I1128 14:00:00.424659 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/33e41487-9a4a-49b9-a429-a888d7c4db38-secret-volume\") pod \"collect-profiles-29405640-5bsch\" (UID: \"33e41487-9a4a-49b9-a429-a888d7c4db38\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405640-5bsch" Nov 28 14:00:00 crc kubenswrapper[4631]: I1128 14:00:00.432616 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xp5qk\" (UniqueName: \"kubernetes.io/projected/33e41487-9a4a-49b9-a429-a888d7c4db38-kube-api-access-xp5qk\") pod \"collect-profiles-29405640-5bsch\" (UID: \"33e41487-9a4a-49b9-a429-a888d7c4db38\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405640-5bsch" Nov 28 14:00:00 crc kubenswrapper[4631]: I1128 14:00:00.474513 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29405640-5bsch" Nov 28 14:00:00 crc kubenswrapper[4631]: I1128 14:00:00.983897 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29405640-5bsch"] Nov 28 14:00:01 crc kubenswrapper[4631]: I1128 14:00:01.816044 4631 generic.go:334] "Generic (PLEG): container finished" podID="33e41487-9a4a-49b9-a429-a888d7c4db38" containerID="cd861850009b7eeb1d28e22c9c54aa7c4e60e9b3ccb6ef9a249c35ff9edd6eda" exitCode=0 Nov 28 14:00:01 crc kubenswrapper[4631]: I1128 14:00:01.816156 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29405640-5bsch" event={"ID":"33e41487-9a4a-49b9-a429-a888d7c4db38","Type":"ContainerDied","Data":"cd861850009b7eeb1d28e22c9c54aa7c4e60e9b3ccb6ef9a249c35ff9edd6eda"} Nov 28 14:00:01 crc kubenswrapper[4631]: I1128 14:00:01.816421 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29405640-5bsch" event={"ID":"33e41487-9a4a-49b9-a429-a888d7c4db38","Type":"ContainerStarted","Data":"88b0ac832afd4215b1cd813fae51fa65097bb64574b103720a1d4b351bd36be4"} Nov 28 14:00:03 crc kubenswrapper[4631]: I1128 14:00:03.148850 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29405640-5bsch" Nov 28 14:00:03 crc kubenswrapper[4631]: I1128 14:00:03.176962 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/33e41487-9a4a-49b9-a429-a888d7c4db38-secret-volume\") pod \"33e41487-9a4a-49b9-a429-a888d7c4db38\" (UID: \"33e41487-9a4a-49b9-a429-a888d7c4db38\") " Nov 28 14:00:03 crc kubenswrapper[4631]: I1128 14:00:03.177322 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/33e41487-9a4a-49b9-a429-a888d7c4db38-config-volume\") pod \"33e41487-9a4a-49b9-a429-a888d7c4db38\" (UID: \"33e41487-9a4a-49b9-a429-a888d7c4db38\") " Nov 28 14:00:03 crc kubenswrapper[4631]: I1128 14:00:03.177437 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xp5qk\" (UniqueName: \"kubernetes.io/projected/33e41487-9a4a-49b9-a429-a888d7c4db38-kube-api-access-xp5qk\") pod \"33e41487-9a4a-49b9-a429-a888d7c4db38\" (UID: \"33e41487-9a4a-49b9-a429-a888d7c4db38\") " Nov 28 14:00:03 crc kubenswrapper[4631]: I1128 14:00:03.177804 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/33e41487-9a4a-49b9-a429-a888d7c4db38-config-volume" (OuterVolumeSpecName: "config-volume") pod "33e41487-9a4a-49b9-a429-a888d7c4db38" (UID: "33e41487-9a4a-49b9-a429-a888d7c4db38"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:00:03 crc kubenswrapper[4631]: I1128 14:00:03.178035 4631 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/33e41487-9a4a-49b9-a429-a888d7c4db38-config-volume\") on node \"crc\" DevicePath \"\"" Nov 28 14:00:03 crc kubenswrapper[4631]: I1128 14:00:03.187529 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/33e41487-9a4a-49b9-a429-a888d7c4db38-kube-api-access-xp5qk" (OuterVolumeSpecName: "kube-api-access-xp5qk") pod "33e41487-9a4a-49b9-a429-a888d7c4db38" (UID: "33e41487-9a4a-49b9-a429-a888d7c4db38"). InnerVolumeSpecName "kube-api-access-xp5qk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:00:03 crc kubenswrapper[4631]: I1128 14:00:03.188987 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/33e41487-9a4a-49b9-a429-a888d7c4db38-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "33e41487-9a4a-49b9-a429-a888d7c4db38" (UID: "33e41487-9a4a-49b9-a429-a888d7c4db38"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:00:03 crc kubenswrapper[4631]: I1128 14:00:03.279190 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xp5qk\" (UniqueName: \"kubernetes.io/projected/33e41487-9a4a-49b9-a429-a888d7c4db38-kube-api-access-xp5qk\") on node \"crc\" DevicePath \"\"" Nov 28 14:00:03 crc kubenswrapper[4631]: I1128 14:00:03.279244 4631 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/33e41487-9a4a-49b9-a429-a888d7c4db38-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 28 14:00:03 crc kubenswrapper[4631]: I1128 14:00:03.836653 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29405640-5bsch" event={"ID":"33e41487-9a4a-49b9-a429-a888d7c4db38","Type":"ContainerDied","Data":"88b0ac832afd4215b1cd813fae51fa65097bb64574b103720a1d4b351bd36be4"} Nov 28 14:00:03 crc kubenswrapper[4631]: I1128 14:00:03.836701 4631 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="88b0ac832afd4215b1cd813fae51fa65097bb64574b103720a1d4b351bd36be4" Nov 28 14:00:03 crc kubenswrapper[4631]: I1128 14:00:03.837040 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29405640-5bsch" Nov 28 14:00:04 crc kubenswrapper[4631]: I1128 14:00:04.216773 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29405595-2v6fk"] Nov 28 14:00:04 crc kubenswrapper[4631]: I1128 14:00:04.226774 4631 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29405595-2v6fk"] Nov 28 14:00:05 crc kubenswrapper[4631]: I1128 14:00:05.536423 4631 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="76a05ca6-a98c-4313-855c-fd6ffaceee4f" path="/var/lib/kubelet/pods/76a05ca6-a98c-4313-855c-fd6ffaceee4f/volumes" Nov 28 14:00:05 crc kubenswrapper[4631]: I1128 14:00:05.783811 4631 scope.go:117] "RemoveContainer" containerID="0f3af793008c1a47e8009812f16cb978e3930b0e22afc292634c25e32a447f21" Nov 28 14:00:07 crc kubenswrapper[4631]: I1128 14:00:07.513676 4631 scope.go:117] "RemoveContainer" containerID="b4f03963475b3c249f705cef33ea3485fc5d4cec16ae5d211e2ba99ac0014be9" Nov 28 14:00:07 crc kubenswrapper[4631]: E1128 14:00:07.514580 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-47sxw_openshift-machine-config-operator(8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1)\"" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" podUID="8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1" Nov 28 14:00:22 crc kubenswrapper[4631]: I1128 14:00:22.514246 4631 scope.go:117] "RemoveContainer" containerID="b4f03963475b3c249f705cef33ea3485fc5d4cec16ae5d211e2ba99ac0014be9" Nov 28 14:00:22 crc kubenswrapper[4631]: E1128 14:00:22.516005 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-47sxw_openshift-machine-config-operator(8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1)\"" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" podUID="8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1" Nov 28 14:00:36 crc kubenswrapper[4631]: I1128 14:00:36.512872 4631 scope.go:117] "RemoveContainer" containerID="b4f03963475b3c249f705cef33ea3485fc5d4cec16ae5d211e2ba99ac0014be9" Nov 28 14:00:36 crc kubenswrapper[4631]: E1128 14:00:36.513718 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-47sxw_openshift-machine-config-operator(8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1)\"" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" podUID="8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1" Nov 28 14:00:47 crc kubenswrapper[4631]: I1128 14:00:47.512757 4631 scope.go:117] "RemoveContainer" containerID="b4f03963475b3c249f705cef33ea3485fc5d4cec16ae5d211e2ba99ac0014be9" Nov 28 14:00:47 crc kubenswrapper[4631]: E1128 14:00:47.513760 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-47sxw_openshift-machine-config-operator(8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1)\"" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" podUID="8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1" Nov 28 14:01:00 crc kubenswrapper[4631]: I1128 14:01:00.152027 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29405641-ngxl6"] Nov 28 14:01:00 crc kubenswrapper[4631]: E1128 14:01:00.154426 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="33e41487-9a4a-49b9-a429-a888d7c4db38" containerName="collect-profiles" Nov 28 14:01:00 crc kubenswrapper[4631]: I1128 14:01:00.154630 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="33e41487-9a4a-49b9-a429-a888d7c4db38" containerName="collect-profiles" Nov 28 14:01:00 crc kubenswrapper[4631]: I1128 14:01:00.155140 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="33e41487-9a4a-49b9-a429-a888d7c4db38" containerName="collect-profiles" Nov 28 14:01:00 crc kubenswrapper[4631]: I1128 14:01:00.155959 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29405641-ngxl6" Nov 28 14:01:00 crc kubenswrapper[4631]: I1128 14:01:00.167016 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29405641-ngxl6"] Nov 28 14:01:00 crc kubenswrapper[4631]: I1128 14:01:00.318965 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2080b2bc-31ce-4061-abc1-6f044c10355b-config-data\") pod \"keystone-cron-29405641-ngxl6\" (UID: \"2080b2bc-31ce-4061-abc1-6f044c10355b\") " pod="openstack/keystone-cron-29405641-ngxl6" Nov 28 14:01:00 crc kubenswrapper[4631]: I1128 14:01:00.319037 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4h4bh\" (UniqueName: \"kubernetes.io/projected/2080b2bc-31ce-4061-abc1-6f044c10355b-kube-api-access-4h4bh\") pod \"keystone-cron-29405641-ngxl6\" (UID: \"2080b2bc-31ce-4061-abc1-6f044c10355b\") " pod="openstack/keystone-cron-29405641-ngxl6" Nov 28 14:01:00 crc kubenswrapper[4631]: I1128 14:01:00.319082 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/2080b2bc-31ce-4061-abc1-6f044c10355b-fernet-keys\") pod \"keystone-cron-29405641-ngxl6\" (UID: \"2080b2bc-31ce-4061-abc1-6f044c10355b\") " pod="openstack/keystone-cron-29405641-ngxl6" Nov 28 14:01:00 crc kubenswrapper[4631]: I1128 14:01:00.319111 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2080b2bc-31ce-4061-abc1-6f044c10355b-combined-ca-bundle\") pod \"keystone-cron-29405641-ngxl6\" (UID: \"2080b2bc-31ce-4061-abc1-6f044c10355b\") " pod="openstack/keystone-cron-29405641-ngxl6" Nov 28 14:01:00 crc kubenswrapper[4631]: I1128 14:01:00.420734 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2080b2bc-31ce-4061-abc1-6f044c10355b-config-data\") pod \"keystone-cron-29405641-ngxl6\" (UID: \"2080b2bc-31ce-4061-abc1-6f044c10355b\") " pod="openstack/keystone-cron-29405641-ngxl6" Nov 28 14:01:00 crc kubenswrapper[4631]: I1128 14:01:00.420837 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4h4bh\" (UniqueName: \"kubernetes.io/projected/2080b2bc-31ce-4061-abc1-6f044c10355b-kube-api-access-4h4bh\") pod \"keystone-cron-29405641-ngxl6\" (UID: \"2080b2bc-31ce-4061-abc1-6f044c10355b\") " pod="openstack/keystone-cron-29405641-ngxl6" Nov 28 14:01:00 crc kubenswrapper[4631]: I1128 14:01:00.420882 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/2080b2bc-31ce-4061-abc1-6f044c10355b-fernet-keys\") pod \"keystone-cron-29405641-ngxl6\" (UID: \"2080b2bc-31ce-4061-abc1-6f044c10355b\") " pod="openstack/keystone-cron-29405641-ngxl6" Nov 28 14:01:00 crc kubenswrapper[4631]: I1128 14:01:00.420926 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2080b2bc-31ce-4061-abc1-6f044c10355b-combined-ca-bundle\") pod \"keystone-cron-29405641-ngxl6\" (UID: \"2080b2bc-31ce-4061-abc1-6f044c10355b\") " pod="openstack/keystone-cron-29405641-ngxl6" Nov 28 14:01:00 crc kubenswrapper[4631]: I1128 14:01:00.427788 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2080b2bc-31ce-4061-abc1-6f044c10355b-combined-ca-bundle\") pod \"keystone-cron-29405641-ngxl6\" (UID: \"2080b2bc-31ce-4061-abc1-6f044c10355b\") " pod="openstack/keystone-cron-29405641-ngxl6" Nov 28 14:01:00 crc kubenswrapper[4631]: I1128 14:01:00.427929 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/2080b2bc-31ce-4061-abc1-6f044c10355b-fernet-keys\") pod \"keystone-cron-29405641-ngxl6\" (UID: \"2080b2bc-31ce-4061-abc1-6f044c10355b\") " pod="openstack/keystone-cron-29405641-ngxl6" Nov 28 14:01:00 crc kubenswrapper[4631]: I1128 14:01:00.440438 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2080b2bc-31ce-4061-abc1-6f044c10355b-config-data\") pod \"keystone-cron-29405641-ngxl6\" (UID: \"2080b2bc-31ce-4061-abc1-6f044c10355b\") " pod="openstack/keystone-cron-29405641-ngxl6" Nov 28 14:01:00 crc kubenswrapper[4631]: I1128 14:01:00.443414 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4h4bh\" (UniqueName: \"kubernetes.io/projected/2080b2bc-31ce-4061-abc1-6f044c10355b-kube-api-access-4h4bh\") pod \"keystone-cron-29405641-ngxl6\" (UID: \"2080b2bc-31ce-4061-abc1-6f044c10355b\") " pod="openstack/keystone-cron-29405641-ngxl6" Nov 28 14:01:00 crc kubenswrapper[4631]: I1128 14:01:00.479422 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29405641-ngxl6" Nov 28 14:01:00 crc kubenswrapper[4631]: I1128 14:01:00.513935 4631 scope.go:117] "RemoveContainer" containerID="b4f03963475b3c249f705cef33ea3485fc5d4cec16ae5d211e2ba99ac0014be9" Nov 28 14:01:00 crc kubenswrapper[4631]: E1128 14:01:00.514268 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-47sxw_openshift-machine-config-operator(8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1)\"" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" podUID="8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1" Nov 28 14:01:00 crc kubenswrapper[4631]: I1128 14:01:00.950654 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29405641-ngxl6"] Nov 28 14:01:01 crc kubenswrapper[4631]: I1128 14:01:01.355409 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29405641-ngxl6" event={"ID":"2080b2bc-31ce-4061-abc1-6f044c10355b","Type":"ContainerStarted","Data":"82d0c3d5ae96ba5cd67b1e80cc3d7e4bcef2e78ac1ece64df3e755abbee5b956"} Nov 28 14:01:01 crc kubenswrapper[4631]: I1128 14:01:01.355460 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29405641-ngxl6" event={"ID":"2080b2bc-31ce-4061-abc1-6f044c10355b","Type":"ContainerStarted","Data":"c96b07d81c8985cb3c4caa696475150a5caa6b8eae824617d663558feeb74df7"} Nov 28 14:01:01 crc kubenswrapper[4631]: I1128 14:01:01.379497 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-cron-29405641-ngxl6" podStartSLOduration=1.379467709 podStartE2EDuration="1.379467709s" podCreationTimestamp="2025-11-28 14:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 14:01:01.3720605 +0000 UTC m=+2418.179363854" watchObservedRunningTime="2025-11-28 14:01:01.379467709 +0000 UTC m=+2418.186771053" Nov 28 14:01:02 crc kubenswrapper[4631]: I1128 14:01:02.365181 4631 generic.go:334] "Generic (PLEG): container finished" podID="86dddd7e-e065-4593-8d47-4d36a2cddb22" containerID="591422d62747b91572874434cfb086d5fe5ebba0af3980e91a9e5a321435801c" exitCode=0 Nov 28 14:01:02 crc kubenswrapper[4631]: I1128 14:01:02.365269 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-pw8wp" event={"ID":"86dddd7e-e065-4593-8d47-4d36a2cddb22","Type":"ContainerDied","Data":"591422d62747b91572874434cfb086d5fe5ebba0af3980e91a9e5a321435801c"} Nov 28 14:01:03 crc kubenswrapper[4631]: I1128 14:01:03.903333 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-pw8wp" Nov 28 14:01:04 crc kubenswrapper[4631]: I1128 14:01:04.020771 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/86dddd7e-e065-4593-8d47-4d36a2cddb22-inventory\") pod \"86dddd7e-e065-4593-8d47-4d36a2cddb22\" (UID: \"86dddd7e-e065-4593-8d47-4d36a2cddb22\") " Nov 28 14:01:04 crc kubenswrapper[4631]: I1128 14:01:04.021517 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/86dddd7e-e065-4593-8d47-4d36a2cddb22-ssh-key\") pod \"86dddd7e-e065-4593-8d47-4d36a2cddb22\" (UID: \"86dddd7e-e065-4593-8d47-4d36a2cddb22\") " Nov 28 14:01:04 crc kubenswrapper[4631]: I1128 14:01:04.021613 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lrgd6\" (UniqueName: \"kubernetes.io/projected/86dddd7e-e065-4593-8d47-4d36a2cddb22-kube-api-access-lrgd6\") pod \"86dddd7e-e065-4593-8d47-4d36a2cddb22\" (UID: \"86dddd7e-e065-4593-8d47-4d36a2cddb22\") " Nov 28 14:01:04 crc kubenswrapper[4631]: I1128 14:01:04.021794 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/86dddd7e-e065-4593-8d47-4d36a2cddb22-libvirt-combined-ca-bundle\") pod \"86dddd7e-e065-4593-8d47-4d36a2cddb22\" (UID: \"86dddd7e-e065-4593-8d47-4d36a2cddb22\") " Nov 28 14:01:04 crc kubenswrapper[4631]: I1128 14:01:04.021960 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/86dddd7e-e065-4593-8d47-4d36a2cddb22-libvirt-secret-0\") pod \"86dddd7e-e065-4593-8d47-4d36a2cddb22\" (UID: \"86dddd7e-e065-4593-8d47-4d36a2cddb22\") " Nov 28 14:01:04 crc kubenswrapper[4631]: I1128 14:01:04.037320 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/86dddd7e-e065-4593-8d47-4d36a2cddb22-kube-api-access-lrgd6" (OuterVolumeSpecName: "kube-api-access-lrgd6") pod "86dddd7e-e065-4593-8d47-4d36a2cddb22" (UID: "86dddd7e-e065-4593-8d47-4d36a2cddb22"). InnerVolumeSpecName "kube-api-access-lrgd6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:01:04 crc kubenswrapper[4631]: I1128 14:01:04.041440 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/86dddd7e-e065-4593-8d47-4d36a2cddb22-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "86dddd7e-e065-4593-8d47-4d36a2cddb22" (UID: "86dddd7e-e065-4593-8d47-4d36a2cddb22"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:01:04 crc kubenswrapper[4631]: I1128 14:01:04.090848 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/86dddd7e-e065-4593-8d47-4d36a2cddb22-inventory" (OuterVolumeSpecName: "inventory") pod "86dddd7e-e065-4593-8d47-4d36a2cddb22" (UID: "86dddd7e-e065-4593-8d47-4d36a2cddb22"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:01:04 crc kubenswrapper[4631]: I1128 14:01:04.092981 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/86dddd7e-e065-4593-8d47-4d36a2cddb22-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "86dddd7e-e065-4593-8d47-4d36a2cddb22" (UID: "86dddd7e-e065-4593-8d47-4d36a2cddb22"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:01:04 crc kubenswrapper[4631]: I1128 14:01:04.096476 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/86dddd7e-e065-4593-8d47-4d36a2cddb22-libvirt-secret-0" (OuterVolumeSpecName: "libvirt-secret-0") pod "86dddd7e-e065-4593-8d47-4d36a2cddb22" (UID: "86dddd7e-e065-4593-8d47-4d36a2cddb22"). InnerVolumeSpecName "libvirt-secret-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:01:04 crc kubenswrapper[4631]: I1128 14:01:04.124010 4631 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/86dddd7e-e065-4593-8d47-4d36a2cddb22-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 14:01:04 crc kubenswrapper[4631]: I1128 14:01:04.124049 4631 reconciler_common.go:293] "Volume detached for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/86dddd7e-e065-4593-8d47-4d36a2cddb22-libvirt-secret-0\") on node \"crc\" DevicePath \"\"" Nov 28 14:01:04 crc kubenswrapper[4631]: I1128 14:01:04.124062 4631 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/86dddd7e-e065-4593-8d47-4d36a2cddb22-inventory\") on node \"crc\" DevicePath \"\"" Nov 28 14:01:04 crc kubenswrapper[4631]: I1128 14:01:04.124071 4631 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/86dddd7e-e065-4593-8d47-4d36a2cddb22-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 28 14:01:04 crc kubenswrapper[4631]: I1128 14:01:04.124079 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lrgd6\" (UniqueName: \"kubernetes.io/projected/86dddd7e-e065-4593-8d47-4d36a2cddb22-kube-api-access-lrgd6\") on node \"crc\" DevicePath \"\"" Nov 28 14:01:04 crc kubenswrapper[4631]: I1128 14:01:04.401928 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-pw8wp" event={"ID":"86dddd7e-e065-4593-8d47-4d36a2cddb22","Type":"ContainerDied","Data":"5ef1ebedaece586346c179e0edf6486a456933646eec9e4130be8acba5475d01"} Nov 28 14:01:04 crc kubenswrapper[4631]: I1128 14:01:04.401981 4631 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5ef1ebedaece586346c179e0edf6486a456933646eec9e4130be8acba5475d01" Nov 28 14:01:04 crc kubenswrapper[4631]: I1128 14:01:04.401994 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-pw8wp" Nov 28 14:01:04 crc kubenswrapper[4631]: I1128 14:01:04.495493 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-5hrhp"] Nov 28 14:01:04 crc kubenswrapper[4631]: E1128 14:01:04.495925 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="86dddd7e-e065-4593-8d47-4d36a2cddb22" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Nov 28 14:01:04 crc kubenswrapper[4631]: I1128 14:01:04.495950 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="86dddd7e-e065-4593-8d47-4d36a2cddb22" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Nov 28 14:01:04 crc kubenswrapper[4631]: I1128 14:01:04.496145 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="86dddd7e-e065-4593-8d47-4d36a2cddb22" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Nov 28 14:01:04 crc kubenswrapper[4631]: I1128 14:01:04.496845 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5hrhp" Nov 28 14:01:04 crc kubenswrapper[4631]: I1128 14:01:04.504324 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 28 14:01:04 crc kubenswrapper[4631]: I1128 14:01:04.507522 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Nov 28 14:01:04 crc kubenswrapper[4631]: I1128 14:01:04.507637 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-vp4ns" Nov 28 14:01:04 crc kubenswrapper[4631]: I1128 14:01:04.507954 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-extra-config" Nov 28 14:01:04 crc kubenswrapper[4631]: I1128 14:01:04.509205 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Nov 28 14:01:04 crc kubenswrapper[4631]: I1128 14:01:04.510264 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 28 14:01:04 crc kubenswrapper[4631]: I1128 14:01:04.512574 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 28 14:01:04 crc kubenswrapper[4631]: I1128 14:01:04.534513 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-5hrhp"] Nov 28 14:01:04 crc kubenswrapper[4631]: I1128 14:01:04.634552 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7692b0ac-e13d-4413-a2c0-e72062ec9110-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-5hrhp\" (UID: \"7692b0ac-e13d-4413-a2c0-e72062ec9110\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5hrhp" Nov 28 14:01:04 crc kubenswrapper[4631]: I1128 14:01:04.634693 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7692b0ac-e13d-4413-a2c0-e72062ec9110-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-5hrhp\" (UID: \"7692b0ac-e13d-4413-a2c0-e72062ec9110\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5hrhp" Nov 28 14:01:04 crc kubenswrapper[4631]: I1128 14:01:04.634721 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/7692b0ac-e13d-4413-a2c0-e72062ec9110-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-5hrhp\" (UID: \"7692b0ac-e13d-4413-a2c0-e72062ec9110\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5hrhp" Nov 28 14:01:04 crc kubenswrapper[4631]: I1128 14:01:04.634760 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/7692b0ac-e13d-4413-a2c0-e72062ec9110-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-5hrhp\" (UID: \"7692b0ac-e13d-4413-a2c0-e72062ec9110\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5hrhp" Nov 28 14:01:04 crc kubenswrapper[4631]: I1128 14:01:04.634782 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/7692b0ac-e13d-4413-a2c0-e72062ec9110-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-5hrhp\" (UID: \"7692b0ac-e13d-4413-a2c0-e72062ec9110\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5hrhp" Nov 28 14:01:04 crc kubenswrapper[4631]: I1128 14:01:04.634822 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/7692b0ac-e13d-4413-a2c0-e72062ec9110-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-5hrhp\" (UID: \"7692b0ac-e13d-4413-a2c0-e72062ec9110\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5hrhp" Nov 28 14:01:04 crc kubenswrapper[4631]: I1128 14:01:04.634870 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/7692b0ac-e13d-4413-a2c0-e72062ec9110-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-5hrhp\" (UID: \"7692b0ac-e13d-4413-a2c0-e72062ec9110\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5hrhp" Nov 28 14:01:04 crc kubenswrapper[4631]: I1128 14:01:04.635004 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7692b0ac-e13d-4413-a2c0-e72062ec9110-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-5hrhp\" (UID: \"7692b0ac-e13d-4413-a2c0-e72062ec9110\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5hrhp" Nov 28 14:01:04 crc kubenswrapper[4631]: I1128 14:01:04.635164 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vxpz7\" (UniqueName: \"kubernetes.io/projected/7692b0ac-e13d-4413-a2c0-e72062ec9110-kube-api-access-vxpz7\") pod \"nova-edpm-deployment-openstack-edpm-ipam-5hrhp\" (UID: \"7692b0ac-e13d-4413-a2c0-e72062ec9110\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5hrhp" Nov 28 14:01:04 crc kubenswrapper[4631]: I1128 14:01:04.737173 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vxpz7\" (UniqueName: \"kubernetes.io/projected/7692b0ac-e13d-4413-a2c0-e72062ec9110-kube-api-access-vxpz7\") pod \"nova-edpm-deployment-openstack-edpm-ipam-5hrhp\" (UID: \"7692b0ac-e13d-4413-a2c0-e72062ec9110\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5hrhp" Nov 28 14:01:04 crc kubenswrapper[4631]: I1128 14:01:04.737336 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7692b0ac-e13d-4413-a2c0-e72062ec9110-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-5hrhp\" (UID: \"7692b0ac-e13d-4413-a2c0-e72062ec9110\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5hrhp" Nov 28 14:01:04 crc kubenswrapper[4631]: I1128 14:01:04.737402 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7692b0ac-e13d-4413-a2c0-e72062ec9110-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-5hrhp\" (UID: \"7692b0ac-e13d-4413-a2c0-e72062ec9110\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5hrhp" Nov 28 14:01:04 crc kubenswrapper[4631]: I1128 14:01:04.737431 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/7692b0ac-e13d-4413-a2c0-e72062ec9110-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-5hrhp\" (UID: \"7692b0ac-e13d-4413-a2c0-e72062ec9110\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5hrhp" Nov 28 14:01:04 crc kubenswrapper[4631]: I1128 14:01:04.737464 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/7692b0ac-e13d-4413-a2c0-e72062ec9110-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-5hrhp\" (UID: \"7692b0ac-e13d-4413-a2c0-e72062ec9110\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5hrhp" Nov 28 14:01:04 crc kubenswrapper[4631]: I1128 14:01:04.737495 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/7692b0ac-e13d-4413-a2c0-e72062ec9110-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-5hrhp\" (UID: \"7692b0ac-e13d-4413-a2c0-e72062ec9110\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5hrhp" Nov 28 14:01:04 crc kubenswrapper[4631]: I1128 14:01:04.737536 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/7692b0ac-e13d-4413-a2c0-e72062ec9110-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-5hrhp\" (UID: \"7692b0ac-e13d-4413-a2c0-e72062ec9110\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5hrhp" Nov 28 14:01:04 crc kubenswrapper[4631]: I1128 14:01:04.737588 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/7692b0ac-e13d-4413-a2c0-e72062ec9110-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-5hrhp\" (UID: \"7692b0ac-e13d-4413-a2c0-e72062ec9110\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5hrhp" Nov 28 14:01:04 crc kubenswrapper[4631]: I1128 14:01:04.737616 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7692b0ac-e13d-4413-a2c0-e72062ec9110-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-5hrhp\" (UID: \"7692b0ac-e13d-4413-a2c0-e72062ec9110\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5hrhp" Nov 28 14:01:04 crc kubenswrapper[4631]: I1128 14:01:04.738616 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/7692b0ac-e13d-4413-a2c0-e72062ec9110-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-5hrhp\" (UID: \"7692b0ac-e13d-4413-a2c0-e72062ec9110\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5hrhp" Nov 28 14:01:04 crc kubenswrapper[4631]: I1128 14:01:04.742349 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7692b0ac-e13d-4413-a2c0-e72062ec9110-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-5hrhp\" (UID: \"7692b0ac-e13d-4413-a2c0-e72062ec9110\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5hrhp" Nov 28 14:01:04 crc kubenswrapper[4631]: I1128 14:01:04.743172 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/7692b0ac-e13d-4413-a2c0-e72062ec9110-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-5hrhp\" (UID: \"7692b0ac-e13d-4413-a2c0-e72062ec9110\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5hrhp" Nov 28 14:01:04 crc kubenswrapper[4631]: I1128 14:01:04.743280 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/7692b0ac-e13d-4413-a2c0-e72062ec9110-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-5hrhp\" (UID: \"7692b0ac-e13d-4413-a2c0-e72062ec9110\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5hrhp" Nov 28 14:01:04 crc kubenswrapper[4631]: I1128 14:01:04.744223 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/7692b0ac-e13d-4413-a2c0-e72062ec9110-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-5hrhp\" (UID: \"7692b0ac-e13d-4413-a2c0-e72062ec9110\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5hrhp" Nov 28 14:01:04 crc kubenswrapper[4631]: I1128 14:01:04.744374 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7692b0ac-e13d-4413-a2c0-e72062ec9110-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-5hrhp\" (UID: \"7692b0ac-e13d-4413-a2c0-e72062ec9110\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5hrhp" Nov 28 14:01:04 crc kubenswrapper[4631]: I1128 14:01:04.744918 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7692b0ac-e13d-4413-a2c0-e72062ec9110-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-5hrhp\" (UID: \"7692b0ac-e13d-4413-a2c0-e72062ec9110\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5hrhp" Nov 28 14:01:04 crc kubenswrapper[4631]: I1128 14:01:04.746246 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/7692b0ac-e13d-4413-a2c0-e72062ec9110-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-5hrhp\" (UID: \"7692b0ac-e13d-4413-a2c0-e72062ec9110\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5hrhp" Nov 28 14:01:04 crc kubenswrapper[4631]: I1128 14:01:04.758272 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vxpz7\" (UniqueName: \"kubernetes.io/projected/7692b0ac-e13d-4413-a2c0-e72062ec9110-kube-api-access-vxpz7\") pod \"nova-edpm-deployment-openstack-edpm-ipam-5hrhp\" (UID: \"7692b0ac-e13d-4413-a2c0-e72062ec9110\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5hrhp" Nov 28 14:01:04 crc kubenswrapper[4631]: I1128 14:01:04.816061 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5hrhp" Nov 28 14:01:05 crc kubenswrapper[4631]: I1128 14:01:05.391600 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-5hrhp"] Nov 28 14:01:05 crc kubenswrapper[4631]: I1128 14:01:05.393040 4631 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 28 14:01:05 crc kubenswrapper[4631]: I1128 14:01:05.416648 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5hrhp" event={"ID":"7692b0ac-e13d-4413-a2c0-e72062ec9110","Type":"ContainerStarted","Data":"7137822baa30b0d4ef24bdcdd0e32db133589c47203aa7c47c3fc844fa74a20d"} Nov 28 14:01:05 crc kubenswrapper[4631]: I1128 14:01:05.418431 4631 generic.go:334] "Generic (PLEG): container finished" podID="2080b2bc-31ce-4061-abc1-6f044c10355b" containerID="82d0c3d5ae96ba5cd67b1e80cc3d7e4bcef2e78ac1ece64df3e755abbee5b956" exitCode=0 Nov 28 14:01:05 crc kubenswrapper[4631]: I1128 14:01:05.418482 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29405641-ngxl6" event={"ID":"2080b2bc-31ce-4061-abc1-6f044c10355b","Type":"ContainerDied","Data":"82d0c3d5ae96ba5cd67b1e80cc3d7e4bcef2e78ac1ece64df3e755abbee5b956"} Nov 28 14:01:06 crc kubenswrapper[4631]: I1128 14:01:06.856473 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29405641-ngxl6" Nov 28 14:01:06 crc kubenswrapper[4631]: I1128 14:01:06.990612 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2080b2bc-31ce-4061-abc1-6f044c10355b-combined-ca-bundle\") pod \"2080b2bc-31ce-4061-abc1-6f044c10355b\" (UID: \"2080b2bc-31ce-4061-abc1-6f044c10355b\") " Nov 28 14:01:06 crc kubenswrapper[4631]: I1128 14:01:06.990696 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4h4bh\" (UniqueName: \"kubernetes.io/projected/2080b2bc-31ce-4061-abc1-6f044c10355b-kube-api-access-4h4bh\") pod \"2080b2bc-31ce-4061-abc1-6f044c10355b\" (UID: \"2080b2bc-31ce-4061-abc1-6f044c10355b\") " Nov 28 14:01:06 crc kubenswrapper[4631]: I1128 14:01:06.990765 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/2080b2bc-31ce-4061-abc1-6f044c10355b-fernet-keys\") pod \"2080b2bc-31ce-4061-abc1-6f044c10355b\" (UID: \"2080b2bc-31ce-4061-abc1-6f044c10355b\") " Nov 28 14:01:06 crc kubenswrapper[4631]: I1128 14:01:06.990870 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2080b2bc-31ce-4061-abc1-6f044c10355b-config-data\") pod \"2080b2bc-31ce-4061-abc1-6f044c10355b\" (UID: \"2080b2bc-31ce-4061-abc1-6f044c10355b\") " Nov 28 14:01:06 crc kubenswrapper[4631]: I1128 14:01:06.996779 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2080b2bc-31ce-4061-abc1-6f044c10355b-kube-api-access-4h4bh" (OuterVolumeSpecName: "kube-api-access-4h4bh") pod "2080b2bc-31ce-4061-abc1-6f044c10355b" (UID: "2080b2bc-31ce-4061-abc1-6f044c10355b"). InnerVolumeSpecName "kube-api-access-4h4bh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:01:06 crc kubenswrapper[4631]: I1128 14:01:06.997578 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2080b2bc-31ce-4061-abc1-6f044c10355b-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "2080b2bc-31ce-4061-abc1-6f044c10355b" (UID: "2080b2bc-31ce-4061-abc1-6f044c10355b"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:01:07 crc kubenswrapper[4631]: I1128 14:01:07.021102 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2080b2bc-31ce-4061-abc1-6f044c10355b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2080b2bc-31ce-4061-abc1-6f044c10355b" (UID: "2080b2bc-31ce-4061-abc1-6f044c10355b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:01:07 crc kubenswrapper[4631]: I1128 14:01:07.044066 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2080b2bc-31ce-4061-abc1-6f044c10355b-config-data" (OuterVolumeSpecName: "config-data") pod "2080b2bc-31ce-4061-abc1-6f044c10355b" (UID: "2080b2bc-31ce-4061-abc1-6f044c10355b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:01:07 crc kubenswrapper[4631]: I1128 14:01:07.094926 4631 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2080b2bc-31ce-4061-abc1-6f044c10355b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 14:01:07 crc kubenswrapper[4631]: I1128 14:01:07.094972 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4h4bh\" (UniqueName: \"kubernetes.io/projected/2080b2bc-31ce-4061-abc1-6f044c10355b-kube-api-access-4h4bh\") on node \"crc\" DevicePath \"\"" Nov 28 14:01:07 crc kubenswrapper[4631]: I1128 14:01:07.094987 4631 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/2080b2bc-31ce-4061-abc1-6f044c10355b-fernet-keys\") on node \"crc\" DevicePath \"\"" Nov 28 14:01:07 crc kubenswrapper[4631]: I1128 14:01:07.094997 4631 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2080b2bc-31ce-4061-abc1-6f044c10355b-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 14:01:07 crc kubenswrapper[4631]: I1128 14:01:07.439441 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5hrhp" event={"ID":"7692b0ac-e13d-4413-a2c0-e72062ec9110","Type":"ContainerStarted","Data":"3c56d645492c97109a5a4b6e10ae5a473ebd68e47fbf26cdb7e888eebd1a0cf4"} Nov 28 14:01:07 crc kubenswrapper[4631]: I1128 14:01:07.442626 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29405641-ngxl6" event={"ID":"2080b2bc-31ce-4061-abc1-6f044c10355b","Type":"ContainerDied","Data":"c96b07d81c8985cb3c4caa696475150a5caa6b8eae824617d663558feeb74df7"} Nov 28 14:01:07 crc kubenswrapper[4631]: I1128 14:01:07.442673 4631 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c96b07d81c8985cb3c4caa696475150a5caa6b8eae824617d663558feeb74df7" Nov 28 14:01:07 crc kubenswrapper[4631]: I1128 14:01:07.442742 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29405641-ngxl6" Nov 28 14:01:07 crc kubenswrapper[4631]: I1128 14:01:07.487593 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5hrhp" podStartSLOduration=2.335742887 podStartE2EDuration="3.487530726s" podCreationTimestamp="2025-11-28 14:01:04 +0000 UTC" firstStartedPulling="2025-11-28 14:01:05.392737749 +0000 UTC m=+2422.200041093" lastFinishedPulling="2025-11-28 14:01:06.544525588 +0000 UTC m=+2423.351828932" observedRunningTime="2025-11-28 14:01:07.474080121 +0000 UTC m=+2424.281383485" watchObservedRunningTime="2025-11-28 14:01:07.487530726 +0000 UTC m=+2424.294834070" Nov 28 14:01:12 crc kubenswrapper[4631]: I1128 14:01:12.513197 4631 scope.go:117] "RemoveContainer" containerID="b4f03963475b3c249f705cef33ea3485fc5d4cec16ae5d211e2ba99ac0014be9" Nov 28 14:01:12 crc kubenswrapper[4631]: E1128 14:01:12.515387 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-47sxw_openshift-machine-config-operator(8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1)\"" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" podUID="8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1" Nov 28 14:01:25 crc kubenswrapper[4631]: I1128 14:01:25.518248 4631 scope.go:117] "RemoveContainer" containerID="b4f03963475b3c249f705cef33ea3485fc5d4cec16ae5d211e2ba99ac0014be9" Nov 28 14:01:25 crc kubenswrapper[4631]: E1128 14:01:25.519025 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-47sxw_openshift-machine-config-operator(8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1)\"" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" podUID="8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1" Nov 28 14:01:36 crc kubenswrapper[4631]: I1128 14:01:36.513398 4631 scope.go:117] "RemoveContainer" containerID="b4f03963475b3c249f705cef33ea3485fc5d4cec16ae5d211e2ba99ac0014be9" Nov 28 14:01:36 crc kubenswrapper[4631]: E1128 14:01:36.514324 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-47sxw_openshift-machine-config-operator(8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1)\"" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" podUID="8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1" Nov 28 14:01:50 crc kubenswrapper[4631]: I1128 14:01:50.513715 4631 scope.go:117] "RemoveContainer" containerID="b4f03963475b3c249f705cef33ea3485fc5d4cec16ae5d211e2ba99ac0014be9" Nov 28 14:01:50 crc kubenswrapper[4631]: E1128 14:01:50.514632 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-47sxw_openshift-machine-config-operator(8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1)\"" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" podUID="8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1" Nov 28 14:02:02 crc kubenswrapper[4631]: I1128 14:02:02.513076 4631 scope.go:117] "RemoveContainer" containerID="b4f03963475b3c249f705cef33ea3485fc5d4cec16ae5d211e2ba99ac0014be9" Nov 28 14:02:02 crc kubenswrapper[4631]: E1128 14:02:02.514319 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-47sxw_openshift-machine-config-operator(8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1)\"" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" podUID="8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1" Nov 28 14:02:17 crc kubenswrapper[4631]: I1128 14:02:17.512786 4631 scope.go:117] "RemoveContainer" containerID="b4f03963475b3c249f705cef33ea3485fc5d4cec16ae5d211e2ba99ac0014be9" Nov 28 14:02:17 crc kubenswrapper[4631]: E1128 14:02:17.513636 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-47sxw_openshift-machine-config-operator(8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1)\"" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" podUID="8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1" Nov 28 14:02:29 crc kubenswrapper[4631]: I1128 14:02:29.513211 4631 scope.go:117] "RemoveContainer" containerID="b4f03963475b3c249f705cef33ea3485fc5d4cec16ae5d211e2ba99ac0014be9" Nov 28 14:02:29 crc kubenswrapper[4631]: E1128 14:02:29.515616 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-47sxw_openshift-machine-config-operator(8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1)\"" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" podUID="8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1" Nov 28 14:02:42 crc kubenswrapper[4631]: I1128 14:02:42.512903 4631 scope.go:117] "RemoveContainer" containerID="b4f03963475b3c249f705cef33ea3485fc5d4cec16ae5d211e2ba99ac0014be9" Nov 28 14:02:42 crc kubenswrapper[4631]: E1128 14:02:42.513677 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-47sxw_openshift-machine-config-operator(8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1)\"" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" podUID="8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1" Nov 28 14:02:56 crc kubenswrapper[4631]: I1128 14:02:56.513625 4631 scope.go:117] "RemoveContainer" containerID="b4f03963475b3c249f705cef33ea3485fc5d4cec16ae5d211e2ba99ac0014be9" Nov 28 14:02:56 crc kubenswrapper[4631]: E1128 14:02:56.514424 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-47sxw_openshift-machine-config-operator(8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1)\"" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" podUID="8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1" Nov 28 14:03:07 crc kubenswrapper[4631]: I1128 14:03:07.514605 4631 scope.go:117] "RemoveContainer" containerID="b4f03963475b3c249f705cef33ea3485fc5d4cec16ae5d211e2ba99ac0014be9" Nov 28 14:03:07 crc kubenswrapper[4631]: E1128 14:03:07.515405 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-47sxw_openshift-machine-config-operator(8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1)\"" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" podUID="8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1" Nov 28 14:03:19 crc kubenswrapper[4631]: I1128 14:03:19.515250 4631 scope.go:117] "RemoveContainer" containerID="b4f03963475b3c249f705cef33ea3485fc5d4cec16ae5d211e2ba99ac0014be9" Nov 28 14:03:19 crc kubenswrapper[4631]: E1128 14:03:19.516503 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-47sxw_openshift-machine-config-operator(8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1)\"" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" podUID="8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1" Nov 28 14:03:32 crc kubenswrapper[4631]: I1128 14:03:32.514045 4631 scope.go:117] "RemoveContainer" containerID="b4f03963475b3c249f705cef33ea3485fc5d4cec16ae5d211e2ba99ac0014be9" Nov 28 14:03:32 crc kubenswrapper[4631]: E1128 14:03:32.514909 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-47sxw_openshift-machine-config-operator(8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1)\"" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" podUID="8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1" Nov 28 14:03:45 crc kubenswrapper[4631]: I1128 14:03:45.514253 4631 scope.go:117] "RemoveContainer" containerID="b4f03963475b3c249f705cef33ea3485fc5d4cec16ae5d211e2ba99ac0014be9" Nov 28 14:03:45 crc kubenswrapper[4631]: E1128 14:03:45.515312 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-47sxw_openshift-machine-config-operator(8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1)\"" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" podUID="8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1" Nov 28 14:04:00 crc kubenswrapper[4631]: I1128 14:04:00.513508 4631 scope.go:117] "RemoveContainer" containerID="b4f03963475b3c249f705cef33ea3485fc5d4cec16ae5d211e2ba99ac0014be9" Nov 28 14:04:00 crc kubenswrapper[4631]: E1128 14:04:00.514458 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-47sxw_openshift-machine-config-operator(8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1)\"" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" podUID="8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1" Nov 28 14:04:06 crc kubenswrapper[4631]: I1128 14:04:06.089842 4631 generic.go:334] "Generic (PLEG): container finished" podID="7692b0ac-e13d-4413-a2c0-e72062ec9110" containerID="3c56d645492c97109a5a4b6e10ae5a473ebd68e47fbf26cdb7e888eebd1a0cf4" exitCode=0 Nov 28 14:04:06 crc kubenswrapper[4631]: I1128 14:04:06.089917 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5hrhp" event={"ID":"7692b0ac-e13d-4413-a2c0-e72062ec9110","Type":"ContainerDied","Data":"3c56d645492c97109a5a4b6e10ae5a473ebd68e47fbf26cdb7e888eebd1a0cf4"} Nov 28 14:04:07 crc kubenswrapper[4631]: I1128 14:04:07.527439 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5hrhp" Nov 28 14:04:07 crc kubenswrapper[4631]: I1128 14:04:07.653839 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/7692b0ac-e13d-4413-a2c0-e72062ec9110-nova-cell1-compute-config-1\") pod \"7692b0ac-e13d-4413-a2c0-e72062ec9110\" (UID: \"7692b0ac-e13d-4413-a2c0-e72062ec9110\") " Nov 28 14:04:07 crc kubenswrapper[4631]: I1128 14:04:07.653944 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7692b0ac-e13d-4413-a2c0-e72062ec9110-nova-combined-ca-bundle\") pod \"7692b0ac-e13d-4413-a2c0-e72062ec9110\" (UID: \"7692b0ac-e13d-4413-a2c0-e72062ec9110\") " Nov 28 14:04:07 crc kubenswrapper[4631]: I1128 14:04:07.653961 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/7692b0ac-e13d-4413-a2c0-e72062ec9110-nova-extra-config-0\") pod \"7692b0ac-e13d-4413-a2c0-e72062ec9110\" (UID: \"7692b0ac-e13d-4413-a2c0-e72062ec9110\") " Nov 28 14:04:07 crc kubenswrapper[4631]: I1128 14:04:07.654051 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/7692b0ac-e13d-4413-a2c0-e72062ec9110-nova-migration-ssh-key-1\") pod \"7692b0ac-e13d-4413-a2c0-e72062ec9110\" (UID: \"7692b0ac-e13d-4413-a2c0-e72062ec9110\") " Nov 28 14:04:07 crc kubenswrapper[4631]: I1128 14:04:07.654091 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7692b0ac-e13d-4413-a2c0-e72062ec9110-inventory\") pod \"7692b0ac-e13d-4413-a2c0-e72062ec9110\" (UID: \"7692b0ac-e13d-4413-a2c0-e72062ec9110\") " Nov 28 14:04:07 crc kubenswrapper[4631]: I1128 14:04:07.654122 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vxpz7\" (UniqueName: \"kubernetes.io/projected/7692b0ac-e13d-4413-a2c0-e72062ec9110-kube-api-access-vxpz7\") pod \"7692b0ac-e13d-4413-a2c0-e72062ec9110\" (UID: \"7692b0ac-e13d-4413-a2c0-e72062ec9110\") " Nov 28 14:04:07 crc kubenswrapper[4631]: I1128 14:04:07.654144 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7692b0ac-e13d-4413-a2c0-e72062ec9110-ssh-key\") pod \"7692b0ac-e13d-4413-a2c0-e72062ec9110\" (UID: \"7692b0ac-e13d-4413-a2c0-e72062ec9110\") " Nov 28 14:04:07 crc kubenswrapper[4631]: I1128 14:04:07.654201 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/7692b0ac-e13d-4413-a2c0-e72062ec9110-nova-migration-ssh-key-0\") pod \"7692b0ac-e13d-4413-a2c0-e72062ec9110\" (UID: \"7692b0ac-e13d-4413-a2c0-e72062ec9110\") " Nov 28 14:04:07 crc kubenswrapper[4631]: I1128 14:04:07.654220 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/7692b0ac-e13d-4413-a2c0-e72062ec9110-nova-cell1-compute-config-0\") pod \"7692b0ac-e13d-4413-a2c0-e72062ec9110\" (UID: \"7692b0ac-e13d-4413-a2c0-e72062ec9110\") " Nov 28 14:04:07 crc kubenswrapper[4631]: I1128 14:04:07.665757 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7692b0ac-e13d-4413-a2c0-e72062ec9110-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "7692b0ac-e13d-4413-a2c0-e72062ec9110" (UID: "7692b0ac-e13d-4413-a2c0-e72062ec9110"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:04:07 crc kubenswrapper[4631]: I1128 14:04:07.672644 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7692b0ac-e13d-4413-a2c0-e72062ec9110-kube-api-access-vxpz7" (OuterVolumeSpecName: "kube-api-access-vxpz7") pod "7692b0ac-e13d-4413-a2c0-e72062ec9110" (UID: "7692b0ac-e13d-4413-a2c0-e72062ec9110"). InnerVolumeSpecName "kube-api-access-vxpz7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:04:07 crc kubenswrapper[4631]: I1128 14:04:07.684968 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7692b0ac-e13d-4413-a2c0-e72062ec9110-nova-extra-config-0" (OuterVolumeSpecName: "nova-extra-config-0") pod "7692b0ac-e13d-4413-a2c0-e72062ec9110" (UID: "7692b0ac-e13d-4413-a2c0-e72062ec9110"). InnerVolumeSpecName "nova-extra-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:04:07 crc kubenswrapper[4631]: I1128 14:04:07.687604 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7692b0ac-e13d-4413-a2c0-e72062ec9110-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "7692b0ac-e13d-4413-a2c0-e72062ec9110" (UID: "7692b0ac-e13d-4413-a2c0-e72062ec9110"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:04:07 crc kubenswrapper[4631]: I1128 14:04:07.689220 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7692b0ac-e13d-4413-a2c0-e72062ec9110-inventory" (OuterVolumeSpecName: "inventory") pod "7692b0ac-e13d-4413-a2c0-e72062ec9110" (UID: "7692b0ac-e13d-4413-a2c0-e72062ec9110"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:04:07 crc kubenswrapper[4631]: I1128 14:04:07.695366 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7692b0ac-e13d-4413-a2c0-e72062ec9110-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "7692b0ac-e13d-4413-a2c0-e72062ec9110" (UID: "7692b0ac-e13d-4413-a2c0-e72062ec9110"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:04:07 crc kubenswrapper[4631]: I1128 14:04:07.717343 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7692b0ac-e13d-4413-a2c0-e72062ec9110-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "7692b0ac-e13d-4413-a2c0-e72062ec9110" (UID: "7692b0ac-e13d-4413-a2c0-e72062ec9110"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:04:07 crc kubenswrapper[4631]: I1128 14:04:07.723994 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7692b0ac-e13d-4413-a2c0-e72062ec9110-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "7692b0ac-e13d-4413-a2c0-e72062ec9110" (UID: "7692b0ac-e13d-4413-a2c0-e72062ec9110"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:04:07 crc kubenswrapper[4631]: I1128 14:04:07.726058 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7692b0ac-e13d-4413-a2c0-e72062ec9110-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "7692b0ac-e13d-4413-a2c0-e72062ec9110" (UID: "7692b0ac-e13d-4413-a2c0-e72062ec9110"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:04:07 crc kubenswrapper[4631]: I1128 14:04:07.757805 4631 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/7692b0ac-e13d-4413-a2c0-e72062ec9110-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Nov 28 14:04:07 crc kubenswrapper[4631]: I1128 14:04:07.757841 4631 reconciler_common.go:293] "Volume detached for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/7692b0ac-e13d-4413-a2c0-e72062ec9110-nova-extra-config-0\") on node \"crc\" DevicePath \"\"" Nov 28 14:04:07 crc kubenswrapper[4631]: I1128 14:04:07.757855 4631 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7692b0ac-e13d-4413-a2c0-e72062ec9110-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 14:04:07 crc kubenswrapper[4631]: I1128 14:04:07.757866 4631 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/7692b0ac-e13d-4413-a2c0-e72062ec9110-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Nov 28 14:04:07 crc kubenswrapper[4631]: I1128 14:04:07.757882 4631 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7692b0ac-e13d-4413-a2c0-e72062ec9110-inventory\") on node \"crc\" DevicePath \"\"" Nov 28 14:04:07 crc kubenswrapper[4631]: I1128 14:04:07.757894 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vxpz7\" (UniqueName: \"kubernetes.io/projected/7692b0ac-e13d-4413-a2c0-e72062ec9110-kube-api-access-vxpz7\") on node \"crc\" DevicePath \"\"" Nov 28 14:04:07 crc kubenswrapper[4631]: I1128 14:04:07.757906 4631 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7692b0ac-e13d-4413-a2c0-e72062ec9110-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 28 14:04:07 crc kubenswrapper[4631]: I1128 14:04:07.757918 4631 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/7692b0ac-e13d-4413-a2c0-e72062ec9110-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Nov 28 14:04:07 crc kubenswrapper[4631]: I1128 14:04:07.757929 4631 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/7692b0ac-e13d-4413-a2c0-e72062ec9110-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Nov 28 14:04:08 crc kubenswrapper[4631]: I1128 14:04:08.112367 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5hrhp" Nov 28 14:04:08 crc kubenswrapper[4631]: I1128 14:04:08.112414 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5hrhp" event={"ID":"7692b0ac-e13d-4413-a2c0-e72062ec9110","Type":"ContainerDied","Data":"7137822baa30b0d4ef24bdcdd0e32db133589c47203aa7c47c3fc844fa74a20d"} Nov 28 14:04:08 crc kubenswrapper[4631]: I1128 14:04:08.112470 4631 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7137822baa30b0d4ef24bdcdd0e32db133589c47203aa7c47c3fc844fa74a20d" Nov 28 14:04:08 crc kubenswrapper[4631]: I1128 14:04:08.264015 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-x9bvm"] Nov 28 14:04:08 crc kubenswrapper[4631]: E1128 14:04:08.264615 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7692b0ac-e13d-4413-a2c0-e72062ec9110" containerName="nova-edpm-deployment-openstack-edpm-ipam" Nov 28 14:04:08 crc kubenswrapper[4631]: I1128 14:04:08.264638 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="7692b0ac-e13d-4413-a2c0-e72062ec9110" containerName="nova-edpm-deployment-openstack-edpm-ipam" Nov 28 14:04:08 crc kubenswrapper[4631]: E1128 14:04:08.264660 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2080b2bc-31ce-4061-abc1-6f044c10355b" containerName="keystone-cron" Nov 28 14:04:08 crc kubenswrapper[4631]: I1128 14:04:08.264667 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="2080b2bc-31ce-4061-abc1-6f044c10355b" containerName="keystone-cron" Nov 28 14:04:08 crc kubenswrapper[4631]: I1128 14:04:08.264884 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="2080b2bc-31ce-4061-abc1-6f044c10355b" containerName="keystone-cron" Nov 28 14:04:08 crc kubenswrapper[4631]: I1128 14:04:08.264908 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="7692b0ac-e13d-4413-a2c0-e72062ec9110" containerName="nova-edpm-deployment-openstack-edpm-ipam" Nov 28 14:04:08 crc kubenswrapper[4631]: I1128 14:04:08.265737 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-x9bvm" Nov 28 14:04:08 crc kubenswrapper[4631]: I1128 14:04:08.273658 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-compute-config-data" Nov 28 14:04:08 crc kubenswrapper[4631]: I1128 14:04:08.274062 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 28 14:04:08 crc kubenswrapper[4631]: I1128 14:04:08.274431 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 28 14:04:08 crc kubenswrapper[4631]: I1128 14:04:08.276741 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-x9bvm"] Nov 28 14:04:08 crc kubenswrapper[4631]: I1128 14:04:08.277442 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 28 14:04:08 crc kubenswrapper[4631]: I1128 14:04:08.277464 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-vp4ns" Nov 28 14:04:08 crc kubenswrapper[4631]: I1128 14:04:08.367906 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/dab7280d-4cb8-4a43-b056-82296e560317-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-x9bvm\" (UID: \"dab7280d-4cb8-4a43-b056-82296e560317\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-x9bvm" Nov 28 14:04:08 crc kubenswrapper[4631]: I1128 14:04:08.368205 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/dab7280d-4cb8-4a43-b056-82296e560317-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-x9bvm\" (UID: \"dab7280d-4cb8-4a43-b056-82296e560317\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-x9bvm" Nov 28 14:04:08 crc kubenswrapper[4631]: I1128 14:04:08.368275 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/dab7280d-4cb8-4a43-b056-82296e560317-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-x9bvm\" (UID: \"dab7280d-4cb8-4a43-b056-82296e560317\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-x9bvm" Nov 28 14:04:08 crc kubenswrapper[4631]: I1128 14:04:08.368509 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/dab7280d-4cb8-4a43-b056-82296e560317-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-x9bvm\" (UID: \"dab7280d-4cb8-4a43-b056-82296e560317\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-x9bvm" Nov 28 14:04:08 crc kubenswrapper[4631]: I1128 14:04:08.368602 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/dab7280d-4cb8-4a43-b056-82296e560317-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-x9bvm\" (UID: \"dab7280d-4cb8-4a43-b056-82296e560317\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-x9bvm" Nov 28 14:04:08 crc kubenswrapper[4631]: I1128 14:04:08.368782 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dab7280d-4cb8-4a43-b056-82296e560317-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-x9bvm\" (UID: \"dab7280d-4cb8-4a43-b056-82296e560317\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-x9bvm" Nov 28 14:04:08 crc kubenswrapper[4631]: I1128 14:04:08.368979 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2997r\" (UniqueName: \"kubernetes.io/projected/dab7280d-4cb8-4a43-b056-82296e560317-kube-api-access-2997r\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-x9bvm\" (UID: \"dab7280d-4cb8-4a43-b056-82296e560317\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-x9bvm" Nov 28 14:04:08 crc kubenswrapper[4631]: I1128 14:04:08.470873 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/dab7280d-4cb8-4a43-b056-82296e560317-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-x9bvm\" (UID: \"dab7280d-4cb8-4a43-b056-82296e560317\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-x9bvm" Nov 28 14:04:08 crc kubenswrapper[4631]: I1128 14:04:08.470971 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dab7280d-4cb8-4a43-b056-82296e560317-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-x9bvm\" (UID: \"dab7280d-4cb8-4a43-b056-82296e560317\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-x9bvm" Nov 28 14:04:08 crc kubenswrapper[4631]: I1128 14:04:08.470995 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2997r\" (UniqueName: \"kubernetes.io/projected/dab7280d-4cb8-4a43-b056-82296e560317-kube-api-access-2997r\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-x9bvm\" (UID: \"dab7280d-4cb8-4a43-b056-82296e560317\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-x9bvm" Nov 28 14:04:08 crc kubenswrapper[4631]: I1128 14:04:08.471044 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/dab7280d-4cb8-4a43-b056-82296e560317-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-x9bvm\" (UID: \"dab7280d-4cb8-4a43-b056-82296e560317\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-x9bvm" Nov 28 14:04:08 crc kubenswrapper[4631]: I1128 14:04:08.471488 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/dab7280d-4cb8-4a43-b056-82296e560317-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-x9bvm\" (UID: \"dab7280d-4cb8-4a43-b056-82296e560317\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-x9bvm" Nov 28 14:04:08 crc kubenswrapper[4631]: I1128 14:04:08.471527 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/dab7280d-4cb8-4a43-b056-82296e560317-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-x9bvm\" (UID: \"dab7280d-4cb8-4a43-b056-82296e560317\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-x9bvm" Nov 28 14:04:08 crc kubenswrapper[4631]: I1128 14:04:08.471611 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/dab7280d-4cb8-4a43-b056-82296e560317-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-x9bvm\" (UID: \"dab7280d-4cb8-4a43-b056-82296e560317\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-x9bvm" Nov 28 14:04:08 crc kubenswrapper[4631]: I1128 14:04:08.475906 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/dab7280d-4cb8-4a43-b056-82296e560317-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-x9bvm\" (UID: \"dab7280d-4cb8-4a43-b056-82296e560317\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-x9bvm" Nov 28 14:04:08 crc kubenswrapper[4631]: I1128 14:04:08.476000 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dab7280d-4cb8-4a43-b056-82296e560317-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-x9bvm\" (UID: \"dab7280d-4cb8-4a43-b056-82296e560317\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-x9bvm" Nov 28 14:04:08 crc kubenswrapper[4631]: I1128 14:04:08.476327 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/dab7280d-4cb8-4a43-b056-82296e560317-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-x9bvm\" (UID: \"dab7280d-4cb8-4a43-b056-82296e560317\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-x9bvm" Nov 28 14:04:08 crc kubenswrapper[4631]: I1128 14:04:08.480839 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/dab7280d-4cb8-4a43-b056-82296e560317-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-x9bvm\" (UID: \"dab7280d-4cb8-4a43-b056-82296e560317\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-x9bvm" Nov 28 14:04:08 crc kubenswrapper[4631]: I1128 14:04:08.481277 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/dab7280d-4cb8-4a43-b056-82296e560317-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-x9bvm\" (UID: \"dab7280d-4cb8-4a43-b056-82296e560317\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-x9bvm" Nov 28 14:04:08 crc kubenswrapper[4631]: I1128 14:04:08.482249 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/dab7280d-4cb8-4a43-b056-82296e560317-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-x9bvm\" (UID: \"dab7280d-4cb8-4a43-b056-82296e560317\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-x9bvm" Nov 28 14:04:08 crc kubenswrapper[4631]: I1128 14:04:08.488626 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2997r\" (UniqueName: \"kubernetes.io/projected/dab7280d-4cb8-4a43-b056-82296e560317-kube-api-access-2997r\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-x9bvm\" (UID: \"dab7280d-4cb8-4a43-b056-82296e560317\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-x9bvm" Nov 28 14:04:08 crc kubenswrapper[4631]: I1128 14:04:08.599118 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-x9bvm" Nov 28 14:04:09 crc kubenswrapper[4631]: I1128 14:04:09.169410 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-x9bvm"] Nov 28 14:04:10 crc kubenswrapper[4631]: I1128 14:04:10.139577 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-x9bvm" event={"ID":"dab7280d-4cb8-4a43-b056-82296e560317","Type":"ContainerStarted","Data":"34750249d9a86295addbad17baa88f13c533f956e35f4b0b7511e4092cb484e4"} Nov 28 14:04:11 crc kubenswrapper[4631]: I1128 14:04:11.153124 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-x9bvm" event={"ID":"dab7280d-4cb8-4a43-b056-82296e560317","Type":"ContainerStarted","Data":"9e272c9be5c1db267208188c1cd7a39c5fc7bc179b5c829a67bccd3b47d15d32"} Nov 28 14:04:11 crc kubenswrapper[4631]: I1128 14:04:11.181891 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-x9bvm" podStartSLOduration=2.40326833 podStartE2EDuration="3.181866202s" podCreationTimestamp="2025-11-28 14:04:08 +0000 UTC" firstStartedPulling="2025-11-28 14:04:09.178594029 +0000 UTC m=+2605.985897373" lastFinishedPulling="2025-11-28 14:04:09.957191901 +0000 UTC m=+2606.764495245" observedRunningTime="2025-11-28 14:04:11.174984076 +0000 UTC m=+2607.982287440" watchObservedRunningTime="2025-11-28 14:04:11.181866202 +0000 UTC m=+2607.989169546" Nov 28 14:04:13 crc kubenswrapper[4631]: I1128 14:04:13.520095 4631 scope.go:117] "RemoveContainer" containerID="b4f03963475b3c249f705cef33ea3485fc5d4cec16ae5d211e2ba99ac0014be9" Nov 28 14:04:13 crc kubenswrapper[4631]: E1128 14:04:13.523734 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-47sxw_openshift-machine-config-operator(8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1)\"" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" podUID="8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1" Nov 28 14:04:24 crc kubenswrapper[4631]: I1128 14:04:24.513278 4631 scope.go:117] "RemoveContainer" containerID="b4f03963475b3c249f705cef33ea3485fc5d4cec16ae5d211e2ba99ac0014be9" Nov 28 14:04:25 crc kubenswrapper[4631]: I1128 14:04:25.280131 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" event={"ID":"8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1","Type":"ContainerStarted","Data":"8dbc40040e6ec9ba26b8327ffa3d7f1a4f6a3ca257f8ca30274daa85de63606a"} Nov 28 14:05:17 crc kubenswrapper[4631]: I1128 14:05:17.104512 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-dptw9"] Nov 28 14:05:17 crc kubenswrapper[4631]: I1128 14:05:17.107890 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dptw9" Nov 28 14:05:17 crc kubenswrapper[4631]: I1128 14:05:17.177454 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-dptw9"] Nov 28 14:05:17 crc kubenswrapper[4631]: I1128 14:05:17.258161 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/85a5bf9e-034b-4eb8-b926-5471fcbf2244-utilities\") pod \"redhat-operators-dptw9\" (UID: \"85a5bf9e-034b-4eb8-b926-5471fcbf2244\") " pod="openshift-marketplace/redhat-operators-dptw9" Nov 28 14:05:17 crc kubenswrapper[4631]: I1128 14:05:17.258326 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/85a5bf9e-034b-4eb8-b926-5471fcbf2244-catalog-content\") pod \"redhat-operators-dptw9\" (UID: \"85a5bf9e-034b-4eb8-b926-5471fcbf2244\") " pod="openshift-marketplace/redhat-operators-dptw9" Nov 28 14:05:17 crc kubenswrapper[4631]: I1128 14:05:17.258421 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cwkms\" (UniqueName: \"kubernetes.io/projected/85a5bf9e-034b-4eb8-b926-5471fcbf2244-kube-api-access-cwkms\") pod \"redhat-operators-dptw9\" (UID: \"85a5bf9e-034b-4eb8-b926-5471fcbf2244\") " pod="openshift-marketplace/redhat-operators-dptw9" Nov 28 14:05:17 crc kubenswrapper[4631]: I1128 14:05:17.359643 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/85a5bf9e-034b-4eb8-b926-5471fcbf2244-catalog-content\") pod \"redhat-operators-dptw9\" (UID: \"85a5bf9e-034b-4eb8-b926-5471fcbf2244\") " pod="openshift-marketplace/redhat-operators-dptw9" Nov 28 14:05:17 crc kubenswrapper[4631]: I1128 14:05:17.359752 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cwkms\" (UniqueName: \"kubernetes.io/projected/85a5bf9e-034b-4eb8-b926-5471fcbf2244-kube-api-access-cwkms\") pod \"redhat-operators-dptw9\" (UID: \"85a5bf9e-034b-4eb8-b926-5471fcbf2244\") " pod="openshift-marketplace/redhat-operators-dptw9" Nov 28 14:05:17 crc kubenswrapper[4631]: I1128 14:05:17.359844 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/85a5bf9e-034b-4eb8-b926-5471fcbf2244-utilities\") pod \"redhat-operators-dptw9\" (UID: \"85a5bf9e-034b-4eb8-b926-5471fcbf2244\") " pod="openshift-marketplace/redhat-operators-dptw9" Nov 28 14:05:17 crc kubenswrapper[4631]: I1128 14:05:17.360430 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/85a5bf9e-034b-4eb8-b926-5471fcbf2244-utilities\") pod \"redhat-operators-dptw9\" (UID: \"85a5bf9e-034b-4eb8-b926-5471fcbf2244\") " pod="openshift-marketplace/redhat-operators-dptw9" Nov 28 14:05:17 crc kubenswrapper[4631]: I1128 14:05:17.360570 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/85a5bf9e-034b-4eb8-b926-5471fcbf2244-catalog-content\") pod \"redhat-operators-dptw9\" (UID: \"85a5bf9e-034b-4eb8-b926-5471fcbf2244\") " pod="openshift-marketplace/redhat-operators-dptw9" Nov 28 14:05:17 crc kubenswrapper[4631]: I1128 14:05:17.409923 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cwkms\" (UniqueName: \"kubernetes.io/projected/85a5bf9e-034b-4eb8-b926-5471fcbf2244-kube-api-access-cwkms\") pod \"redhat-operators-dptw9\" (UID: \"85a5bf9e-034b-4eb8-b926-5471fcbf2244\") " pod="openshift-marketplace/redhat-operators-dptw9" Nov 28 14:05:17 crc kubenswrapper[4631]: I1128 14:05:17.429994 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dptw9" Nov 28 14:05:17 crc kubenswrapper[4631]: I1128 14:05:17.964625 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-dptw9"] Nov 28 14:05:18 crc kubenswrapper[4631]: I1128 14:05:18.778364 4631 generic.go:334] "Generic (PLEG): container finished" podID="85a5bf9e-034b-4eb8-b926-5471fcbf2244" containerID="f6762609063f154d59baae59a3246f867d60f07bc575a9fe9197bbba6e36eefc" exitCode=0 Nov 28 14:05:18 crc kubenswrapper[4631]: I1128 14:05:18.778603 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dptw9" event={"ID":"85a5bf9e-034b-4eb8-b926-5471fcbf2244","Type":"ContainerDied","Data":"f6762609063f154d59baae59a3246f867d60f07bc575a9fe9197bbba6e36eefc"} Nov 28 14:05:18 crc kubenswrapper[4631]: I1128 14:05:18.778828 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dptw9" event={"ID":"85a5bf9e-034b-4eb8-b926-5471fcbf2244","Type":"ContainerStarted","Data":"b29d706c210d8af5c4ae89210886bb2cd3aefeee0fd102bf5f2123a87ef6fa98"} Nov 28 14:05:19 crc kubenswrapper[4631]: I1128 14:05:19.792462 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dptw9" event={"ID":"85a5bf9e-034b-4eb8-b926-5471fcbf2244","Type":"ContainerStarted","Data":"a528535457355454b98f35b23449bdd4fc6f6c7f8345d2c140ee9be7dcdea503"} Nov 28 14:05:23 crc kubenswrapper[4631]: I1128 14:05:23.829746 4631 generic.go:334] "Generic (PLEG): container finished" podID="85a5bf9e-034b-4eb8-b926-5471fcbf2244" containerID="a528535457355454b98f35b23449bdd4fc6f6c7f8345d2c140ee9be7dcdea503" exitCode=0 Nov 28 14:05:23 crc kubenswrapper[4631]: I1128 14:05:23.829934 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dptw9" event={"ID":"85a5bf9e-034b-4eb8-b926-5471fcbf2244","Type":"ContainerDied","Data":"a528535457355454b98f35b23449bdd4fc6f6c7f8345d2c140ee9be7dcdea503"} Nov 28 14:05:24 crc kubenswrapper[4631]: I1128 14:05:24.840820 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dptw9" event={"ID":"85a5bf9e-034b-4eb8-b926-5471fcbf2244","Type":"ContainerStarted","Data":"a7a5bade4dc0751a4a4d7d2c813da5cad538eb8532645639d925727bac266645"} Nov 28 14:05:24 crc kubenswrapper[4631]: I1128 14:05:24.869117 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-dptw9" podStartSLOduration=2.373420876 podStartE2EDuration="7.869094352s" podCreationTimestamp="2025-11-28 14:05:17 +0000 UTC" firstStartedPulling="2025-11-28 14:05:18.781394637 +0000 UTC m=+2675.588697981" lastFinishedPulling="2025-11-28 14:05:24.277068113 +0000 UTC m=+2681.084371457" observedRunningTime="2025-11-28 14:05:24.859263085 +0000 UTC m=+2681.666566429" watchObservedRunningTime="2025-11-28 14:05:24.869094352 +0000 UTC m=+2681.676397696" Nov 28 14:05:27 crc kubenswrapper[4631]: I1128 14:05:27.432467 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-dptw9" Nov 28 14:05:27 crc kubenswrapper[4631]: I1128 14:05:27.433511 4631 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-dptw9" Nov 28 14:05:28 crc kubenswrapper[4631]: I1128 14:05:28.499505 4631 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-dptw9" podUID="85a5bf9e-034b-4eb8-b926-5471fcbf2244" containerName="registry-server" probeResult="failure" output=< Nov 28 14:05:28 crc kubenswrapper[4631]: timeout: failed to connect service ":50051" within 1s Nov 28 14:05:28 crc kubenswrapper[4631]: > Nov 28 14:05:37 crc kubenswrapper[4631]: I1128 14:05:37.482669 4631 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-dptw9" Nov 28 14:05:37 crc kubenswrapper[4631]: I1128 14:05:37.540001 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-dptw9" Nov 28 14:05:37 crc kubenswrapper[4631]: I1128 14:05:37.720867 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-dptw9"] Nov 28 14:05:38 crc kubenswrapper[4631]: I1128 14:05:38.957010 4631 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-dptw9" podUID="85a5bf9e-034b-4eb8-b926-5471fcbf2244" containerName="registry-server" containerID="cri-o://a7a5bade4dc0751a4a4d7d2c813da5cad538eb8532645639d925727bac266645" gracePeriod=2 Nov 28 14:05:39 crc kubenswrapper[4631]: I1128 14:05:39.452640 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dptw9" Nov 28 14:05:39 crc kubenswrapper[4631]: I1128 14:05:39.618028 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cwkms\" (UniqueName: \"kubernetes.io/projected/85a5bf9e-034b-4eb8-b926-5471fcbf2244-kube-api-access-cwkms\") pod \"85a5bf9e-034b-4eb8-b926-5471fcbf2244\" (UID: \"85a5bf9e-034b-4eb8-b926-5471fcbf2244\") " Nov 28 14:05:39 crc kubenswrapper[4631]: I1128 14:05:39.618099 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/85a5bf9e-034b-4eb8-b926-5471fcbf2244-utilities\") pod \"85a5bf9e-034b-4eb8-b926-5471fcbf2244\" (UID: \"85a5bf9e-034b-4eb8-b926-5471fcbf2244\") " Nov 28 14:05:39 crc kubenswrapper[4631]: I1128 14:05:39.618375 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/85a5bf9e-034b-4eb8-b926-5471fcbf2244-catalog-content\") pod \"85a5bf9e-034b-4eb8-b926-5471fcbf2244\" (UID: \"85a5bf9e-034b-4eb8-b926-5471fcbf2244\") " Nov 28 14:05:39 crc kubenswrapper[4631]: I1128 14:05:39.619236 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/85a5bf9e-034b-4eb8-b926-5471fcbf2244-utilities" (OuterVolumeSpecName: "utilities") pod "85a5bf9e-034b-4eb8-b926-5471fcbf2244" (UID: "85a5bf9e-034b-4eb8-b926-5471fcbf2244"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 14:05:39 crc kubenswrapper[4631]: I1128 14:05:39.632425 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/85a5bf9e-034b-4eb8-b926-5471fcbf2244-kube-api-access-cwkms" (OuterVolumeSpecName: "kube-api-access-cwkms") pod "85a5bf9e-034b-4eb8-b926-5471fcbf2244" (UID: "85a5bf9e-034b-4eb8-b926-5471fcbf2244"). InnerVolumeSpecName "kube-api-access-cwkms". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:05:39 crc kubenswrapper[4631]: I1128 14:05:39.722345 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cwkms\" (UniqueName: \"kubernetes.io/projected/85a5bf9e-034b-4eb8-b926-5471fcbf2244-kube-api-access-cwkms\") on node \"crc\" DevicePath \"\"" Nov 28 14:05:39 crc kubenswrapper[4631]: I1128 14:05:39.723174 4631 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/85a5bf9e-034b-4eb8-b926-5471fcbf2244-utilities\") on node \"crc\" DevicePath \"\"" Nov 28 14:05:39 crc kubenswrapper[4631]: I1128 14:05:39.748341 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/85a5bf9e-034b-4eb8-b926-5471fcbf2244-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "85a5bf9e-034b-4eb8-b926-5471fcbf2244" (UID: "85a5bf9e-034b-4eb8-b926-5471fcbf2244"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 14:05:39 crc kubenswrapper[4631]: I1128 14:05:39.824546 4631 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/85a5bf9e-034b-4eb8-b926-5471fcbf2244-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 28 14:05:39 crc kubenswrapper[4631]: I1128 14:05:39.968616 4631 generic.go:334] "Generic (PLEG): container finished" podID="85a5bf9e-034b-4eb8-b926-5471fcbf2244" containerID="a7a5bade4dc0751a4a4d7d2c813da5cad538eb8532645639d925727bac266645" exitCode=0 Nov 28 14:05:39 crc kubenswrapper[4631]: I1128 14:05:39.968676 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dptw9" event={"ID":"85a5bf9e-034b-4eb8-b926-5471fcbf2244","Type":"ContainerDied","Data":"a7a5bade4dc0751a4a4d7d2c813da5cad538eb8532645639d925727bac266645"} Nov 28 14:05:39 crc kubenswrapper[4631]: I1128 14:05:39.968712 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dptw9" event={"ID":"85a5bf9e-034b-4eb8-b926-5471fcbf2244","Type":"ContainerDied","Data":"b29d706c210d8af5c4ae89210886bb2cd3aefeee0fd102bf5f2123a87ef6fa98"} Nov 28 14:05:39 crc kubenswrapper[4631]: I1128 14:05:39.968704 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dptw9" Nov 28 14:05:39 crc kubenswrapper[4631]: I1128 14:05:39.968730 4631 scope.go:117] "RemoveContainer" containerID="a7a5bade4dc0751a4a4d7d2c813da5cad538eb8532645639d925727bac266645" Nov 28 14:05:39 crc kubenswrapper[4631]: I1128 14:05:39.994568 4631 scope.go:117] "RemoveContainer" containerID="a528535457355454b98f35b23449bdd4fc6f6c7f8345d2c140ee9be7dcdea503" Nov 28 14:05:40 crc kubenswrapper[4631]: I1128 14:05:40.009465 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-dptw9"] Nov 28 14:05:40 crc kubenswrapper[4631]: I1128 14:05:40.028700 4631 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-dptw9"] Nov 28 14:05:40 crc kubenswrapper[4631]: I1128 14:05:40.039641 4631 scope.go:117] "RemoveContainer" containerID="f6762609063f154d59baae59a3246f867d60f07bc575a9fe9197bbba6e36eefc" Nov 28 14:05:40 crc kubenswrapper[4631]: I1128 14:05:40.073897 4631 scope.go:117] "RemoveContainer" containerID="a7a5bade4dc0751a4a4d7d2c813da5cad538eb8532645639d925727bac266645" Nov 28 14:05:40 crc kubenswrapper[4631]: E1128 14:05:40.074483 4631 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a7a5bade4dc0751a4a4d7d2c813da5cad538eb8532645639d925727bac266645\": container with ID starting with a7a5bade4dc0751a4a4d7d2c813da5cad538eb8532645639d925727bac266645 not found: ID does not exist" containerID="a7a5bade4dc0751a4a4d7d2c813da5cad538eb8532645639d925727bac266645" Nov 28 14:05:40 crc kubenswrapper[4631]: I1128 14:05:40.074545 4631 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a7a5bade4dc0751a4a4d7d2c813da5cad538eb8532645639d925727bac266645"} err="failed to get container status \"a7a5bade4dc0751a4a4d7d2c813da5cad538eb8532645639d925727bac266645\": rpc error: code = NotFound desc = could not find container \"a7a5bade4dc0751a4a4d7d2c813da5cad538eb8532645639d925727bac266645\": container with ID starting with a7a5bade4dc0751a4a4d7d2c813da5cad538eb8532645639d925727bac266645 not found: ID does not exist" Nov 28 14:05:40 crc kubenswrapper[4631]: I1128 14:05:40.074583 4631 scope.go:117] "RemoveContainer" containerID="a528535457355454b98f35b23449bdd4fc6f6c7f8345d2c140ee9be7dcdea503" Nov 28 14:05:40 crc kubenswrapper[4631]: E1128 14:05:40.074987 4631 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a528535457355454b98f35b23449bdd4fc6f6c7f8345d2c140ee9be7dcdea503\": container with ID starting with a528535457355454b98f35b23449bdd4fc6f6c7f8345d2c140ee9be7dcdea503 not found: ID does not exist" containerID="a528535457355454b98f35b23449bdd4fc6f6c7f8345d2c140ee9be7dcdea503" Nov 28 14:05:40 crc kubenswrapper[4631]: I1128 14:05:40.075022 4631 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a528535457355454b98f35b23449bdd4fc6f6c7f8345d2c140ee9be7dcdea503"} err="failed to get container status \"a528535457355454b98f35b23449bdd4fc6f6c7f8345d2c140ee9be7dcdea503\": rpc error: code = NotFound desc = could not find container \"a528535457355454b98f35b23449bdd4fc6f6c7f8345d2c140ee9be7dcdea503\": container with ID starting with a528535457355454b98f35b23449bdd4fc6f6c7f8345d2c140ee9be7dcdea503 not found: ID does not exist" Nov 28 14:05:40 crc kubenswrapper[4631]: I1128 14:05:40.075046 4631 scope.go:117] "RemoveContainer" containerID="f6762609063f154d59baae59a3246f867d60f07bc575a9fe9197bbba6e36eefc" Nov 28 14:05:40 crc kubenswrapper[4631]: E1128 14:05:40.075349 4631 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f6762609063f154d59baae59a3246f867d60f07bc575a9fe9197bbba6e36eefc\": container with ID starting with f6762609063f154d59baae59a3246f867d60f07bc575a9fe9197bbba6e36eefc not found: ID does not exist" containerID="f6762609063f154d59baae59a3246f867d60f07bc575a9fe9197bbba6e36eefc" Nov 28 14:05:40 crc kubenswrapper[4631]: I1128 14:05:40.075383 4631 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f6762609063f154d59baae59a3246f867d60f07bc575a9fe9197bbba6e36eefc"} err="failed to get container status \"f6762609063f154d59baae59a3246f867d60f07bc575a9fe9197bbba6e36eefc\": rpc error: code = NotFound desc = could not find container \"f6762609063f154d59baae59a3246f867d60f07bc575a9fe9197bbba6e36eefc\": container with ID starting with f6762609063f154d59baae59a3246f867d60f07bc575a9fe9197bbba6e36eefc not found: ID does not exist" Nov 28 14:05:41 crc kubenswrapper[4631]: I1128 14:05:41.534912 4631 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="85a5bf9e-034b-4eb8-b926-5471fcbf2244" path="/var/lib/kubelet/pods/85a5bf9e-034b-4eb8-b926-5471fcbf2244/volumes" Nov 28 14:05:42 crc kubenswrapper[4631]: I1128 14:05:42.927344 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-62kfw"] Nov 28 14:05:42 crc kubenswrapper[4631]: E1128 14:05:42.928219 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="85a5bf9e-034b-4eb8-b926-5471fcbf2244" containerName="extract-content" Nov 28 14:05:42 crc kubenswrapper[4631]: I1128 14:05:42.928239 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="85a5bf9e-034b-4eb8-b926-5471fcbf2244" containerName="extract-content" Nov 28 14:05:42 crc kubenswrapper[4631]: E1128 14:05:42.928262 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="85a5bf9e-034b-4eb8-b926-5471fcbf2244" containerName="registry-server" Nov 28 14:05:42 crc kubenswrapper[4631]: I1128 14:05:42.928271 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="85a5bf9e-034b-4eb8-b926-5471fcbf2244" containerName="registry-server" Nov 28 14:05:42 crc kubenswrapper[4631]: E1128 14:05:42.928896 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="85a5bf9e-034b-4eb8-b926-5471fcbf2244" containerName="extract-utilities" Nov 28 14:05:42 crc kubenswrapper[4631]: I1128 14:05:42.928912 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="85a5bf9e-034b-4eb8-b926-5471fcbf2244" containerName="extract-utilities" Nov 28 14:05:42 crc kubenswrapper[4631]: I1128 14:05:42.929189 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="85a5bf9e-034b-4eb8-b926-5471fcbf2244" containerName="registry-server" Nov 28 14:05:42 crc kubenswrapper[4631]: I1128 14:05:42.931605 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-62kfw" Nov 28 14:05:42 crc kubenswrapper[4631]: I1128 14:05:42.939789 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-62kfw"] Nov 28 14:05:43 crc kubenswrapper[4631]: I1128 14:05:43.086459 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/da976722-0313-4d07-8ead-6f9dcb3bfb03-catalog-content\") pod \"community-operators-62kfw\" (UID: \"da976722-0313-4d07-8ead-6f9dcb3bfb03\") " pod="openshift-marketplace/community-operators-62kfw" Nov 28 14:05:43 crc kubenswrapper[4631]: I1128 14:05:43.086514 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/da976722-0313-4d07-8ead-6f9dcb3bfb03-utilities\") pod \"community-operators-62kfw\" (UID: \"da976722-0313-4d07-8ead-6f9dcb3bfb03\") " pod="openshift-marketplace/community-operators-62kfw" Nov 28 14:05:43 crc kubenswrapper[4631]: I1128 14:05:43.086602 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cvllb\" (UniqueName: \"kubernetes.io/projected/da976722-0313-4d07-8ead-6f9dcb3bfb03-kube-api-access-cvllb\") pod \"community-operators-62kfw\" (UID: \"da976722-0313-4d07-8ead-6f9dcb3bfb03\") " pod="openshift-marketplace/community-operators-62kfw" Nov 28 14:05:43 crc kubenswrapper[4631]: I1128 14:05:43.189148 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cvllb\" (UniqueName: \"kubernetes.io/projected/da976722-0313-4d07-8ead-6f9dcb3bfb03-kube-api-access-cvllb\") pod \"community-operators-62kfw\" (UID: \"da976722-0313-4d07-8ead-6f9dcb3bfb03\") " pod="openshift-marketplace/community-operators-62kfw" Nov 28 14:05:43 crc kubenswrapper[4631]: I1128 14:05:43.189356 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/da976722-0313-4d07-8ead-6f9dcb3bfb03-catalog-content\") pod \"community-operators-62kfw\" (UID: \"da976722-0313-4d07-8ead-6f9dcb3bfb03\") " pod="openshift-marketplace/community-operators-62kfw" Nov 28 14:05:43 crc kubenswrapper[4631]: I1128 14:05:43.189397 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/da976722-0313-4d07-8ead-6f9dcb3bfb03-utilities\") pod \"community-operators-62kfw\" (UID: \"da976722-0313-4d07-8ead-6f9dcb3bfb03\") " pod="openshift-marketplace/community-operators-62kfw" Nov 28 14:05:43 crc kubenswrapper[4631]: I1128 14:05:43.189975 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/da976722-0313-4d07-8ead-6f9dcb3bfb03-catalog-content\") pod \"community-operators-62kfw\" (UID: \"da976722-0313-4d07-8ead-6f9dcb3bfb03\") " pod="openshift-marketplace/community-operators-62kfw" Nov 28 14:05:43 crc kubenswrapper[4631]: I1128 14:05:43.190034 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/da976722-0313-4d07-8ead-6f9dcb3bfb03-utilities\") pod \"community-operators-62kfw\" (UID: \"da976722-0313-4d07-8ead-6f9dcb3bfb03\") " pod="openshift-marketplace/community-operators-62kfw" Nov 28 14:05:43 crc kubenswrapper[4631]: I1128 14:05:43.210223 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cvllb\" (UniqueName: \"kubernetes.io/projected/da976722-0313-4d07-8ead-6f9dcb3bfb03-kube-api-access-cvllb\") pod \"community-operators-62kfw\" (UID: \"da976722-0313-4d07-8ead-6f9dcb3bfb03\") " pod="openshift-marketplace/community-operators-62kfw" Nov 28 14:05:43 crc kubenswrapper[4631]: I1128 14:05:43.257018 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-62kfw" Nov 28 14:05:43 crc kubenswrapper[4631]: I1128 14:05:43.996206 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-62kfw"] Nov 28 14:05:44 crc kubenswrapper[4631]: W1128 14:05:44.006641 4631 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podda976722_0313_4d07_8ead_6f9dcb3bfb03.slice/crio-bc3f8717d8bdb1806b682fd1a69e79ef543daa1423a56fca9c362d9bad7d3f65 WatchSource:0}: Error finding container bc3f8717d8bdb1806b682fd1a69e79ef543daa1423a56fca9c362d9bad7d3f65: Status 404 returned error can't find the container with id bc3f8717d8bdb1806b682fd1a69e79ef543daa1423a56fca9c362d9bad7d3f65 Nov 28 14:05:45 crc kubenswrapper[4631]: I1128 14:05:45.023000 4631 generic.go:334] "Generic (PLEG): container finished" podID="da976722-0313-4d07-8ead-6f9dcb3bfb03" containerID="f4d36af863e6f4f984ee7c7ff9abc69f2b236ba1c9f43725261c902c1b20c534" exitCode=0 Nov 28 14:05:45 crc kubenswrapper[4631]: I1128 14:05:45.023079 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-62kfw" event={"ID":"da976722-0313-4d07-8ead-6f9dcb3bfb03","Type":"ContainerDied","Data":"f4d36af863e6f4f984ee7c7ff9abc69f2b236ba1c9f43725261c902c1b20c534"} Nov 28 14:05:45 crc kubenswrapper[4631]: I1128 14:05:45.023108 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-62kfw" event={"ID":"da976722-0313-4d07-8ead-6f9dcb3bfb03","Type":"ContainerStarted","Data":"bc3f8717d8bdb1806b682fd1a69e79ef543daa1423a56fca9c362d9bad7d3f65"} Nov 28 14:05:46 crc kubenswrapper[4631]: I1128 14:05:46.033913 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-62kfw" event={"ID":"da976722-0313-4d07-8ead-6f9dcb3bfb03","Type":"ContainerStarted","Data":"8845d7819ad2597e97a074dd42206270ef384e7c0a9ecfe361f905d26c5e6719"} Nov 28 14:05:47 crc kubenswrapper[4631]: I1128 14:05:47.045708 4631 generic.go:334] "Generic (PLEG): container finished" podID="da976722-0313-4d07-8ead-6f9dcb3bfb03" containerID="8845d7819ad2597e97a074dd42206270ef384e7c0a9ecfe361f905d26c5e6719" exitCode=0 Nov 28 14:05:47 crc kubenswrapper[4631]: I1128 14:05:47.045782 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-62kfw" event={"ID":"da976722-0313-4d07-8ead-6f9dcb3bfb03","Type":"ContainerDied","Data":"8845d7819ad2597e97a074dd42206270ef384e7c0a9ecfe361f905d26c5e6719"} Nov 28 14:05:48 crc kubenswrapper[4631]: I1128 14:05:48.055401 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-62kfw" event={"ID":"da976722-0313-4d07-8ead-6f9dcb3bfb03","Type":"ContainerStarted","Data":"301e3f12f4dcdce552fb86ebd49cdbfcdf4222c879a10248067bc99a01c9201a"} Nov 28 14:05:48 crc kubenswrapper[4631]: I1128 14:05:48.091953 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-62kfw" podStartSLOduration=3.44389106 podStartE2EDuration="6.091933299s" podCreationTimestamp="2025-11-28 14:05:42 +0000 UTC" firstStartedPulling="2025-11-28 14:05:45.027830967 +0000 UTC m=+2701.835134311" lastFinishedPulling="2025-11-28 14:05:47.675873216 +0000 UTC m=+2704.483176550" observedRunningTime="2025-11-28 14:05:48.0721244 +0000 UTC m=+2704.879427754" watchObservedRunningTime="2025-11-28 14:05:48.091933299 +0000 UTC m=+2704.899236643" Nov 28 14:05:53 crc kubenswrapper[4631]: I1128 14:05:53.257570 4631 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-62kfw" Nov 28 14:05:53 crc kubenswrapper[4631]: I1128 14:05:53.258466 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-62kfw" Nov 28 14:05:53 crc kubenswrapper[4631]: I1128 14:05:53.360218 4631 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-62kfw" Nov 28 14:05:54 crc kubenswrapper[4631]: I1128 14:05:54.167520 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-62kfw" Nov 28 14:05:54 crc kubenswrapper[4631]: I1128 14:05:54.249996 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-62kfw"] Nov 28 14:05:56 crc kubenswrapper[4631]: I1128 14:05:56.130187 4631 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-62kfw" podUID="da976722-0313-4d07-8ead-6f9dcb3bfb03" containerName="registry-server" containerID="cri-o://301e3f12f4dcdce552fb86ebd49cdbfcdf4222c879a10248067bc99a01c9201a" gracePeriod=2 Nov 28 14:05:57 crc kubenswrapper[4631]: I1128 14:05:57.162400 4631 generic.go:334] "Generic (PLEG): container finished" podID="da976722-0313-4d07-8ead-6f9dcb3bfb03" containerID="301e3f12f4dcdce552fb86ebd49cdbfcdf4222c879a10248067bc99a01c9201a" exitCode=0 Nov 28 14:05:57 crc kubenswrapper[4631]: I1128 14:05:57.162559 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-62kfw" event={"ID":"da976722-0313-4d07-8ead-6f9dcb3bfb03","Type":"ContainerDied","Data":"301e3f12f4dcdce552fb86ebd49cdbfcdf4222c879a10248067bc99a01c9201a"} Nov 28 14:05:57 crc kubenswrapper[4631]: I1128 14:05:57.162685 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-62kfw" event={"ID":"da976722-0313-4d07-8ead-6f9dcb3bfb03","Type":"ContainerDied","Data":"bc3f8717d8bdb1806b682fd1a69e79ef543daa1423a56fca9c362d9bad7d3f65"} Nov 28 14:05:57 crc kubenswrapper[4631]: I1128 14:05:57.162702 4631 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bc3f8717d8bdb1806b682fd1a69e79ef543daa1423a56fca9c362d9bad7d3f65" Nov 28 14:05:57 crc kubenswrapper[4631]: I1128 14:05:57.194752 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-62kfw" Nov 28 14:05:57 crc kubenswrapper[4631]: I1128 14:05:57.306835 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cvllb\" (UniqueName: \"kubernetes.io/projected/da976722-0313-4d07-8ead-6f9dcb3bfb03-kube-api-access-cvllb\") pod \"da976722-0313-4d07-8ead-6f9dcb3bfb03\" (UID: \"da976722-0313-4d07-8ead-6f9dcb3bfb03\") " Nov 28 14:05:57 crc kubenswrapper[4631]: I1128 14:05:57.307217 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/da976722-0313-4d07-8ead-6f9dcb3bfb03-catalog-content\") pod \"da976722-0313-4d07-8ead-6f9dcb3bfb03\" (UID: \"da976722-0313-4d07-8ead-6f9dcb3bfb03\") " Nov 28 14:05:57 crc kubenswrapper[4631]: I1128 14:05:57.307325 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/da976722-0313-4d07-8ead-6f9dcb3bfb03-utilities\") pod \"da976722-0313-4d07-8ead-6f9dcb3bfb03\" (UID: \"da976722-0313-4d07-8ead-6f9dcb3bfb03\") " Nov 28 14:05:57 crc kubenswrapper[4631]: I1128 14:05:57.308529 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/da976722-0313-4d07-8ead-6f9dcb3bfb03-utilities" (OuterVolumeSpecName: "utilities") pod "da976722-0313-4d07-8ead-6f9dcb3bfb03" (UID: "da976722-0313-4d07-8ead-6f9dcb3bfb03"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 14:05:57 crc kubenswrapper[4631]: I1128 14:05:57.322980 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/da976722-0313-4d07-8ead-6f9dcb3bfb03-kube-api-access-cvllb" (OuterVolumeSpecName: "kube-api-access-cvllb") pod "da976722-0313-4d07-8ead-6f9dcb3bfb03" (UID: "da976722-0313-4d07-8ead-6f9dcb3bfb03"). InnerVolumeSpecName "kube-api-access-cvllb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:05:57 crc kubenswrapper[4631]: I1128 14:05:57.366605 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/da976722-0313-4d07-8ead-6f9dcb3bfb03-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "da976722-0313-4d07-8ead-6f9dcb3bfb03" (UID: "da976722-0313-4d07-8ead-6f9dcb3bfb03"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 14:05:57 crc kubenswrapper[4631]: I1128 14:05:57.409158 4631 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/da976722-0313-4d07-8ead-6f9dcb3bfb03-utilities\") on node \"crc\" DevicePath \"\"" Nov 28 14:05:57 crc kubenswrapper[4631]: I1128 14:05:57.409231 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cvllb\" (UniqueName: \"kubernetes.io/projected/da976722-0313-4d07-8ead-6f9dcb3bfb03-kube-api-access-cvllb\") on node \"crc\" DevicePath \"\"" Nov 28 14:05:57 crc kubenswrapper[4631]: I1128 14:05:57.409254 4631 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/da976722-0313-4d07-8ead-6f9dcb3bfb03-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 28 14:05:58 crc kubenswrapper[4631]: I1128 14:05:58.172924 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-62kfw" Nov 28 14:05:58 crc kubenswrapper[4631]: I1128 14:05:58.205851 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-62kfw"] Nov 28 14:05:58 crc kubenswrapper[4631]: I1128 14:05:58.214543 4631 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-62kfw"] Nov 28 14:05:59 crc kubenswrapper[4631]: I1128 14:05:59.524040 4631 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="da976722-0313-4d07-8ead-6f9dcb3bfb03" path="/var/lib/kubelet/pods/da976722-0313-4d07-8ead-6f9dcb3bfb03/volumes" Nov 28 14:06:49 crc kubenswrapper[4631]: I1128 14:06:49.635304 4631 patch_prober.go:28] interesting pod/machine-config-daemon-47sxw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 28 14:06:49 crc kubenswrapper[4631]: I1128 14:06:49.635887 4631 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" podUID="8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 28 14:07:11 crc kubenswrapper[4631]: I1128 14:07:11.891388 4631 generic.go:334] "Generic (PLEG): container finished" podID="dab7280d-4cb8-4a43-b056-82296e560317" containerID="9e272c9be5c1db267208188c1cd7a39c5fc7bc179b5c829a67bccd3b47d15d32" exitCode=0 Nov 28 14:07:11 crc kubenswrapper[4631]: I1128 14:07:11.891502 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-x9bvm" event={"ID":"dab7280d-4cb8-4a43-b056-82296e560317","Type":"ContainerDied","Data":"9e272c9be5c1db267208188c1cd7a39c5fc7bc179b5c829a67bccd3b47d15d32"} Nov 28 14:07:13 crc kubenswrapper[4631]: I1128 14:07:13.372111 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-x9bvm" Nov 28 14:07:13 crc kubenswrapper[4631]: I1128 14:07:13.492136 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dab7280d-4cb8-4a43-b056-82296e560317-telemetry-combined-ca-bundle\") pod \"dab7280d-4cb8-4a43-b056-82296e560317\" (UID: \"dab7280d-4cb8-4a43-b056-82296e560317\") " Nov 28 14:07:13 crc kubenswrapper[4631]: I1128 14:07:13.492217 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2997r\" (UniqueName: \"kubernetes.io/projected/dab7280d-4cb8-4a43-b056-82296e560317-kube-api-access-2997r\") pod \"dab7280d-4cb8-4a43-b056-82296e560317\" (UID: \"dab7280d-4cb8-4a43-b056-82296e560317\") " Nov 28 14:07:13 crc kubenswrapper[4631]: I1128 14:07:13.492273 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/dab7280d-4cb8-4a43-b056-82296e560317-ceilometer-compute-config-data-0\") pod \"dab7280d-4cb8-4a43-b056-82296e560317\" (UID: \"dab7280d-4cb8-4a43-b056-82296e560317\") " Nov 28 14:07:13 crc kubenswrapper[4631]: I1128 14:07:13.492410 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/dab7280d-4cb8-4a43-b056-82296e560317-ceilometer-compute-config-data-2\") pod \"dab7280d-4cb8-4a43-b056-82296e560317\" (UID: \"dab7280d-4cb8-4a43-b056-82296e560317\") " Nov 28 14:07:13 crc kubenswrapper[4631]: I1128 14:07:13.492446 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/dab7280d-4cb8-4a43-b056-82296e560317-ssh-key\") pod \"dab7280d-4cb8-4a43-b056-82296e560317\" (UID: \"dab7280d-4cb8-4a43-b056-82296e560317\") " Nov 28 14:07:13 crc kubenswrapper[4631]: I1128 14:07:13.492465 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/dab7280d-4cb8-4a43-b056-82296e560317-inventory\") pod \"dab7280d-4cb8-4a43-b056-82296e560317\" (UID: \"dab7280d-4cb8-4a43-b056-82296e560317\") " Nov 28 14:07:13 crc kubenswrapper[4631]: I1128 14:07:13.492643 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/dab7280d-4cb8-4a43-b056-82296e560317-ceilometer-compute-config-data-1\") pod \"dab7280d-4cb8-4a43-b056-82296e560317\" (UID: \"dab7280d-4cb8-4a43-b056-82296e560317\") " Nov 28 14:07:13 crc kubenswrapper[4631]: I1128 14:07:13.498772 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dab7280d-4cb8-4a43-b056-82296e560317-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "dab7280d-4cb8-4a43-b056-82296e560317" (UID: "dab7280d-4cb8-4a43-b056-82296e560317"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:07:13 crc kubenswrapper[4631]: I1128 14:07:13.498883 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dab7280d-4cb8-4a43-b056-82296e560317-kube-api-access-2997r" (OuterVolumeSpecName: "kube-api-access-2997r") pod "dab7280d-4cb8-4a43-b056-82296e560317" (UID: "dab7280d-4cb8-4a43-b056-82296e560317"). InnerVolumeSpecName "kube-api-access-2997r". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:07:13 crc kubenswrapper[4631]: I1128 14:07:13.523844 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dab7280d-4cb8-4a43-b056-82296e560317-inventory" (OuterVolumeSpecName: "inventory") pod "dab7280d-4cb8-4a43-b056-82296e560317" (UID: "dab7280d-4cb8-4a43-b056-82296e560317"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:07:13 crc kubenswrapper[4631]: I1128 14:07:13.531250 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dab7280d-4cb8-4a43-b056-82296e560317-ceilometer-compute-config-data-2" (OuterVolumeSpecName: "ceilometer-compute-config-data-2") pod "dab7280d-4cb8-4a43-b056-82296e560317" (UID: "dab7280d-4cb8-4a43-b056-82296e560317"). InnerVolumeSpecName "ceilometer-compute-config-data-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:07:13 crc kubenswrapper[4631]: I1128 14:07:13.533856 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dab7280d-4cb8-4a43-b056-82296e560317-ceilometer-compute-config-data-0" (OuterVolumeSpecName: "ceilometer-compute-config-data-0") pod "dab7280d-4cb8-4a43-b056-82296e560317" (UID: "dab7280d-4cb8-4a43-b056-82296e560317"). InnerVolumeSpecName "ceilometer-compute-config-data-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:07:13 crc kubenswrapper[4631]: I1128 14:07:13.534504 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dab7280d-4cb8-4a43-b056-82296e560317-ceilometer-compute-config-data-1" (OuterVolumeSpecName: "ceilometer-compute-config-data-1") pod "dab7280d-4cb8-4a43-b056-82296e560317" (UID: "dab7280d-4cb8-4a43-b056-82296e560317"). InnerVolumeSpecName "ceilometer-compute-config-data-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:07:13 crc kubenswrapper[4631]: I1128 14:07:13.543402 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dab7280d-4cb8-4a43-b056-82296e560317-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "dab7280d-4cb8-4a43-b056-82296e560317" (UID: "dab7280d-4cb8-4a43-b056-82296e560317"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:07:13 crc kubenswrapper[4631]: I1128 14:07:13.595777 4631 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/dab7280d-4cb8-4a43-b056-82296e560317-ceilometer-compute-config-data-1\") on node \"crc\" DevicePath \"\"" Nov 28 14:07:13 crc kubenswrapper[4631]: I1128 14:07:13.596095 4631 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dab7280d-4cb8-4a43-b056-82296e560317-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 14:07:13 crc kubenswrapper[4631]: I1128 14:07:13.596107 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2997r\" (UniqueName: \"kubernetes.io/projected/dab7280d-4cb8-4a43-b056-82296e560317-kube-api-access-2997r\") on node \"crc\" DevicePath \"\"" Nov 28 14:07:13 crc kubenswrapper[4631]: I1128 14:07:13.596118 4631 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/dab7280d-4cb8-4a43-b056-82296e560317-ceilometer-compute-config-data-0\") on node \"crc\" DevicePath \"\"" Nov 28 14:07:13 crc kubenswrapper[4631]: I1128 14:07:13.596170 4631 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/dab7280d-4cb8-4a43-b056-82296e560317-ceilometer-compute-config-data-2\") on node \"crc\" DevicePath \"\"" Nov 28 14:07:13 crc kubenswrapper[4631]: I1128 14:07:13.597355 4631 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/dab7280d-4cb8-4a43-b056-82296e560317-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 28 14:07:13 crc kubenswrapper[4631]: I1128 14:07:13.597390 4631 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/dab7280d-4cb8-4a43-b056-82296e560317-inventory\") on node \"crc\" DevicePath \"\"" Nov 28 14:07:13 crc kubenswrapper[4631]: I1128 14:07:13.911523 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-x9bvm" event={"ID":"dab7280d-4cb8-4a43-b056-82296e560317","Type":"ContainerDied","Data":"34750249d9a86295addbad17baa88f13c533f956e35f4b0b7511e4092cb484e4"} Nov 28 14:07:13 crc kubenswrapper[4631]: I1128 14:07:13.911564 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-x9bvm" Nov 28 14:07:13 crc kubenswrapper[4631]: I1128 14:07:13.911570 4631 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="34750249d9a86295addbad17baa88f13c533f956e35f4b0b7511e4092cb484e4" Nov 28 14:07:19 crc kubenswrapper[4631]: I1128 14:07:19.634641 4631 patch_prober.go:28] interesting pod/machine-config-daemon-47sxw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 28 14:07:19 crc kubenswrapper[4631]: I1128 14:07:19.635350 4631 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" podUID="8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 28 14:07:49 crc kubenswrapper[4631]: I1128 14:07:49.635163 4631 patch_prober.go:28] interesting pod/machine-config-daemon-47sxw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 28 14:07:49 crc kubenswrapper[4631]: I1128 14:07:49.635736 4631 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" podUID="8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 28 14:07:49 crc kubenswrapper[4631]: I1128 14:07:49.635785 4631 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" Nov 28 14:07:49 crc kubenswrapper[4631]: I1128 14:07:49.637054 4631 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"8dbc40040e6ec9ba26b8327ffa3d7f1a4f6a3ca257f8ca30274daa85de63606a"} pod="openshift-machine-config-operator/machine-config-daemon-47sxw" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 28 14:07:49 crc kubenswrapper[4631]: I1128 14:07:49.637180 4631 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" podUID="8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1" containerName="machine-config-daemon" containerID="cri-o://8dbc40040e6ec9ba26b8327ffa3d7f1a4f6a3ca257f8ca30274daa85de63606a" gracePeriod=600 Nov 28 14:07:50 crc kubenswrapper[4631]: I1128 14:07:50.244583 4631 generic.go:334] "Generic (PLEG): container finished" podID="8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1" containerID="8dbc40040e6ec9ba26b8327ffa3d7f1a4f6a3ca257f8ca30274daa85de63606a" exitCode=0 Nov 28 14:07:50 crc kubenswrapper[4631]: I1128 14:07:50.244632 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" event={"ID":"8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1","Type":"ContainerDied","Data":"8dbc40040e6ec9ba26b8327ffa3d7f1a4f6a3ca257f8ca30274daa85de63606a"} Nov 28 14:07:50 crc kubenswrapper[4631]: I1128 14:07:50.244668 4631 scope.go:117] "RemoveContainer" containerID="b4f03963475b3c249f705cef33ea3485fc5d4cec16ae5d211e2ba99ac0014be9" Nov 28 14:07:51 crc kubenswrapper[4631]: I1128 14:07:51.257154 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" event={"ID":"8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1","Type":"ContainerStarted","Data":"d188114d99f65821ef24c4d713f07ed3cbe0cd10a5aaa23d6346374210035dc4"} Nov 28 14:08:14 crc kubenswrapper[4631]: I1128 14:08:14.928409 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/tempest-tests-tempest"] Nov 28 14:08:14 crc kubenswrapper[4631]: E1128 14:08:14.929390 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dab7280d-4cb8-4a43-b056-82296e560317" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Nov 28 14:08:14 crc kubenswrapper[4631]: I1128 14:08:14.929408 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="dab7280d-4cb8-4a43-b056-82296e560317" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Nov 28 14:08:14 crc kubenswrapper[4631]: E1128 14:08:14.929420 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="da976722-0313-4d07-8ead-6f9dcb3bfb03" containerName="extract-content" Nov 28 14:08:14 crc kubenswrapper[4631]: I1128 14:08:14.929427 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="da976722-0313-4d07-8ead-6f9dcb3bfb03" containerName="extract-content" Nov 28 14:08:14 crc kubenswrapper[4631]: E1128 14:08:14.929456 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="da976722-0313-4d07-8ead-6f9dcb3bfb03" containerName="extract-utilities" Nov 28 14:08:14 crc kubenswrapper[4631]: I1128 14:08:14.929463 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="da976722-0313-4d07-8ead-6f9dcb3bfb03" containerName="extract-utilities" Nov 28 14:08:14 crc kubenswrapper[4631]: E1128 14:08:14.929474 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="da976722-0313-4d07-8ead-6f9dcb3bfb03" containerName="registry-server" Nov 28 14:08:14 crc kubenswrapper[4631]: I1128 14:08:14.929481 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="da976722-0313-4d07-8ead-6f9dcb3bfb03" containerName="registry-server" Nov 28 14:08:14 crc kubenswrapper[4631]: I1128 14:08:14.929699 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="da976722-0313-4d07-8ead-6f9dcb3bfb03" containerName="registry-server" Nov 28 14:08:14 crc kubenswrapper[4631]: I1128 14:08:14.929717 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="dab7280d-4cb8-4a43-b056-82296e560317" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Nov 28 14:08:14 crc kubenswrapper[4631]: I1128 14:08:14.930596 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Nov 28 14:08:14 crc kubenswrapper[4631]: I1128 14:08:14.933907 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"test-operator-controller-priv-key" Nov 28 14:08:14 crc kubenswrapper[4631]: I1128 14:08:14.934215 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-tqp7g" Nov 28 14:08:14 crc kubenswrapper[4631]: I1128 14:08:14.937871 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Nov 28 14:08:14 crc kubenswrapper[4631]: I1128 14:08:14.938101 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-custom-data-s0" Nov 28 14:08:14 crc kubenswrapper[4631]: I1128 14:08:14.944386 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Nov 28 14:08:15 crc kubenswrapper[4631]: I1128 14:08:15.126918 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/183d2a7b-cd4e-423c-82c0-a72e2ebec301-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"183d2a7b-cd4e-423c-82c0-a72e2ebec301\") " pod="openstack/tempest-tests-tempest" Nov 28 14:08:15 crc kubenswrapper[4631]: I1128 14:08:15.127011 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/183d2a7b-cd4e-423c-82c0-a72e2ebec301-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"183d2a7b-cd4e-423c-82c0-a72e2ebec301\") " pod="openstack/tempest-tests-tempest" Nov 28 14:08:15 crc kubenswrapper[4631]: I1128 14:08:15.127036 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jnsvj\" (UniqueName: \"kubernetes.io/projected/183d2a7b-cd4e-423c-82c0-a72e2ebec301-kube-api-access-jnsvj\") pod \"tempest-tests-tempest\" (UID: \"183d2a7b-cd4e-423c-82c0-a72e2ebec301\") " pod="openstack/tempest-tests-tempest" Nov 28 14:08:15 crc kubenswrapper[4631]: I1128 14:08:15.127089 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"tempest-tests-tempest\" (UID: \"183d2a7b-cd4e-423c-82c0-a72e2ebec301\") " pod="openstack/tempest-tests-tempest" Nov 28 14:08:15 crc kubenswrapper[4631]: I1128 14:08:15.127108 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/183d2a7b-cd4e-423c-82c0-a72e2ebec301-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"183d2a7b-cd4e-423c-82c0-a72e2ebec301\") " pod="openstack/tempest-tests-tempest" Nov 28 14:08:15 crc kubenswrapper[4631]: I1128 14:08:15.127123 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/183d2a7b-cd4e-423c-82c0-a72e2ebec301-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"183d2a7b-cd4e-423c-82c0-a72e2ebec301\") " pod="openstack/tempest-tests-tempest" Nov 28 14:08:15 crc kubenswrapper[4631]: I1128 14:08:15.127172 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/183d2a7b-cd4e-423c-82c0-a72e2ebec301-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"183d2a7b-cd4e-423c-82c0-a72e2ebec301\") " pod="openstack/tempest-tests-tempest" Nov 28 14:08:15 crc kubenswrapper[4631]: I1128 14:08:15.127199 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/183d2a7b-cd4e-423c-82c0-a72e2ebec301-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"183d2a7b-cd4e-423c-82c0-a72e2ebec301\") " pod="openstack/tempest-tests-tempest" Nov 28 14:08:15 crc kubenswrapper[4631]: I1128 14:08:15.127223 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/183d2a7b-cd4e-423c-82c0-a72e2ebec301-config-data\") pod \"tempest-tests-tempest\" (UID: \"183d2a7b-cd4e-423c-82c0-a72e2ebec301\") " pod="openstack/tempest-tests-tempest" Nov 28 14:08:15 crc kubenswrapper[4631]: I1128 14:08:15.229273 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"tempest-tests-tempest\" (UID: \"183d2a7b-cd4e-423c-82c0-a72e2ebec301\") " pod="openstack/tempest-tests-tempest" Nov 28 14:08:15 crc kubenswrapper[4631]: I1128 14:08:15.229342 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/183d2a7b-cd4e-423c-82c0-a72e2ebec301-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"183d2a7b-cd4e-423c-82c0-a72e2ebec301\") " pod="openstack/tempest-tests-tempest" Nov 28 14:08:15 crc kubenswrapper[4631]: I1128 14:08:15.229367 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/183d2a7b-cd4e-423c-82c0-a72e2ebec301-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"183d2a7b-cd4e-423c-82c0-a72e2ebec301\") " pod="openstack/tempest-tests-tempest" Nov 28 14:08:15 crc kubenswrapper[4631]: I1128 14:08:15.229439 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/183d2a7b-cd4e-423c-82c0-a72e2ebec301-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"183d2a7b-cd4e-423c-82c0-a72e2ebec301\") " pod="openstack/tempest-tests-tempest" Nov 28 14:08:15 crc kubenswrapper[4631]: I1128 14:08:15.229477 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/183d2a7b-cd4e-423c-82c0-a72e2ebec301-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"183d2a7b-cd4e-423c-82c0-a72e2ebec301\") " pod="openstack/tempest-tests-tempest" Nov 28 14:08:15 crc kubenswrapper[4631]: I1128 14:08:15.229509 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/183d2a7b-cd4e-423c-82c0-a72e2ebec301-config-data\") pod \"tempest-tests-tempest\" (UID: \"183d2a7b-cd4e-423c-82c0-a72e2ebec301\") " pod="openstack/tempest-tests-tempest" Nov 28 14:08:15 crc kubenswrapper[4631]: I1128 14:08:15.229552 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/183d2a7b-cd4e-423c-82c0-a72e2ebec301-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"183d2a7b-cd4e-423c-82c0-a72e2ebec301\") " pod="openstack/tempest-tests-tempest" Nov 28 14:08:15 crc kubenswrapper[4631]: I1128 14:08:15.229605 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/183d2a7b-cd4e-423c-82c0-a72e2ebec301-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"183d2a7b-cd4e-423c-82c0-a72e2ebec301\") " pod="openstack/tempest-tests-tempest" Nov 28 14:08:15 crc kubenswrapper[4631]: I1128 14:08:15.229636 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jnsvj\" (UniqueName: \"kubernetes.io/projected/183d2a7b-cd4e-423c-82c0-a72e2ebec301-kube-api-access-jnsvj\") pod \"tempest-tests-tempest\" (UID: \"183d2a7b-cd4e-423c-82c0-a72e2ebec301\") " pod="openstack/tempest-tests-tempest" Nov 28 14:08:15 crc kubenswrapper[4631]: I1128 14:08:15.229649 4631 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"tempest-tests-tempest\" (UID: \"183d2a7b-cd4e-423c-82c0-a72e2ebec301\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/tempest-tests-tempest" Nov 28 14:08:15 crc kubenswrapper[4631]: I1128 14:08:15.230923 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/183d2a7b-cd4e-423c-82c0-a72e2ebec301-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"183d2a7b-cd4e-423c-82c0-a72e2ebec301\") " pod="openstack/tempest-tests-tempest" Nov 28 14:08:15 crc kubenswrapper[4631]: I1128 14:08:15.231033 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/183d2a7b-cd4e-423c-82c0-a72e2ebec301-config-data\") pod \"tempest-tests-tempest\" (UID: \"183d2a7b-cd4e-423c-82c0-a72e2ebec301\") " pod="openstack/tempest-tests-tempest" Nov 28 14:08:15 crc kubenswrapper[4631]: I1128 14:08:15.231396 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/183d2a7b-cd4e-423c-82c0-a72e2ebec301-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"183d2a7b-cd4e-423c-82c0-a72e2ebec301\") " pod="openstack/tempest-tests-tempest" Nov 28 14:08:15 crc kubenswrapper[4631]: I1128 14:08:15.231834 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/183d2a7b-cd4e-423c-82c0-a72e2ebec301-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"183d2a7b-cd4e-423c-82c0-a72e2ebec301\") " pod="openstack/tempest-tests-tempest" Nov 28 14:08:15 crc kubenswrapper[4631]: I1128 14:08:15.240044 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/183d2a7b-cd4e-423c-82c0-a72e2ebec301-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"183d2a7b-cd4e-423c-82c0-a72e2ebec301\") " pod="openstack/tempest-tests-tempest" Nov 28 14:08:15 crc kubenswrapper[4631]: I1128 14:08:15.240423 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/183d2a7b-cd4e-423c-82c0-a72e2ebec301-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"183d2a7b-cd4e-423c-82c0-a72e2ebec301\") " pod="openstack/tempest-tests-tempest" Nov 28 14:08:15 crc kubenswrapper[4631]: I1128 14:08:15.240464 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/183d2a7b-cd4e-423c-82c0-a72e2ebec301-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"183d2a7b-cd4e-423c-82c0-a72e2ebec301\") " pod="openstack/tempest-tests-tempest" Nov 28 14:08:15 crc kubenswrapper[4631]: I1128 14:08:15.249664 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jnsvj\" (UniqueName: \"kubernetes.io/projected/183d2a7b-cd4e-423c-82c0-a72e2ebec301-kube-api-access-jnsvj\") pod \"tempest-tests-tempest\" (UID: \"183d2a7b-cd4e-423c-82c0-a72e2ebec301\") " pod="openstack/tempest-tests-tempest" Nov 28 14:08:15 crc kubenswrapper[4631]: I1128 14:08:15.260461 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"tempest-tests-tempest\" (UID: \"183d2a7b-cd4e-423c-82c0-a72e2ebec301\") " pod="openstack/tempest-tests-tempest" Nov 28 14:08:15 crc kubenswrapper[4631]: I1128 14:08:15.562937 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Nov 28 14:08:16 crc kubenswrapper[4631]: I1128 14:08:16.122048 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Nov 28 14:08:16 crc kubenswrapper[4631]: I1128 14:08:16.139812 4631 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 28 14:08:16 crc kubenswrapper[4631]: I1128 14:08:16.533463 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"183d2a7b-cd4e-423c-82c0-a72e2ebec301","Type":"ContainerStarted","Data":"dd2f2386d84ab97e49a9ea37a9a080e34a266c70a3359e5e32353fdb48b712e2"} Nov 28 14:08:55 crc kubenswrapper[4631]: E1128 14:08:55.838670 4631 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified" Nov 28 14:08:55 crc kubenswrapper[4631]: E1128 14:08:55.841248 4631 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:tempest-tests-tempest-tests-runner,Image:quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:test-operator-ephemeral-workdir,ReadOnly:false,MountPath:/var/lib/tempest,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-temporary,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:false,MountPath:/etc/test_operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-logs,ReadOnly:false,MountPath:/var/lib/tempest/external_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/etc/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/var/lib/tempest/.config/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config-secret,ReadOnly:false,MountPath:/etc/openstack/secure.yaml,SubPath:secure.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ca-certs,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ssh-key,ReadOnly:false,MountPath:/var/lib/tempest/id_ecdsa,SubPath:ssh_key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-jnsvj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42480,RunAsNonRoot:*false,ReadOnlyRootFilesystem:*false,AllowPrivilegeEscalation:*true,RunAsGroup:*42480,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-custom-data-s0,},Optional:nil,},SecretRef:nil,},EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-env-vars-s0,},Optional:nil,},SecretRef:nil,},},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod tempest-tests-tempest_openstack(183d2a7b-cd4e-423c-82c0-a72e2ebec301): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 28 14:08:55 crc kubenswrapper[4631]: E1128 14:08:55.842623 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/tempest-tests-tempest" podUID="183d2a7b-cd4e-423c-82c0-a72e2ebec301" Nov 28 14:08:55 crc kubenswrapper[4631]: E1128 14:08:55.934055 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified\\\"\"" pod="openstack/tempest-tests-tempest" podUID="183d2a7b-cd4e-423c-82c0-a72e2ebec301" Nov 28 14:09:07 crc kubenswrapper[4631]: I1128 14:09:07.425834 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Nov 28 14:09:09 crc kubenswrapper[4631]: I1128 14:09:09.059394 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"183d2a7b-cd4e-423c-82c0-a72e2ebec301","Type":"ContainerStarted","Data":"a9c0440be1ccf0bdf15da7e6f8fb2dfc5bb1144d0074dcdc8d85c8bf7f03545d"} Nov 28 14:09:09 crc kubenswrapper[4631]: I1128 14:09:09.083231 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/tempest-tests-tempest" podStartSLOduration=4.800146837 podStartE2EDuration="56.083209072s" podCreationTimestamp="2025-11-28 14:08:13 +0000 UTC" firstStartedPulling="2025-11-28 14:08:16.139601843 +0000 UTC m=+2852.946905187" lastFinishedPulling="2025-11-28 14:09:07.422664078 +0000 UTC m=+2904.229967422" observedRunningTime="2025-11-28 14:09:09.0789956 +0000 UTC m=+2905.886298944" watchObservedRunningTime="2025-11-28 14:09:09.083209072 +0000 UTC m=+2905.890512416" Nov 28 14:10:19 crc kubenswrapper[4631]: I1128 14:10:19.634371 4631 patch_prober.go:28] interesting pod/machine-config-daemon-47sxw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 28 14:10:19 crc kubenswrapper[4631]: I1128 14:10:19.635077 4631 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" podUID="8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 28 14:10:49 crc kubenswrapper[4631]: I1128 14:10:49.635279 4631 patch_prober.go:28] interesting pod/machine-config-daemon-47sxw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 28 14:10:49 crc kubenswrapper[4631]: I1128 14:10:49.635918 4631 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" podUID="8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 28 14:11:19 crc kubenswrapper[4631]: I1128 14:11:19.635616 4631 patch_prober.go:28] interesting pod/machine-config-daemon-47sxw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 28 14:11:19 crc kubenswrapper[4631]: I1128 14:11:19.636169 4631 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" podUID="8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 28 14:11:19 crc kubenswrapper[4631]: I1128 14:11:19.636218 4631 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" Nov 28 14:11:19 crc kubenswrapper[4631]: I1128 14:11:19.637265 4631 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"d188114d99f65821ef24c4d713f07ed3cbe0cd10a5aaa23d6346374210035dc4"} pod="openshift-machine-config-operator/machine-config-daemon-47sxw" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 28 14:11:19 crc kubenswrapper[4631]: I1128 14:11:19.637403 4631 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" podUID="8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1" containerName="machine-config-daemon" containerID="cri-o://d188114d99f65821ef24c4d713f07ed3cbe0cd10a5aaa23d6346374210035dc4" gracePeriod=600 Nov 28 14:11:19 crc kubenswrapper[4631]: E1128 14:11:19.777369 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-47sxw_openshift-machine-config-operator(8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1)\"" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" podUID="8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1" Nov 28 14:11:20 crc kubenswrapper[4631]: I1128 14:11:20.262965 4631 generic.go:334] "Generic (PLEG): container finished" podID="8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1" containerID="d188114d99f65821ef24c4d713f07ed3cbe0cd10a5aaa23d6346374210035dc4" exitCode=0 Nov 28 14:11:20 crc kubenswrapper[4631]: I1128 14:11:20.263016 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" event={"ID":"8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1","Type":"ContainerDied","Data":"d188114d99f65821ef24c4d713f07ed3cbe0cd10a5aaa23d6346374210035dc4"} Nov 28 14:11:20 crc kubenswrapper[4631]: I1128 14:11:20.263074 4631 scope.go:117] "RemoveContainer" containerID="8dbc40040e6ec9ba26b8327ffa3d7f1a4f6a3ca257f8ca30274daa85de63606a" Nov 28 14:11:20 crc kubenswrapper[4631]: I1128 14:11:20.264602 4631 scope.go:117] "RemoveContainer" containerID="d188114d99f65821ef24c4d713f07ed3cbe0cd10a5aaa23d6346374210035dc4" Nov 28 14:11:20 crc kubenswrapper[4631]: E1128 14:11:20.265386 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-47sxw_openshift-machine-config-operator(8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1)\"" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" podUID="8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1" Nov 28 14:11:35 crc kubenswrapper[4631]: I1128 14:11:35.513357 4631 scope.go:117] "RemoveContainer" containerID="d188114d99f65821ef24c4d713f07ed3cbe0cd10a5aaa23d6346374210035dc4" Nov 28 14:11:35 crc kubenswrapper[4631]: E1128 14:11:35.514032 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-47sxw_openshift-machine-config-operator(8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1)\"" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" podUID="8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1" Nov 28 14:11:36 crc kubenswrapper[4631]: I1128 14:11:36.148771 4631 generic.go:334] "Generic (PLEG): container finished" podID="183d2a7b-cd4e-423c-82c0-a72e2ebec301" containerID="a9c0440be1ccf0bdf15da7e6f8fb2dfc5bb1144d0074dcdc8d85c8bf7f03545d" exitCode=0 Nov 28 14:11:36 crc kubenswrapper[4631]: I1128 14:11:36.148854 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"183d2a7b-cd4e-423c-82c0-a72e2ebec301","Type":"ContainerDied","Data":"a9c0440be1ccf0bdf15da7e6f8fb2dfc5bb1144d0074dcdc8d85c8bf7f03545d"} Nov 28 14:11:37 crc kubenswrapper[4631]: I1128 14:11:37.652761 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Nov 28 14:11:37 crc kubenswrapper[4631]: I1128 14:11:37.828329 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/183d2a7b-cd4e-423c-82c0-a72e2ebec301-ssh-key\") pod \"183d2a7b-cd4e-423c-82c0-a72e2ebec301\" (UID: \"183d2a7b-cd4e-423c-82c0-a72e2ebec301\") " Nov 28 14:11:37 crc kubenswrapper[4631]: I1128 14:11:37.828382 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/183d2a7b-cd4e-423c-82c0-a72e2ebec301-test-operator-ephemeral-temporary\") pod \"183d2a7b-cd4e-423c-82c0-a72e2ebec301\" (UID: \"183d2a7b-cd4e-423c-82c0-a72e2ebec301\") " Nov 28 14:11:37 crc kubenswrapper[4631]: I1128 14:11:37.828441 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/183d2a7b-cd4e-423c-82c0-a72e2ebec301-openstack-config\") pod \"183d2a7b-cd4e-423c-82c0-a72e2ebec301\" (UID: \"183d2a7b-cd4e-423c-82c0-a72e2ebec301\") " Nov 28 14:11:37 crc kubenswrapper[4631]: I1128 14:11:37.828479 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/183d2a7b-cd4e-423c-82c0-a72e2ebec301-config-data\") pod \"183d2a7b-cd4e-423c-82c0-a72e2ebec301\" (UID: \"183d2a7b-cd4e-423c-82c0-a72e2ebec301\") " Nov 28 14:11:37 crc kubenswrapper[4631]: I1128 14:11:37.828568 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/183d2a7b-cd4e-423c-82c0-a72e2ebec301-ca-certs\") pod \"183d2a7b-cd4e-423c-82c0-a72e2ebec301\" (UID: \"183d2a7b-cd4e-423c-82c0-a72e2ebec301\") " Nov 28 14:11:37 crc kubenswrapper[4631]: I1128 14:11:37.828693 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-logs\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"183d2a7b-cd4e-423c-82c0-a72e2ebec301\" (UID: \"183d2a7b-cd4e-423c-82c0-a72e2ebec301\") " Nov 28 14:11:37 crc kubenswrapper[4631]: I1128 14:11:37.828740 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/183d2a7b-cd4e-423c-82c0-a72e2ebec301-test-operator-ephemeral-workdir\") pod \"183d2a7b-cd4e-423c-82c0-a72e2ebec301\" (UID: \"183d2a7b-cd4e-423c-82c0-a72e2ebec301\") " Nov 28 14:11:37 crc kubenswrapper[4631]: I1128 14:11:37.828773 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/183d2a7b-cd4e-423c-82c0-a72e2ebec301-openstack-config-secret\") pod \"183d2a7b-cd4e-423c-82c0-a72e2ebec301\" (UID: \"183d2a7b-cd4e-423c-82c0-a72e2ebec301\") " Nov 28 14:11:37 crc kubenswrapper[4631]: I1128 14:11:37.828792 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jnsvj\" (UniqueName: \"kubernetes.io/projected/183d2a7b-cd4e-423c-82c0-a72e2ebec301-kube-api-access-jnsvj\") pod \"183d2a7b-cd4e-423c-82c0-a72e2ebec301\" (UID: \"183d2a7b-cd4e-423c-82c0-a72e2ebec301\") " Nov 28 14:11:37 crc kubenswrapper[4631]: I1128 14:11:37.829253 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/183d2a7b-cd4e-423c-82c0-a72e2ebec301-config-data" (OuterVolumeSpecName: "config-data") pod "183d2a7b-cd4e-423c-82c0-a72e2ebec301" (UID: "183d2a7b-cd4e-423c-82c0-a72e2ebec301"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:11:37 crc kubenswrapper[4631]: I1128 14:11:37.829946 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/183d2a7b-cd4e-423c-82c0-a72e2ebec301-test-operator-ephemeral-temporary" (OuterVolumeSpecName: "test-operator-ephemeral-temporary") pod "183d2a7b-cd4e-423c-82c0-a72e2ebec301" (UID: "183d2a7b-cd4e-423c-82c0-a72e2ebec301"). InnerVolumeSpecName "test-operator-ephemeral-temporary". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 14:11:37 crc kubenswrapper[4631]: I1128 14:11:37.833460 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/183d2a7b-cd4e-423c-82c0-a72e2ebec301-test-operator-ephemeral-workdir" (OuterVolumeSpecName: "test-operator-ephemeral-workdir") pod "183d2a7b-cd4e-423c-82c0-a72e2ebec301" (UID: "183d2a7b-cd4e-423c-82c0-a72e2ebec301"). InnerVolumeSpecName "test-operator-ephemeral-workdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 14:11:37 crc kubenswrapper[4631]: I1128 14:11:37.835182 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage12-crc" (OuterVolumeSpecName: "test-operator-logs") pod "183d2a7b-cd4e-423c-82c0-a72e2ebec301" (UID: "183d2a7b-cd4e-423c-82c0-a72e2ebec301"). InnerVolumeSpecName "local-storage12-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 28 14:11:37 crc kubenswrapper[4631]: I1128 14:11:37.836357 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/183d2a7b-cd4e-423c-82c0-a72e2ebec301-kube-api-access-jnsvj" (OuterVolumeSpecName: "kube-api-access-jnsvj") pod "183d2a7b-cd4e-423c-82c0-a72e2ebec301" (UID: "183d2a7b-cd4e-423c-82c0-a72e2ebec301"). InnerVolumeSpecName "kube-api-access-jnsvj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:11:37 crc kubenswrapper[4631]: I1128 14:11:37.866715 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/183d2a7b-cd4e-423c-82c0-a72e2ebec301-ca-certs" (OuterVolumeSpecName: "ca-certs") pod "183d2a7b-cd4e-423c-82c0-a72e2ebec301" (UID: "183d2a7b-cd4e-423c-82c0-a72e2ebec301"). InnerVolumeSpecName "ca-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:11:37 crc kubenswrapper[4631]: I1128 14:11:37.868521 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/183d2a7b-cd4e-423c-82c0-a72e2ebec301-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "183d2a7b-cd4e-423c-82c0-a72e2ebec301" (UID: "183d2a7b-cd4e-423c-82c0-a72e2ebec301"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:11:37 crc kubenswrapper[4631]: I1128 14:11:37.869294 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/183d2a7b-cd4e-423c-82c0-a72e2ebec301-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "183d2a7b-cd4e-423c-82c0-a72e2ebec301" (UID: "183d2a7b-cd4e-423c-82c0-a72e2ebec301"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:11:37 crc kubenswrapper[4631]: I1128 14:11:37.894782 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/183d2a7b-cd4e-423c-82c0-a72e2ebec301-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "183d2a7b-cd4e-423c-82c0-a72e2ebec301" (UID: "183d2a7b-cd4e-423c-82c0-a72e2ebec301"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:11:37 crc kubenswrapper[4631]: I1128 14:11:37.931058 4631 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/183d2a7b-cd4e-423c-82c0-a72e2ebec301-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 28 14:11:37 crc kubenswrapper[4631]: I1128 14:11:37.931108 4631 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/183d2a7b-cd4e-423c-82c0-a72e2ebec301-test-operator-ephemeral-temporary\") on node \"crc\" DevicePath \"\"" Nov 28 14:11:37 crc kubenswrapper[4631]: I1128 14:11:37.931129 4631 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/183d2a7b-cd4e-423c-82c0-a72e2ebec301-openstack-config\") on node \"crc\" DevicePath \"\"" Nov 28 14:11:37 crc kubenswrapper[4631]: I1128 14:11:37.931143 4631 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/183d2a7b-cd4e-423c-82c0-a72e2ebec301-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 14:11:37 crc kubenswrapper[4631]: I1128 14:11:37.931159 4631 reconciler_common.go:293] "Volume detached for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/183d2a7b-cd4e-423c-82c0-a72e2ebec301-ca-certs\") on node \"crc\" DevicePath \"\"" Nov 28 14:11:37 crc kubenswrapper[4631]: I1128 14:11:37.931207 4631 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" " Nov 28 14:11:37 crc kubenswrapper[4631]: I1128 14:11:37.931224 4631 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/183d2a7b-cd4e-423c-82c0-a72e2ebec301-test-operator-ephemeral-workdir\") on node \"crc\" DevicePath \"\"" Nov 28 14:11:37 crc kubenswrapper[4631]: I1128 14:11:37.931236 4631 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/183d2a7b-cd4e-423c-82c0-a72e2ebec301-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Nov 28 14:11:37 crc kubenswrapper[4631]: I1128 14:11:37.931249 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jnsvj\" (UniqueName: \"kubernetes.io/projected/183d2a7b-cd4e-423c-82c0-a72e2ebec301-kube-api-access-jnsvj\") on node \"crc\" DevicePath \"\"" Nov 28 14:11:37 crc kubenswrapper[4631]: I1128 14:11:37.958991 4631 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage12-crc" (UniqueName: "kubernetes.io/local-volume/local-storage12-crc") on node "crc" Nov 28 14:11:38 crc kubenswrapper[4631]: I1128 14:11:38.033575 4631 reconciler_common.go:293] "Volume detached for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" DevicePath \"\"" Nov 28 14:11:38 crc kubenswrapper[4631]: I1128 14:11:38.171407 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"183d2a7b-cd4e-423c-82c0-a72e2ebec301","Type":"ContainerDied","Data":"dd2f2386d84ab97e49a9ea37a9a080e34a266c70a3359e5e32353fdb48b712e2"} Nov 28 14:11:38 crc kubenswrapper[4631]: I1128 14:11:38.171468 4631 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dd2f2386d84ab97e49a9ea37a9a080e34a266c70a3359e5e32353fdb48b712e2" Nov 28 14:11:38 crc kubenswrapper[4631]: I1128 14:11:38.171491 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Nov 28 14:11:40 crc kubenswrapper[4631]: I1128 14:11:40.775163 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Nov 28 14:11:40 crc kubenswrapper[4631]: E1128 14:11:40.776330 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="183d2a7b-cd4e-423c-82c0-a72e2ebec301" containerName="tempest-tests-tempest-tests-runner" Nov 28 14:11:40 crc kubenswrapper[4631]: I1128 14:11:40.776355 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="183d2a7b-cd4e-423c-82c0-a72e2ebec301" containerName="tempest-tests-tempest-tests-runner" Nov 28 14:11:40 crc kubenswrapper[4631]: I1128 14:11:40.776644 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="183d2a7b-cd4e-423c-82c0-a72e2ebec301" containerName="tempest-tests-tempest-tests-runner" Nov 28 14:11:40 crc kubenswrapper[4631]: I1128 14:11:40.777685 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 28 14:11:40 crc kubenswrapper[4631]: I1128 14:11:40.780666 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-tqp7g" Nov 28 14:11:40 crc kubenswrapper[4631]: I1128 14:11:40.802515 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Nov 28 14:11:40 crc kubenswrapper[4631]: I1128 14:11:40.892733 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"7fffbcdc-ae30-4841-9161-7a0f06bb7f67\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 28 14:11:40 crc kubenswrapper[4631]: I1128 14:11:40.892888 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nknkz\" (UniqueName: \"kubernetes.io/projected/7fffbcdc-ae30-4841-9161-7a0f06bb7f67-kube-api-access-nknkz\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"7fffbcdc-ae30-4841-9161-7a0f06bb7f67\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 28 14:11:40 crc kubenswrapper[4631]: I1128 14:11:40.994510 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nknkz\" (UniqueName: \"kubernetes.io/projected/7fffbcdc-ae30-4841-9161-7a0f06bb7f67-kube-api-access-nknkz\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"7fffbcdc-ae30-4841-9161-7a0f06bb7f67\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 28 14:11:40 crc kubenswrapper[4631]: I1128 14:11:40.994659 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"7fffbcdc-ae30-4841-9161-7a0f06bb7f67\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 28 14:11:40 crc kubenswrapper[4631]: I1128 14:11:40.995196 4631 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"7fffbcdc-ae30-4841-9161-7a0f06bb7f67\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 28 14:11:41 crc kubenswrapper[4631]: I1128 14:11:41.024219 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"7fffbcdc-ae30-4841-9161-7a0f06bb7f67\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 28 14:11:41 crc kubenswrapper[4631]: I1128 14:11:41.028460 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nknkz\" (UniqueName: \"kubernetes.io/projected/7fffbcdc-ae30-4841-9161-7a0f06bb7f67-kube-api-access-nknkz\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"7fffbcdc-ae30-4841-9161-7a0f06bb7f67\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 28 14:11:41 crc kubenswrapper[4631]: I1128 14:11:41.101147 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 28 14:11:41 crc kubenswrapper[4631]: I1128 14:11:41.579716 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Nov 28 14:11:42 crc kubenswrapper[4631]: I1128 14:11:42.203055 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"7fffbcdc-ae30-4841-9161-7a0f06bb7f67","Type":"ContainerStarted","Data":"4a80105e7675323ad0e0454fe9928944ec6532c51fb8cf3dd9595a5969d14641"} Nov 28 14:11:43 crc kubenswrapper[4631]: I1128 14:11:43.215455 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"7fffbcdc-ae30-4841-9161-7a0f06bb7f67","Type":"ContainerStarted","Data":"217ac1af47e9d96481f4c3f2d2a8787cdc76222b309f52460636d36aae33337d"} Nov 28 14:11:43 crc kubenswrapper[4631]: I1128 14:11:43.246378 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" podStartSLOduration=2.082646385 podStartE2EDuration="3.246355474s" podCreationTimestamp="2025-11-28 14:11:40 +0000 UTC" firstStartedPulling="2025-11-28 14:11:41.593324053 +0000 UTC m=+3058.400627397" lastFinishedPulling="2025-11-28 14:11:42.757033142 +0000 UTC m=+3059.564336486" observedRunningTime="2025-11-28 14:11:43.242548192 +0000 UTC m=+3060.049851536" watchObservedRunningTime="2025-11-28 14:11:43.246355474 +0000 UTC m=+3060.053658818" Nov 28 14:11:49 crc kubenswrapper[4631]: I1128 14:11:49.513583 4631 scope.go:117] "RemoveContainer" containerID="d188114d99f65821ef24c4d713f07ed3cbe0cd10a5aaa23d6346374210035dc4" Nov 28 14:11:49 crc kubenswrapper[4631]: E1128 14:11:49.514257 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-47sxw_openshift-machine-config-operator(8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1)\"" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" podUID="8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1" Nov 28 14:12:01 crc kubenswrapper[4631]: I1128 14:12:01.514320 4631 scope.go:117] "RemoveContainer" containerID="d188114d99f65821ef24c4d713f07ed3cbe0cd10a5aaa23d6346374210035dc4" Nov 28 14:12:01 crc kubenswrapper[4631]: E1128 14:12:01.515022 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-47sxw_openshift-machine-config-operator(8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1)\"" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" podUID="8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1" Nov 28 14:12:06 crc kubenswrapper[4631]: I1128 14:12:06.105052 4631 scope.go:117] "RemoveContainer" containerID="301e3f12f4dcdce552fb86ebd49cdbfcdf4222c879a10248067bc99a01c9201a" Nov 28 14:12:06 crc kubenswrapper[4631]: I1128 14:12:06.160588 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-rrfmx/must-gather-8h6ns"] Nov 28 14:12:06 crc kubenswrapper[4631]: I1128 14:12:06.168837 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-rrfmx/must-gather-8h6ns" Nov 28 14:12:06 crc kubenswrapper[4631]: I1128 14:12:06.173600 4631 scope.go:117] "RemoveContainer" containerID="8845d7819ad2597e97a074dd42206270ef384e7c0a9ecfe361f905d26c5e6719" Nov 28 14:12:06 crc kubenswrapper[4631]: I1128 14:12:06.187757 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-rrfmx"/"kube-root-ca.crt" Nov 28 14:12:06 crc kubenswrapper[4631]: I1128 14:12:06.188080 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-rrfmx"/"openshift-service-ca.crt" Nov 28 14:12:06 crc kubenswrapper[4631]: I1128 14:12:06.189239 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-rrfmx"/"default-dockercfg-r5lrr" Nov 28 14:12:06 crc kubenswrapper[4631]: I1128 14:12:06.241647 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-rrfmx/must-gather-8h6ns"] Nov 28 14:12:06 crc kubenswrapper[4631]: I1128 14:12:06.246932 4631 scope.go:117] "RemoveContainer" containerID="f4d36af863e6f4f984ee7c7ff9abc69f2b236ba1c9f43725261c902c1b20c534" Nov 28 14:12:06 crc kubenswrapper[4631]: I1128 14:12:06.355478 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mh457\" (UniqueName: \"kubernetes.io/projected/5796c446-01db-4372-aabb-dc74c82f687f-kube-api-access-mh457\") pod \"must-gather-8h6ns\" (UID: \"5796c446-01db-4372-aabb-dc74c82f687f\") " pod="openshift-must-gather-rrfmx/must-gather-8h6ns" Nov 28 14:12:06 crc kubenswrapper[4631]: I1128 14:12:06.355560 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/5796c446-01db-4372-aabb-dc74c82f687f-must-gather-output\") pod \"must-gather-8h6ns\" (UID: \"5796c446-01db-4372-aabb-dc74c82f687f\") " pod="openshift-must-gather-rrfmx/must-gather-8h6ns" Nov 28 14:12:06 crc kubenswrapper[4631]: I1128 14:12:06.456969 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/5796c446-01db-4372-aabb-dc74c82f687f-must-gather-output\") pod \"must-gather-8h6ns\" (UID: \"5796c446-01db-4372-aabb-dc74c82f687f\") " pod="openshift-must-gather-rrfmx/must-gather-8h6ns" Nov 28 14:12:06 crc kubenswrapper[4631]: I1128 14:12:06.457129 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mh457\" (UniqueName: \"kubernetes.io/projected/5796c446-01db-4372-aabb-dc74c82f687f-kube-api-access-mh457\") pod \"must-gather-8h6ns\" (UID: \"5796c446-01db-4372-aabb-dc74c82f687f\") " pod="openshift-must-gather-rrfmx/must-gather-8h6ns" Nov 28 14:12:06 crc kubenswrapper[4631]: I1128 14:12:06.458173 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/5796c446-01db-4372-aabb-dc74c82f687f-must-gather-output\") pod \"must-gather-8h6ns\" (UID: \"5796c446-01db-4372-aabb-dc74c82f687f\") " pod="openshift-must-gather-rrfmx/must-gather-8h6ns" Nov 28 14:12:06 crc kubenswrapper[4631]: I1128 14:12:06.479938 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mh457\" (UniqueName: \"kubernetes.io/projected/5796c446-01db-4372-aabb-dc74c82f687f-kube-api-access-mh457\") pod \"must-gather-8h6ns\" (UID: \"5796c446-01db-4372-aabb-dc74c82f687f\") " pod="openshift-must-gather-rrfmx/must-gather-8h6ns" Nov 28 14:12:06 crc kubenswrapper[4631]: I1128 14:12:06.489627 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-rrfmx/must-gather-8h6ns" Nov 28 14:12:07 crc kubenswrapper[4631]: I1128 14:12:07.092899 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-rrfmx/must-gather-8h6ns"] Nov 28 14:12:07 crc kubenswrapper[4631]: I1128 14:12:07.420553 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-rrfmx/must-gather-8h6ns" event={"ID":"5796c446-01db-4372-aabb-dc74c82f687f","Type":"ContainerStarted","Data":"1d683b498d6b4c30b5525beee3f4451811130dca1a7d2d636eb002accdad5476"} Nov 28 14:12:12 crc kubenswrapper[4631]: I1128 14:12:12.470341 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-rrfmx/must-gather-8h6ns" event={"ID":"5796c446-01db-4372-aabb-dc74c82f687f","Type":"ContainerStarted","Data":"cc7e4d2bf296df0e3c26dab3ddf3ac2ff265c052cc5a9fe3a50621fcc68bc68d"} Nov 28 14:12:12 crc kubenswrapper[4631]: I1128 14:12:12.514628 4631 scope.go:117] "RemoveContainer" containerID="d188114d99f65821ef24c4d713f07ed3cbe0cd10a5aaa23d6346374210035dc4" Nov 28 14:12:12 crc kubenswrapper[4631]: E1128 14:12:12.514938 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-47sxw_openshift-machine-config-operator(8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1)\"" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" podUID="8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1" Nov 28 14:12:13 crc kubenswrapper[4631]: I1128 14:12:13.480909 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-rrfmx/must-gather-8h6ns" event={"ID":"5796c446-01db-4372-aabb-dc74c82f687f","Type":"ContainerStarted","Data":"d648d8aa6f3d6d2803aa909de93b07f2613d8f0882da4a7ea4ec2ed32fd50ddc"} Nov 28 14:12:13 crc kubenswrapper[4631]: I1128 14:12:13.502010 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-rrfmx/must-gather-8h6ns" podStartSLOduration=3.441618858 podStartE2EDuration="7.501988621s" podCreationTimestamp="2025-11-28 14:12:06 +0000 UTC" firstStartedPulling="2025-11-28 14:12:07.086139864 +0000 UTC m=+3083.893443208" lastFinishedPulling="2025-11-28 14:12:11.146509627 +0000 UTC m=+3087.953812971" observedRunningTime="2025-11-28 14:12:13.500777321 +0000 UTC m=+3090.308080665" watchObservedRunningTime="2025-11-28 14:12:13.501988621 +0000 UTC m=+3090.309291965" Nov 28 14:12:17 crc kubenswrapper[4631]: I1128 14:12:17.751053 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-rrfmx/crc-debug-twkwf"] Nov 28 14:12:17 crc kubenswrapper[4631]: I1128 14:12:17.752999 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-rrfmx/crc-debug-twkwf" Nov 28 14:12:17 crc kubenswrapper[4631]: I1128 14:12:17.826894 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/4710b62b-e003-4264-ba82-aca9f02a6d6b-host\") pod \"crc-debug-twkwf\" (UID: \"4710b62b-e003-4264-ba82-aca9f02a6d6b\") " pod="openshift-must-gather-rrfmx/crc-debug-twkwf" Nov 28 14:12:17 crc kubenswrapper[4631]: I1128 14:12:17.827032 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gpmbs\" (UniqueName: \"kubernetes.io/projected/4710b62b-e003-4264-ba82-aca9f02a6d6b-kube-api-access-gpmbs\") pod \"crc-debug-twkwf\" (UID: \"4710b62b-e003-4264-ba82-aca9f02a6d6b\") " pod="openshift-must-gather-rrfmx/crc-debug-twkwf" Nov 28 14:12:17 crc kubenswrapper[4631]: I1128 14:12:17.929149 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/4710b62b-e003-4264-ba82-aca9f02a6d6b-host\") pod \"crc-debug-twkwf\" (UID: \"4710b62b-e003-4264-ba82-aca9f02a6d6b\") " pod="openshift-must-gather-rrfmx/crc-debug-twkwf" Nov 28 14:12:17 crc kubenswrapper[4631]: I1128 14:12:17.929305 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gpmbs\" (UniqueName: \"kubernetes.io/projected/4710b62b-e003-4264-ba82-aca9f02a6d6b-kube-api-access-gpmbs\") pod \"crc-debug-twkwf\" (UID: \"4710b62b-e003-4264-ba82-aca9f02a6d6b\") " pod="openshift-must-gather-rrfmx/crc-debug-twkwf" Nov 28 14:12:17 crc kubenswrapper[4631]: I1128 14:12:17.929332 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/4710b62b-e003-4264-ba82-aca9f02a6d6b-host\") pod \"crc-debug-twkwf\" (UID: \"4710b62b-e003-4264-ba82-aca9f02a6d6b\") " pod="openshift-must-gather-rrfmx/crc-debug-twkwf" Nov 28 14:12:17 crc kubenswrapper[4631]: I1128 14:12:17.953046 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gpmbs\" (UniqueName: \"kubernetes.io/projected/4710b62b-e003-4264-ba82-aca9f02a6d6b-kube-api-access-gpmbs\") pod \"crc-debug-twkwf\" (UID: \"4710b62b-e003-4264-ba82-aca9f02a6d6b\") " pod="openshift-must-gather-rrfmx/crc-debug-twkwf" Nov 28 14:12:18 crc kubenswrapper[4631]: I1128 14:12:18.082062 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-rrfmx/crc-debug-twkwf" Nov 28 14:12:18 crc kubenswrapper[4631]: I1128 14:12:18.540554 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-rrfmx/crc-debug-twkwf" event={"ID":"4710b62b-e003-4264-ba82-aca9f02a6d6b","Type":"ContainerStarted","Data":"d546ee28daf2589890cefb5f162842d2b34aa2efa42c67df9421306ad1a0b80e"} Nov 28 14:12:25 crc kubenswrapper[4631]: I1128 14:12:25.513960 4631 scope.go:117] "RemoveContainer" containerID="d188114d99f65821ef24c4d713f07ed3cbe0cd10a5aaa23d6346374210035dc4" Nov 28 14:12:25 crc kubenswrapper[4631]: E1128 14:12:25.515044 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-47sxw_openshift-machine-config-operator(8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1)\"" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" podUID="8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1" Nov 28 14:12:31 crc kubenswrapper[4631]: I1128 14:12:31.917296 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-sp4vr"] Nov 28 14:12:31 crc kubenswrapper[4631]: I1128 14:12:31.921793 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-sp4vr" Nov 28 14:12:31 crc kubenswrapper[4631]: I1128 14:12:31.944918 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-sp4vr"] Nov 28 14:12:32 crc kubenswrapper[4631]: I1128 14:12:32.056445 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8756b05d-b515-4a0b-a823-74b6c68337ce-utilities\") pod \"certified-operators-sp4vr\" (UID: \"8756b05d-b515-4a0b-a823-74b6c68337ce\") " pod="openshift-marketplace/certified-operators-sp4vr" Nov 28 14:12:32 crc kubenswrapper[4631]: I1128 14:12:32.056537 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8756b05d-b515-4a0b-a823-74b6c68337ce-catalog-content\") pod \"certified-operators-sp4vr\" (UID: \"8756b05d-b515-4a0b-a823-74b6c68337ce\") " pod="openshift-marketplace/certified-operators-sp4vr" Nov 28 14:12:32 crc kubenswrapper[4631]: I1128 14:12:32.056635 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t8zfx\" (UniqueName: \"kubernetes.io/projected/8756b05d-b515-4a0b-a823-74b6c68337ce-kube-api-access-t8zfx\") pod \"certified-operators-sp4vr\" (UID: \"8756b05d-b515-4a0b-a823-74b6c68337ce\") " pod="openshift-marketplace/certified-operators-sp4vr" Nov 28 14:12:32 crc kubenswrapper[4631]: I1128 14:12:32.158107 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8756b05d-b515-4a0b-a823-74b6c68337ce-utilities\") pod \"certified-operators-sp4vr\" (UID: \"8756b05d-b515-4a0b-a823-74b6c68337ce\") " pod="openshift-marketplace/certified-operators-sp4vr" Nov 28 14:12:32 crc kubenswrapper[4631]: I1128 14:12:32.158213 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8756b05d-b515-4a0b-a823-74b6c68337ce-catalog-content\") pod \"certified-operators-sp4vr\" (UID: \"8756b05d-b515-4a0b-a823-74b6c68337ce\") " pod="openshift-marketplace/certified-operators-sp4vr" Nov 28 14:12:32 crc kubenswrapper[4631]: I1128 14:12:32.158332 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t8zfx\" (UniqueName: \"kubernetes.io/projected/8756b05d-b515-4a0b-a823-74b6c68337ce-kube-api-access-t8zfx\") pod \"certified-operators-sp4vr\" (UID: \"8756b05d-b515-4a0b-a823-74b6c68337ce\") " pod="openshift-marketplace/certified-operators-sp4vr" Nov 28 14:12:32 crc kubenswrapper[4631]: I1128 14:12:32.158674 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8756b05d-b515-4a0b-a823-74b6c68337ce-utilities\") pod \"certified-operators-sp4vr\" (UID: \"8756b05d-b515-4a0b-a823-74b6c68337ce\") " pod="openshift-marketplace/certified-operators-sp4vr" Nov 28 14:12:32 crc kubenswrapper[4631]: I1128 14:12:32.158812 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8756b05d-b515-4a0b-a823-74b6c68337ce-catalog-content\") pod \"certified-operators-sp4vr\" (UID: \"8756b05d-b515-4a0b-a823-74b6c68337ce\") " pod="openshift-marketplace/certified-operators-sp4vr" Nov 28 14:12:32 crc kubenswrapper[4631]: I1128 14:12:32.179829 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t8zfx\" (UniqueName: \"kubernetes.io/projected/8756b05d-b515-4a0b-a823-74b6c68337ce-kube-api-access-t8zfx\") pod \"certified-operators-sp4vr\" (UID: \"8756b05d-b515-4a0b-a823-74b6c68337ce\") " pod="openshift-marketplace/certified-operators-sp4vr" Nov 28 14:12:32 crc kubenswrapper[4631]: I1128 14:12:32.254082 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-sp4vr" Nov 28 14:12:32 crc kubenswrapper[4631]: I1128 14:12:32.720884 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-rrfmx/crc-debug-twkwf" event={"ID":"4710b62b-e003-4264-ba82-aca9f02a6d6b","Type":"ContainerStarted","Data":"f745700aa1123fc42680159ed9862f7fc833a0aacb2f835a64d4db7f605413ce"} Nov 28 14:12:32 crc kubenswrapper[4631]: I1128 14:12:32.772680 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-rrfmx/crc-debug-twkwf" podStartSLOduration=1.7107478189999998 podStartE2EDuration="15.772660118s" podCreationTimestamp="2025-11-28 14:12:17 +0000 UTC" firstStartedPulling="2025-11-28 14:12:18.14060585 +0000 UTC m=+3094.947909194" lastFinishedPulling="2025-11-28 14:12:32.202518149 +0000 UTC m=+3109.009821493" observedRunningTime="2025-11-28 14:12:32.762700164 +0000 UTC m=+3109.570003508" watchObservedRunningTime="2025-11-28 14:12:32.772660118 +0000 UTC m=+3109.579963462" Nov 28 14:12:32 crc kubenswrapper[4631]: I1128 14:12:32.944137 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-sp4vr"] Nov 28 14:12:32 crc kubenswrapper[4631]: W1128 14:12:32.951980 4631 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8756b05d_b515_4a0b_a823_74b6c68337ce.slice/crio-27460e65bccb7643cf548f5d517ec489c252e3cbfd4f95efbbe39fdc5e3d97f1 WatchSource:0}: Error finding container 27460e65bccb7643cf548f5d517ec489c252e3cbfd4f95efbbe39fdc5e3d97f1: Status 404 returned error can't find the container with id 27460e65bccb7643cf548f5d517ec489c252e3cbfd4f95efbbe39fdc5e3d97f1 Nov 28 14:12:33 crc kubenswrapper[4631]: I1128 14:12:33.730524 4631 generic.go:334] "Generic (PLEG): container finished" podID="8756b05d-b515-4a0b-a823-74b6c68337ce" containerID="5b30b00addd36195acbee6942ef08b367708305ab7e3bdb0a5c5ea5e93ede3cb" exitCode=0 Nov 28 14:12:33 crc kubenswrapper[4631]: I1128 14:12:33.730777 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sp4vr" event={"ID":"8756b05d-b515-4a0b-a823-74b6c68337ce","Type":"ContainerDied","Data":"5b30b00addd36195acbee6942ef08b367708305ab7e3bdb0a5c5ea5e93ede3cb"} Nov 28 14:12:33 crc kubenswrapper[4631]: I1128 14:12:33.732375 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sp4vr" event={"ID":"8756b05d-b515-4a0b-a823-74b6c68337ce","Type":"ContainerStarted","Data":"27460e65bccb7643cf548f5d517ec489c252e3cbfd4f95efbbe39fdc5e3d97f1"} Nov 28 14:12:34 crc kubenswrapper[4631]: I1128 14:12:34.304582 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-5dqv2"] Nov 28 14:12:34 crc kubenswrapper[4631]: I1128 14:12:34.307089 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5dqv2" Nov 28 14:12:34 crc kubenswrapper[4631]: I1128 14:12:34.328878 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-5dqv2"] Nov 28 14:12:34 crc kubenswrapper[4631]: I1128 14:12:34.424665 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pc9df\" (UniqueName: \"kubernetes.io/projected/a5372eee-50fc-4826-9054-ddd95b4ed426-kube-api-access-pc9df\") pod \"redhat-marketplace-5dqv2\" (UID: \"a5372eee-50fc-4826-9054-ddd95b4ed426\") " pod="openshift-marketplace/redhat-marketplace-5dqv2" Nov 28 14:12:34 crc kubenswrapper[4631]: I1128 14:12:34.424859 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a5372eee-50fc-4826-9054-ddd95b4ed426-utilities\") pod \"redhat-marketplace-5dqv2\" (UID: \"a5372eee-50fc-4826-9054-ddd95b4ed426\") " pod="openshift-marketplace/redhat-marketplace-5dqv2" Nov 28 14:12:34 crc kubenswrapper[4631]: I1128 14:12:34.424926 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a5372eee-50fc-4826-9054-ddd95b4ed426-catalog-content\") pod \"redhat-marketplace-5dqv2\" (UID: \"a5372eee-50fc-4826-9054-ddd95b4ed426\") " pod="openshift-marketplace/redhat-marketplace-5dqv2" Nov 28 14:12:34 crc kubenswrapper[4631]: I1128 14:12:34.527648 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pc9df\" (UniqueName: \"kubernetes.io/projected/a5372eee-50fc-4826-9054-ddd95b4ed426-kube-api-access-pc9df\") pod \"redhat-marketplace-5dqv2\" (UID: \"a5372eee-50fc-4826-9054-ddd95b4ed426\") " pod="openshift-marketplace/redhat-marketplace-5dqv2" Nov 28 14:12:34 crc kubenswrapper[4631]: I1128 14:12:34.527865 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a5372eee-50fc-4826-9054-ddd95b4ed426-utilities\") pod \"redhat-marketplace-5dqv2\" (UID: \"a5372eee-50fc-4826-9054-ddd95b4ed426\") " pod="openshift-marketplace/redhat-marketplace-5dqv2" Nov 28 14:12:34 crc kubenswrapper[4631]: I1128 14:12:34.528440 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a5372eee-50fc-4826-9054-ddd95b4ed426-catalog-content\") pod \"redhat-marketplace-5dqv2\" (UID: \"a5372eee-50fc-4826-9054-ddd95b4ed426\") " pod="openshift-marketplace/redhat-marketplace-5dqv2" Nov 28 14:12:34 crc kubenswrapper[4631]: I1128 14:12:34.528619 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a5372eee-50fc-4826-9054-ddd95b4ed426-utilities\") pod \"redhat-marketplace-5dqv2\" (UID: \"a5372eee-50fc-4826-9054-ddd95b4ed426\") " pod="openshift-marketplace/redhat-marketplace-5dqv2" Nov 28 14:12:34 crc kubenswrapper[4631]: I1128 14:12:34.529037 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a5372eee-50fc-4826-9054-ddd95b4ed426-catalog-content\") pod \"redhat-marketplace-5dqv2\" (UID: \"a5372eee-50fc-4826-9054-ddd95b4ed426\") " pod="openshift-marketplace/redhat-marketplace-5dqv2" Nov 28 14:12:34 crc kubenswrapper[4631]: I1128 14:12:34.556913 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pc9df\" (UniqueName: \"kubernetes.io/projected/a5372eee-50fc-4826-9054-ddd95b4ed426-kube-api-access-pc9df\") pod \"redhat-marketplace-5dqv2\" (UID: \"a5372eee-50fc-4826-9054-ddd95b4ed426\") " pod="openshift-marketplace/redhat-marketplace-5dqv2" Nov 28 14:12:34 crc kubenswrapper[4631]: I1128 14:12:34.631421 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5dqv2" Nov 28 14:12:35 crc kubenswrapper[4631]: I1128 14:12:35.318806 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-5dqv2"] Nov 28 14:12:35 crc kubenswrapper[4631]: I1128 14:12:35.757575 4631 generic.go:334] "Generic (PLEG): container finished" podID="a5372eee-50fc-4826-9054-ddd95b4ed426" containerID="c9d1ab70b920fefed3d3887f6768848e237034125e663fbe672553a6ce20a9d2" exitCode=0 Nov 28 14:12:35 crc kubenswrapper[4631]: I1128 14:12:35.757759 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5dqv2" event={"ID":"a5372eee-50fc-4826-9054-ddd95b4ed426","Type":"ContainerDied","Data":"c9d1ab70b920fefed3d3887f6768848e237034125e663fbe672553a6ce20a9d2"} Nov 28 14:12:35 crc kubenswrapper[4631]: I1128 14:12:35.757901 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5dqv2" event={"ID":"a5372eee-50fc-4826-9054-ddd95b4ed426","Type":"ContainerStarted","Data":"7f3500b1258cddce8dc34c797590ee12ab50731ccb9574de98826480feaf0178"} Nov 28 14:12:35 crc kubenswrapper[4631]: I1128 14:12:35.762048 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sp4vr" event={"ID":"8756b05d-b515-4a0b-a823-74b6c68337ce","Type":"ContainerStarted","Data":"8cd23e92bf7b8726ae718ca6fddc15ab0b29d8320244a2b9daebceeb0764d517"} Nov 28 14:12:37 crc kubenswrapper[4631]: I1128 14:12:37.787032 4631 generic.go:334] "Generic (PLEG): container finished" podID="8756b05d-b515-4a0b-a823-74b6c68337ce" containerID="8cd23e92bf7b8726ae718ca6fddc15ab0b29d8320244a2b9daebceeb0764d517" exitCode=0 Nov 28 14:12:37 crc kubenswrapper[4631]: I1128 14:12:37.787240 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sp4vr" event={"ID":"8756b05d-b515-4a0b-a823-74b6c68337ce","Type":"ContainerDied","Data":"8cd23e92bf7b8726ae718ca6fddc15ab0b29d8320244a2b9daebceeb0764d517"} Nov 28 14:12:38 crc kubenswrapper[4631]: I1128 14:12:38.801263 4631 generic.go:334] "Generic (PLEG): container finished" podID="a5372eee-50fc-4826-9054-ddd95b4ed426" containerID="f2a2ea3d1162c39900b83bebf7e8625c87c5679080f92468ee16eb9d80d5da89" exitCode=0 Nov 28 14:12:38 crc kubenswrapper[4631]: I1128 14:12:38.801381 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5dqv2" event={"ID":"a5372eee-50fc-4826-9054-ddd95b4ed426","Type":"ContainerDied","Data":"f2a2ea3d1162c39900b83bebf7e8625c87c5679080f92468ee16eb9d80d5da89"} Nov 28 14:12:39 crc kubenswrapper[4631]: I1128 14:12:39.814638 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5dqv2" event={"ID":"a5372eee-50fc-4826-9054-ddd95b4ed426","Type":"ContainerStarted","Data":"133fee652109f1922b9d4b283f025441889404feb95688bffb7daaa74d116c12"} Nov 28 14:12:39 crc kubenswrapper[4631]: I1128 14:12:39.824498 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sp4vr" event={"ID":"8756b05d-b515-4a0b-a823-74b6c68337ce","Type":"ContainerStarted","Data":"9eda831c0fd2446aacaa9d9f93f297726146b28e6e67e3d8e746230c940ed7f4"} Nov 28 14:12:39 crc kubenswrapper[4631]: I1128 14:12:39.845360 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-5dqv2" podStartSLOduration=2.020953058 podStartE2EDuration="5.845341622s" podCreationTimestamp="2025-11-28 14:12:34 +0000 UTC" firstStartedPulling="2025-11-28 14:12:35.758983064 +0000 UTC m=+3112.566286408" lastFinishedPulling="2025-11-28 14:12:39.583371628 +0000 UTC m=+3116.390674972" observedRunningTime="2025-11-28 14:12:39.838124126 +0000 UTC m=+3116.645427470" watchObservedRunningTime="2025-11-28 14:12:39.845341622 +0000 UTC m=+3116.652644966" Nov 28 14:12:39 crc kubenswrapper[4631]: I1128 14:12:39.913685 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-sp4vr" podStartSLOduration=3.290160496 podStartE2EDuration="8.913652483s" podCreationTimestamp="2025-11-28 14:12:31 +0000 UTC" firstStartedPulling="2025-11-28 14:12:33.734239955 +0000 UTC m=+3110.541543299" lastFinishedPulling="2025-11-28 14:12:39.357731942 +0000 UTC m=+3116.165035286" observedRunningTime="2025-11-28 14:12:39.877421587 +0000 UTC m=+3116.684724941" watchObservedRunningTime="2025-11-28 14:12:39.913652483 +0000 UTC m=+3116.720955847" Nov 28 14:12:40 crc kubenswrapper[4631]: I1128 14:12:40.513500 4631 scope.go:117] "RemoveContainer" containerID="d188114d99f65821ef24c4d713f07ed3cbe0cd10a5aaa23d6346374210035dc4" Nov 28 14:12:40 crc kubenswrapper[4631]: E1128 14:12:40.514391 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-47sxw_openshift-machine-config-operator(8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1)\"" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" podUID="8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1" Nov 28 14:12:42 crc kubenswrapper[4631]: I1128 14:12:42.254956 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-sp4vr" Nov 28 14:12:42 crc kubenswrapper[4631]: I1128 14:12:42.255048 4631 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-sp4vr" Nov 28 14:12:43 crc kubenswrapper[4631]: I1128 14:12:43.317158 4631 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-sp4vr" podUID="8756b05d-b515-4a0b-a823-74b6c68337ce" containerName="registry-server" probeResult="failure" output=< Nov 28 14:12:43 crc kubenswrapper[4631]: timeout: failed to connect service ":50051" within 1s Nov 28 14:12:43 crc kubenswrapper[4631]: > Nov 28 14:12:44 crc kubenswrapper[4631]: I1128 14:12:44.632145 4631 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-5dqv2" Nov 28 14:12:44 crc kubenswrapper[4631]: I1128 14:12:44.633356 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-5dqv2" Nov 28 14:12:44 crc kubenswrapper[4631]: I1128 14:12:44.684946 4631 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-5dqv2" Nov 28 14:12:44 crc kubenswrapper[4631]: I1128 14:12:44.919903 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-5dqv2" Nov 28 14:12:47 crc kubenswrapper[4631]: I1128 14:12:47.894276 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-5dqv2"] Nov 28 14:12:47 crc kubenswrapper[4631]: I1128 14:12:47.908161 4631 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-5dqv2" podUID="a5372eee-50fc-4826-9054-ddd95b4ed426" containerName="registry-server" containerID="cri-o://133fee652109f1922b9d4b283f025441889404feb95688bffb7daaa74d116c12" gracePeriod=2 Nov 28 14:12:48 crc kubenswrapper[4631]: I1128 14:12:48.920554 4631 generic.go:334] "Generic (PLEG): container finished" podID="a5372eee-50fc-4826-9054-ddd95b4ed426" containerID="133fee652109f1922b9d4b283f025441889404feb95688bffb7daaa74d116c12" exitCode=0 Nov 28 14:12:48 crc kubenswrapper[4631]: I1128 14:12:48.920634 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5dqv2" event={"ID":"a5372eee-50fc-4826-9054-ddd95b4ed426","Type":"ContainerDied","Data":"133fee652109f1922b9d4b283f025441889404feb95688bffb7daaa74d116c12"} Nov 28 14:12:48 crc kubenswrapper[4631]: I1128 14:12:48.921672 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5dqv2" event={"ID":"a5372eee-50fc-4826-9054-ddd95b4ed426","Type":"ContainerDied","Data":"7f3500b1258cddce8dc34c797590ee12ab50731ccb9574de98826480feaf0178"} Nov 28 14:12:48 crc kubenswrapper[4631]: I1128 14:12:48.921697 4631 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7f3500b1258cddce8dc34c797590ee12ab50731ccb9574de98826480feaf0178" Nov 28 14:12:49 crc kubenswrapper[4631]: I1128 14:12:49.089456 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5dqv2" Nov 28 14:12:49 crc kubenswrapper[4631]: I1128 14:12:49.170234 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a5372eee-50fc-4826-9054-ddd95b4ed426-utilities\") pod \"a5372eee-50fc-4826-9054-ddd95b4ed426\" (UID: \"a5372eee-50fc-4826-9054-ddd95b4ed426\") " Nov 28 14:12:49 crc kubenswrapper[4631]: I1128 14:12:49.170397 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a5372eee-50fc-4826-9054-ddd95b4ed426-catalog-content\") pod \"a5372eee-50fc-4826-9054-ddd95b4ed426\" (UID: \"a5372eee-50fc-4826-9054-ddd95b4ed426\") " Nov 28 14:12:49 crc kubenswrapper[4631]: I1128 14:12:49.170456 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pc9df\" (UniqueName: \"kubernetes.io/projected/a5372eee-50fc-4826-9054-ddd95b4ed426-kube-api-access-pc9df\") pod \"a5372eee-50fc-4826-9054-ddd95b4ed426\" (UID: \"a5372eee-50fc-4826-9054-ddd95b4ed426\") " Nov 28 14:12:49 crc kubenswrapper[4631]: I1128 14:12:49.171769 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a5372eee-50fc-4826-9054-ddd95b4ed426-utilities" (OuterVolumeSpecName: "utilities") pod "a5372eee-50fc-4826-9054-ddd95b4ed426" (UID: "a5372eee-50fc-4826-9054-ddd95b4ed426"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 14:12:49 crc kubenswrapper[4631]: I1128 14:12:49.184875 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a5372eee-50fc-4826-9054-ddd95b4ed426-kube-api-access-pc9df" (OuterVolumeSpecName: "kube-api-access-pc9df") pod "a5372eee-50fc-4826-9054-ddd95b4ed426" (UID: "a5372eee-50fc-4826-9054-ddd95b4ed426"). InnerVolumeSpecName "kube-api-access-pc9df". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:12:49 crc kubenswrapper[4631]: I1128 14:12:49.197012 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a5372eee-50fc-4826-9054-ddd95b4ed426-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a5372eee-50fc-4826-9054-ddd95b4ed426" (UID: "a5372eee-50fc-4826-9054-ddd95b4ed426"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 14:12:49 crc kubenswrapper[4631]: I1128 14:12:49.272695 4631 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a5372eee-50fc-4826-9054-ddd95b4ed426-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 28 14:12:49 crc kubenswrapper[4631]: I1128 14:12:49.272736 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pc9df\" (UniqueName: \"kubernetes.io/projected/a5372eee-50fc-4826-9054-ddd95b4ed426-kube-api-access-pc9df\") on node \"crc\" DevicePath \"\"" Nov 28 14:12:49 crc kubenswrapper[4631]: I1128 14:12:49.272750 4631 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a5372eee-50fc-4826-9054-ddd95b4ed426-utilities\") on node \"crc\" DevicePath \"\"" Nov 28 14:12:49 crc kubenswrapper[4631]: I1128 14:12:49.930184 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5dqv2" Nov 28 14:12:49 crc kubenswrapper[4631]: I1128 14:12:49.956912 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-5dqv2"] Nov 28 14:12:49 crc kubenswrapper[4631]: I1128 14:12:49.967801 4631 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-5dqv2"] Nov 28 14:12:51 crc kubenswrapper[4631]: I1128 14:12:51.525092 4631 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a5372eee-50fc-4826-9054-ddd95b4ed426" path="/var/lib/kubelet/pods/a5372eee-50fc-4826-9054-ddd95b4ed426/volumes" Nov 28 14:12:52 crc kubenswrapper[4631]: I1128 14:12:52.314234 4631 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-sp4vr" Nov 28 14:12:52 crc kubenswrapper[4631]: I1128 14:12:52.379580 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-sp4vr" Nov 28 14:12:53 crc kubenswrapper[4631]: I1128 14:12:53.488191 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-sp4vr"] Nov 28 14:12:53 crc kubenswrapper[4631]: I1128 14:12:53.967227 4631 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-sp4vr" podUID="8756b05d-b515-4a0b-a823-74b6c68337ce" containerName="registry-server" containerID="cri-o://9eda831c0fd2446aacaa9d9f93f297726146b28e6e67e3d8e746230c940ed7f4" gracePeriod=2 Nov 28 14:12:54 crc kubenswrapper[4631]: I1128 14:12:54.471681 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-sp4vr" Nov 28 14:12:54 crc kubenswrapper[4631]: I1128 14:12:54.585496 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8756b05d-b515-4a0b-a823-74b6c68337ce-catalog-content\") pod \"8756b05d-b515-4a0b-a823-74b6c68337ce\" (UID: \"8756b05d-b515-4a0b-a823-74b6c68337ce\") " Nov 28 14:12:54 crc kubenswrapper[4631]: I1128 14:12:54.585634 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8756b05d-b515-4a0b-a823-74b6c68337ce-utilities\") pod \"8756b05d-b515-4a0b-a823-74b6c68337ce\" (UID: \"8756b05d-b515-4a0b-a823-74b6c68337ce\") " Nov 28 14:12:54 crc kubenswrapper[4631]: I1128 14:12:54.585703 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t8zfx\" (UniqueName: \"kubernetes.io/projected/8756b05d-b515-4a0b-a823-74b6c68337ce-kube-api-access-t8zfx\") pod \"8756b05d-b515-4a0b-a823-74b6c68337ce\" (UID: \"8756b05d-b515-4a0b-a823-74b6c68337ce\") " Nov 28 14:12:54 crc kubenswrapper[4631]: I1128 14:12:54.586489 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8756b05d-b515-4a0b-a823-74b6c68337ce-utilities" (OuterVolumeSpecName: "utilities") pod "8756b05d-b515-4a0b-a823-74b6c68337ce" (UID: "8756b05d-b515-4a0b-a823-74b6c68337ce"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 14:12:54 crc kubenswrapper[4631]: I1128 14:12:54.586599 4631 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8756b05d-b515-4a0b-a823-74b6c68337ce-utilities\") on node \"crc\" DevicePath \"\"" Nov 28 14:12:54 crc kubenswrapper[4631]: I1128 14:12:54.602658 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8756b05d-b515-4a0b-a823-74b6c68337ce-kube-api-access-t8zfx" (OuterVolumeSpecName: "kube-api-access-t8zfx") pod "8756b05d-b515-4a0b-a823-74b6c68337ce" (UID: "8756b05d-b515-4a0b-a823-74b6c68337ce"). InnerVolumeSpecName "kube-api-access-t8zfx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:12:54 crc kubenswrapper[4631]: I1128 14:12:54.641885 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8756b05d-b515-4a0b-a823-74b6c68337ce-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8756b05d-b515-4a0b-a823-74b6c68337ce" (UID: "8756b05d-b515-4a0b-a823-74b6c68337ce"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 14:12:54 crc kubenswrapper[4631]: I1128 14:12:54.688680 4631 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8756b05d-b515-4a0b-a823-74b6c68337ce-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 28 14:12:54 crc kubenswrapper[4631]: I1128 14:12:54.688966 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t8zfx\" (UniqueName: \"kubernetes.io/projected/8756b05d-b515-4a0b-a823-74b6c68337ce-kube-api-access-t8zfx\") on node \"crc\" DevicePath \"\"" Nov 28 14:12:54 crc kubenswrapper[4631]: I1128 14:12:54.979027 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-sp4vr" Nov 28 14:12:54 crc kubenswrapper[4631]: I1128 14:12:54.979108 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sp4vr" event={"ID":"8756b05d-b515-4a0b-a823-74b6c68337ce","Type":"ContainerDied","Data":"9eda831c0fd2446aacaa9d9f93f297726146b28e6e67e3d8e746230c940ed7f4"} Nov 28 14:12:54 crc kubenswrapper[4631]: I1128 14:12:54.979200 4631 scope.go:117] "RemoveContainer" containerID="9eda831c0fd2446aacaa9d9f93f297726146b28e6e67e3d8e746230c940ed7f4" Nov 28 14:12:54 crc kubenswrapper[4631]: I1128 14:12:54.983522 4631 generic.go:334] "Generic (PLEG): container finished" podID="8756b05d-b515-4a0b-a823-74b6c68337ce" containerID="9eda831c0fd2446aacaa9d9f93f297726146b28e6e67e3d8e746230c940ed7f4" exitCode=0 Nov 28 14:12:54 crc kubenswrapper[4631]: I1128 14:12:54.983693 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sp4vr" event={"ID":"8756b05d-b515-4a0b-a823-74b6c68337ce","Type":"ContainerDied","Data":"27460e65bccb7643cf548f5d517ec489c252e3cbfd4f95efbbe39fdc5e3d97f1"} Nov 28 14:12:55 crc kubenswrapper[4631]: I1128 14:12:55.011251 4631 scope.go:117] "RemoveContainer" containerID="8cd23e92bf7b8726ae718ca6fddc15ab0b29d8320244a2b9daebceeb0764d517" Nov 28 14:12:55 crc kubenswrapper[4631]: I1128 14:12:55.037993 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-sp4vr"] Nov 28 14:12:55 crc kubenswrapper[4631]: I1128 14:12:55.044508 4631 scope.go:117] "RemoveContainer" containerID="5b30b00addd36195acbee6942ef08b367708305ab7e3bdb0a5c5ea5e93ede3cb" Nov 28 14:12:55 crc kubenswrapper[4631]: I1128 14:12:55.048735 4631 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-sp4vr"] Nov 28 14:12:55 crc kubenswrapper[4631]: I1128 14:12:55.094177 4631 scope.go:117] "RemoveContainer" containerID="9eda831c0fd2446aacaa9d9f93f297726146b28e6e67e3d8e746230c940ed7f4" Nov 28 14:12:55 crc kubenswrapper[4631]: E1128 14:12:55.094781 4631 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9eda831c0fd2446aacaa9d9f93f297726146b28e6e67e3d8e746230c940ed7f4\": container with ID starting with 9eda831c0fd2446aacaa9d9f93f297726146b28e6e67e3d8e746230c940ed7f4 not found: ID does not exist" containerID="9eda831c0fd2446aacaa9d9f93f297726146b28e6e67e3d8e746230c940ed7f4" Nov 28 14:12:55 crc kubenswrapper[4631]: I1128 14:12:55.094818 4631 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9eda831c0fd2446aacaa9d9f93f297726146b28e6e67e3d8e746230c940ed7f4"} err="failed to get container status \"9eda831c0fd2446aacaa9d9f93f297726146b28e6e67e3d8e746230c940ed7f4\": rpc error: code = NotFound desc = could not find container \"9eda831c0fd2446aacaa9d9f93f297726146b28e6e67e3d8e746230c940ed7f4\": container with ID starting with 9eda831c0fd2446aacaa9d9f93f297726146b28e6e67e3d8e746230c940ed7f4 not found: ID does not exist" Nov 28 14:12:55 crc kubenswrapper[4631]: I1128 14:12:55.094839 4631 scope.go:117] "RemoveContainer" containerID="8cd23e92bf7b8726ae718ca6fddc15ab0b29d8320244a2b9daebceeb0764d517" Nov 28 14:12:55 crc kubenswrapper[4631]: E1128 14:12:55.095166 4631 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8cd23e92bf7b8726ae718ca6fddc15ab0b29d8320244a2b9daebceeb0764d517\": container with ID starting with 8cd23e92bf7b8726ae718ca6fddc15ab0b29d8320244a2b9daebceeb0764d517 not found: ID does not exist" containerID="8cd23e92bf7b8726ae718ca6fddc15ab0b29d8320244a2b9daebceeb0764d517" Nov 28 14:12:55 crc kubenswrapper[4631]: I1128 14:12:55.095192 4631 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8cd23e92bf7b8726ae718ca6fddc15ab0b29d8320244a2b9daebceeb0764d517"} err="failed to get container status \"8cd23e92bf7b8726ae718ca6fddc15ab0b29d8320244a2b9daebceeb0764d517\": rpc error: code = NotFound desc = could not find container \"8cd23e92bf7b8726ae718ca6fddc15ab0b29d8320244a2b9daebceeb0764d517\": container with ID starting with 8cd23e92bf7b8726ae718ca6fddc15ab0b29d8320244a2b9daebceeb0764d517 not found: ID does not exist" Nov 28 14:12:55 crc kubenswrapper[4631]: I1128 14:12:55.095209 4631 scope.go:117] "RemoveContainer" containerID="5b30b00addd36195acbee6942ef08b367708305ab7e3bdb0a5c5ea5e93ede3cb" Nov 28 14:12:55 crc kubenswrapper[4631]: E1128 14:12:55.096449 4631 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5b30b00addd36195acbee6942ef08b367708305ab7e3bdb0a5c5ea5e93ede3cb\": container with ID starting with 5b30b00addd36195acbee6942ef08b367708305ab7e3bdb0a5c5ea5e93ede3cb not found: ID does not exist" containerID="5b30b00addd36195acbee6942ef08b367708305ab7e3bdb0a5c5ea5e93ede3cb" Nov 28 14:12:55 crc kubenswrapper[4631]: I1128 14:12:55.096470 4631 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5b30b00addd36195acbee6942ef08b367708305ab7e3bdb0a5c5ea5e93ede3cb"} err="failed to get container status \"5b30b00addd36195acbee6942ef08b367708305ab7e3bdb0a5c5ea5e93ede3cb\": rpc error: code = NotFound desc = could not find container \"5b30b00addd36195acbee6942ef08b367708305ab7e3bdb0a5c5ea5e93ede3cb\": container with ID starting with 5b30b00addd36195acbee6942ef08b367708305ab7e3bdb0a5c5ea5e93ede3cb not found: ID does not exist" Nov 28 14:12:55 crc kubenswrapper[4631]: I1128 14:12:55.524599 4631 scope.go:117] "RemoveContainer" containerID="d188114d99f65821ef24c4d713f07ed3cbe0cd10a5aaa23d6346374210035dc4" Nov 28 14:12:55 crc kubenswrapper[4631]: E1128 14:12:55.525049 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-47sxw_openshift-machine-config-operator(8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1)\"" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" podUID="8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1" Nov 28 14:12:55 crc kubenswrapper[4631]: I1128 14:12:55.543934 4631 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8756b05d-b515-4a0b-a823-74b6c68337ce" path="/var/lib/kubelet/pods/8756b05d-b515-4a0b-a823-74b6c68337ce/volumes" Nov 28 14:13:06 crc kubenswrapper[4631]: I1128 14:13:06.513669 4631 scope.go:117] "RemoveContainer" containerID="d188114d99f65821ef24c4d713f07ed3cbe0cd10a5aaa23d6346374210035dc4" Nov 28 14:13:06 crc kubenswrapper[4631]: E1128 14:13:06.514444 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-47sxw_openshift-machine-config-operator(8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1)\"" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" podUID="8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1" Nov 28 14:13:16 crc kubenswrapper[4631]: I1128 14:13:16.192456 4631 generic.go:334] "Generic (PLEG): container finished" podID="4710b62b-e003-4264-ba82-aca9f02a6d6b" containerID="f745700aa1123fc42680159ed9862f7fc833a0aacb2f835a64d4db7f605413ce" exitCode=0 Nov 28 14:13:16 crc kubenswrapper[4631]: I1128 14:13:16.192617 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-rrfmx/crc-debug-twkwf" event={"ID":"4710b62b-e003-4264-ba82-aca9f02a6d6b","Type":"ContainerDied","Data":"f745700aa1123fc42680159ed9862f7fc833a0aacb2f835a64d4db7f605413ce"} Nov 28 14:13:17 crc kubenswrapper[4631]: I1128 14:13:17.332657 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-rrfmx/crc-debug-twkwf" Nov 28 14:13:17 crc kubenswrapper[4631]: I1128 14:13:17.372098 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-rrfmx/crc-debug-twkwf"] Nov 28 14:13:17 crc kubenswrapper[4631]: I1128 14:13:17.384428 4631 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-rrfmx/crc-debug-twkwf"] Nov 28 14:13:17 crc kubenswrapper[4631]: I1128 14:13:17.496208 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/4710b62b-e003-4264-ba82-aca9f02a6d6b-host\") pod \"4710b62b-e003-4264-ba82-aca9f02a6d6b\" (UID: \"4710b62b-e003-4264-ba82-aca9f02a6d6b\") " Nov 28 14:13:17 crc kubenswrapper[4631]: I1128 14:13:17.496423 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gpmbs\" (UniqueName: \"kubernetes.io/projected/4710b62b-e003-4264-ba82-aca9f02a6d6b-kube-api-access-gpmbs\") pod \"4710b62b-e003-4264-ba82-aca9f02a6d6b\" (UID: \"4710b62b-e003-4264-ba82-aca9f02a6d6b\") " Nov 28 14:13:17 crc kubenswrapper[4631]: I1128 14:13:17.496648 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4710b62b-e003-4264-ba82-aca9f02a6d6b-host" (OuterVolumeSpecName: "host") pod "4710b62b-e003-4264-ba82-aca9f02a6d6b" (UID: "4710b62b-e003-4264-ba82-aca9f02a6d6b"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 14:13:17 crc kubenswrapper[4631]: I1128 14:13:17.496837 4631 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/4710b62b-e003-4264-ba82-aca9f02a6d6b-host\") on node \"crc\" DevicePath \"\"" Nov 28 14:13:17 crc kubenswrapper[4631]: I1128 14:13:17.504484 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4710b62b-e003-4264-ba82-aca9f02a6d6b-kube-api-access-gpmbs" (OuterVolumeSpecName: "kube-api-access-gpmbs") pod "4710b62b-e003-4264-ba82-aca9f02a6d6b" (UID: "4710b62b-e003-4264-ba82-aca9f02a6d6b"). InnerVolumeSpecName "kube-api-access-gpmbs". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:13:17 crc kubenswrapper[4631]: I1128 14:13:17.515054 4631 scope.go:117] "RemoveContainer" containerID="d188114d99f65821ef24c4d713f07ed3cbe0cd10a5aaa23d6346374210035dc4" Nov 28 14:13:17 crc kubenswrapper[4631]: E1128 14:13:17.515371 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-47sxw_openshift-machine-config-operator(8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1)\"" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" podUID="8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1" Nov 28 14:13:17 crc kubenswrapper[4631]: I1128 14:13:17.525419 4631 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4710b62b-e003-4264-ba82-aca9f02a6d6b" path="/var/lib/kubelet/pods/4710b62b-e003-4264-ba82-aca9f02a6d6b/volumes" Nov 28 14:13:17 crc kubenswrapper[4631]: I1128 14:13:17.599070 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gpmbs\" (UniqueName: \"kubernetes.io/projected/4710b62b-e003-4264-ba82-aca9f02a6d6b-kube-api-access-gpmbs\") on node \"crc\" DevicePath \"\"" Nov 28 14:13:18 crc kubenswrapper[4631]: I1128 14:13:18.214715 4631 scope.go:117] "RemoveContainer" containerID="f745700aa1123fc42680159ed9862f7fc833a0aacb2f835a64d4db7f605413ce" Nov 28 14:13:18 crc kubenswrapper[4631]: I1128 14:13:18.214950 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-rrfmx/crc-debug-twkwf" Nov 28 14:13:18 crc kubenswrapper[4631]: I1128 14:13:18.601994 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-rrfmx/crc-debug-vrtbd"] Nov 28 14:13:18 crc kubenswrapper[4631]: E1128 14:13:18.604362 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a5372eee-50fc-4826-9054-ddd95b4ed426" containerName="extract-content" Nov 28 14:13:18 crc kubenswrapper[4631]: I1128 14:13:18.604458 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="a5372eee-50fc-4826-9054-ddd95b4ed426" containerName="extract-content" Nov 28 14:13:18 crc kubenswrapper[4631]: E1128 14:13:18.604524 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a5372eee-50fc-4826-9054-ddd95b4ed426" containerName="extract-utilities" Nov 28 14:13:18 crc kubenswrapper[4631]: I1128 14:13:18.604582 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="a5372eee-50fc-4826-9054-ddd95b4ed426" containerName="extract-utilities" Nov 28 14:13:18 crc kubenswrapper[4631]: E1128 14:13:18.604645 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8756b05d-b515-4a0b-a823-74b6c68337ce" containerName="extract-utilities" Nov 28 14:13:18 crc kubenswrapper[4631]: I1128 14:13:18.604712 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="8756b05d-b515-4a0b-a823-74b6c68337ce" containerName="extract-utilities" Nov 28 14:13:18 crc kubenswrapper[4631]: E1128 14:13:18.604807 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4710b62b-e003-4264-ba82-aca9f02a6d6b" containerName="container-00" Nov 28 14:13:18 crc kubenswrapper[4631]: I1128 14:13:18.604878 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="4710b62b-e003-4264-ba82-aca9f02a6d6b" containerName="container-00" Nov 28 14:13:18 crc kubenswrapper[4631]: E1128 14:13:18.604947 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8756b05d-b515-4a0b-a823-74b6c68337ce" containerName="registry-server" Nov 28 14:13:18 crc kubenswrapper[4631]: I1128 14:13:18.605009 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="8756b05d-b515-4a0b-a823-74b6c68337ce" containerName="registry-server" Nov 28 14:13:18 crc kubenswrapper[4631]: E1128 14:13:18.605076 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a5372eee-50fc-4826-9054-ddd95b4ed426" containerName="registry-server" Nov 28 14:13:18 crc kubenswrapper[4631]: I1128 14:13:18.605154 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="a5372eee-50fc-4826-9054-ddd95b4ed426" containerName="registry-server" Nov 28 14:13:18 crc kubenswrapper[4631]: E1128 14:13:18.605237 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8756b05d-b515-4a0b-a823-74b6c68337ce" containerName="extract-content" Nov 28 14:13:18 crc kubenswrapper[4631]: I1128 14:13:18.605317 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="8756b05d-b515-4a0b-a823-74b6c68337ce" containerName="extract-content" Nov 28 14:13:18 crc kubenswrapper[4631]: I1128 14:13:18.605598 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="8756b05d-b515-4a0b-a823-74b6c68337ce" containerName="registry-server" Nov 28 14:13:18 crc kubenswrapper[4631]: I1128 14:13:18.605669 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="a5372eee-50fc-4826-9054-ddd95b4ed426" containerName="registry-server" Nov 28 14:13:18 crc kubenswrapper[4631]: I1128 14:13:18.605739 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="4710b62b-e003-4264-ba82-aca9f02a6d6b" containerName="container-00" Nov 28 14:13:18 crc kubenswrapper[4631]: I1128 14:13:18.606681 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-rrfmx/crc-debug-vrtbd" Nov 28 14:13:18 crc kubenswrapper[4631]: I1128 14:13:18.731787 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ntkxt\" (UniqueName: \"kubernetes.io/projected/7c903619-e943-4d4e-a0ef-5c0fd47233e6-kube-api-access-ntkxt\") pod \"crc-debug-vrtbd\" (UID: \"7c903619-e943-4d4e-a0ef-5c0fd47233e6\") " pod="openshift-must-gather-rrfmx/crc-debug-vrtbd" Nov 28 14:13:18 crc kubenswrapper[4631]: I1128 14:13:18.731997 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/7c903619-e943-4d4e-a0ef-5c0fd47233e6-host\") pod \"crc-debug-vrtbd\" (UID: \"7c903619-e943-4d4e-a0ef-5c0fd47233e6\") " pod="openshift-must-gather-rrfmx/crc-debug-vrtbd" Nov 28 14:13:18 crc kubenswrapper[4631]: I1128 14:13:18.834142 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ntkxt\" (UniqueName: \"kubernetes.io/projected/7c903619-e943-4d4e-a0ef-5c0fd47233e6-kube-api-access-ntkxt\") pod \"crc-debug-vrtbd\" (UID: \"7c903619-e943-4d4e-a0ef-5c0fd47233e6\") " pod="openshift-must-gather-rrfmx/crc-debug-vrtbd" Nov 28 14:13:18 crc kubenswrapper[4631]: I1128 14:13:18.834222 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/7c903619-e943-4d4e-a0ef-5c0fd47233e6-host\") pod \"crc-debug-vrtbd\" (UID: \"7c903619-e943-4d4e-a0ef-5c0fd47233e6\") " pod="openshift-must-gather-rrfmx/crc-debug-vrtbd" Nov 28 14:13:18 crc kubenswrapper[4631]: I1128 14:13:18.834506 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/7c903619-e943-4d4e-a0ef-5c0fd47233e6-host\") pod \"crc-debug-vrtbd\" (UID: \"7c903619-e943-4d4e-a0ef-5c0fd47233e6\") " pod="openshift-must-gather-rrfmx/crc-debug-vrtbd" Nov 28 14:13:18 crc kubenswrapper[4631]: I1128 14:13:18.856191 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ntkxt\" (UniqueName: \"kubernetes.io/projected/7c903619-e943-4d4e-a0ef-5c0fd47233e6-kube-api-access-ntkxt\") pod \"crc-debug-vrtbd\" (UID: \"7c903619-e943-4d4e-a0ef-5c0fd47233e6\") " pod="openshift-must-gather-rrfmx/crc-debug-vrtbd" Nov 28 14:13:18 crc kubenswrapper[4631]: I1128 14:13:18.934836 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-rrfmx/crc-debug-vrtbd" Nov 28 14:13:19 crc kubenswrapper[4631]: I1128 14:13:19.227016 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-rrfmx/crc-debug-vrtbd" event={"ID":"7c903619-e943-4d4e-a0ef-5c0fd47233e6","Type":"ContainerStarted","Data":"ed7262d52dc097387077537ee0d5188810bf05845929f3ad9562cb5c4c58ec99"} Nov 28 14:13:20 crc kubenswrapper[4631]: I1128 14:13:20.244422 4631 generic.go:334] "Generic (PLEG): container finished" podID="7c903619-e943-4d4e-a0ef-5c0fd47233e6" containerID="593af8305f9d618c761232d77a909c162a80291381836ff787eeddd58a6f1a31" exitCode=0 Nov 28 14:13:20 crc kubenswrapper[4631]: I1128 14:13:20.244966 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-rrfmx/crc-debug-vrtbd" event={"ID":"7c903619-e943-4d4e-a0ef-5c0fd47233e6","Type":"ContainerDied","Data":"593af8305f9d618c761232d77a909c162a80291381836ff787eeddd58a6f1a31"} Nov 28 14:13:20 crc kubenswrapper[4631]: I1128 14:13:20.709438 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-rrfmx/crc-debug-vrtbd"] Nov 28 14:13:20 crc kubenswrapper[4631]: I1128 14:13:20.724443 4631 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-rrfmx/crc-debug-vrtbd"] Nov 28 14:13:21 crc kubenswrapper[4631]: I1128 14:13:21.358559 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-rrfmx/crc-debug-vrtbd" Nov 28 14:13:21 crc kubenswrapper[4631]: I1128 14:13:21.487435 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ntkxt\" (UniqueName: \"kubernetes.io/projected/7c903619-e943-4d4e-a0ef-5c0fd47233e6-kube-api-access-ntkxt\") pod \"7c903619-e943-4d4e-a0ef-5c0fd47233e6\" (UID: \"7c903619-e943-4d4e-a0ef-5c0fd47233e6\") " Nov 28 14:13:21 crc kubenswrapper[4631]: I1128 14:13:21.487766 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/7c903619-e943-4d4e-a0ef-5c0fd47233e6-host\") pod \"7c903619-e943-4d4e-a0ef-5c0fd47233e6\" (UID: \"7c903619-e943-4d4e-a0ef-5c0fd47233e6\") " Nov 28 14:13:21 crc kubenswrapper[4631]: I1128 14:13:21.488038 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7c903619-e943-4d4e-a0ef-5c0fd47233e6-host" (OuterVolumeSpecName: "host") pod "7c903619-e943-4d4e-a0ef-5c0fd47233e6" (UID: "7c903619-e943-4d4e-a0ef-5c0fd47233e6"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 14:13:21 crc kubenswrapper[4631]: I1128 14:13:21.488329 4631 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/7c903619-e943-4d4e-a0ef-5c0fd47233e6-host\") on node \"crc\" DevicePath \"\"" Nov 28 14:13:21 crc kubenswrapper[4631]: I1128 14:13:21.774682 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7c903619-e943-4d4e-a0ef-5c0fd47233e6-kube-api-access-ntkxt" (OuterVolumeSpecName: "kube-api-access-ntkxt") pod "7c903619-e943-4d4e-a0ef-5c0fd47233e6" (UID: "7c903619-e943-4d4e-a0ef-5c0fd47233e6"). InnerVolumeSpecName "kube-api-access-ntkxt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:13:21 crc kubenswrapper[4631]: I1128 14:13:21.793276 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ntkxt\" (UniqueName: \"kubernetes.io/projected/7c903619-e943-4d4e-a0ef-5c0fd47233e6-kube-api-access-ntkxt\") on node \"crc\" DevicePath \"\"" Nov 28 14:13:21 crc kubenswrapper[4631]: I1128 14:13:21.881832 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-rrfmx/crc-debug-vfgdh"] Nov 28 14:13:21 crc kubenswrapper[4631]: E1128 14:13:21.882270 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7c903619-e943-4d4e-a0ef-5c0fd47233e6" containerName="container-00" Nov 28 14:13:21 crc kubenswrapper[4631]: I1128 14:13:21.882314 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="7c903619-e943-4d4e-a0ef-5c0fd47233e6" containerName="container-00" Nov 28 14:13:21 crc kubenswrapper[4631]: I1128 14:13:21.882520 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="7c903619-e943-4d4e-a0ef-5c0fd47233e6" containerName="container-00" Nov 28 14:13:21 crc kubenswrapper[4631]: I1128 14:13:21.883196 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-rrfmx/crc-debug-vfgdh" Nov 28 14:13:22 crc kubenswrapper[4631]: I1128 14:13:22.026089 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g6h4w\" (UniqueName: \"kubernetes.io/projected/7c5efc82-a549-4aa1-9a1a-cc24584097d0-kube-api-access-g6h4w\") pod \"crc-debug-vfgdh\" (UID: \"7c5efc82-a549-4aa1-9a1a-cc24584097d0\") " pod="openshift-must-gather-rrfmx/crc-debug-vfgdh" Nov 28 14:13:22 crc kubenswrapper[4631]: I1128 14:13:22.026490 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/7c5efc82-a549-4aa1-9a1a-cc24584097d0-host\") pod \"crc-debug-vfgdh\" (UID: \"7c5efc82-a549-4aa1-9a1a-cc24584097d0\") " pod="openshift-must-gather-rrfmx/crc-debug-vfgdh" Nov 28 14:13:22 crc kubenswrapper[4631]: I1128 14:13:22.128916 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g6h4w\" (UniqueName: \"kubernetes.io/projected/7c5efc82-a549-4aa1-9a1a-cc24584097d0-kube-api-access-g6h4w\") pod \"crc-debug-vfgdh\" (UID: \"7c5efc82-a549-4aa1-9a1a-cc24584097d0\") " pod="openshift-must-gather-rrfmx/crc-debug-vfgdh" Nov 28 14:13:22 crc kubenswrapper[4631]: I1128 14:13:22.129102 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/7c5efc82-a549-4aa1-9a1a-cc24584097d0-host\") pod \"crc-debug-vfgdh\" (UID: \"7c5efc82-a549-4aa1-9a1a-cc24584097d0\") " pod="openshift-must-gather-rrfmx/crc-debug-vfgdh" Nov 28 14:13:22 crc kubenswrapper[4631]: I1128 14:13:22.129239 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/7c5efc82-a549-4aa1-9a1a-cc24584097d0-host\") pod \"crc-debug-vfgdh\" (UID: \"7c5efc82-a549-4aa1-9a1a-cc24584097d0\") " pod="openshift-must-gather-rrfmx/crc-debug-vfgdh" Nov 28 14:13:22 crc kubenswrapper[4631]: I1128 14:13:22.148951 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g6h4w\" (UniqueName: \"kubernetes.io/projected/7c5efc82-a549-4aa1-9a1a-cc24584097d0-kube-api-access-g6h4w\") pod \"crc-debug-vfgdh\" (UID: \"7c5efc82-a549-4aa1-9a1a-cc24584097d0\") " pod="openshift-must-gather-rrfmx/crc-debug-vfgdh" Nov 28 14:13:22 crc kubenswrapper[4631]: I1128 14:13:22.212151 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-rrfmx/crc-debug-vfgdh" Nov 28 14:13:22 crc kubenswrapper[4631]: W1128 14:13:22.266040 4631 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7c5efc82_a549_4aa1_9a1a_cc24584097d0.slice/crio-09b091c831cd12cb6a85d3e4f4513d30d9db34e90c0ca0b0895dbb9b9649ff6e WatchSource:0}: Error finding container 09b091c831cd12cb6a85d3e4f4513d30d9db34e90c0ca0b0895dbb9b9649ff6e: Status 404 returned error can't find the container with id 09b091c831cd12cb6a85d3e4f4513d30d9db34e90c0ca0b0895dbb9b9649ff6e Nov 28 14:13:22 crc kubenswrapper[4631]: I1128 14:13:22.267170 4631 scope.go:117] "RemoveContainer" containerID="593af8305f9d618c761232d77a909c162a80291381836ff787eeddd58a6f1a31" Nov 28 14:13:22 crc kubenswrapper[4631]: I1128 14:13:22.267218 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-rrfmx/crc-debug-vrtbd" Nov 28 14:13:23 crc kubenswrapper[4631]: I1128 14:13:23.280839 4631 generic.go:334] "Generic (PLEG): container finished" podID="7c5efc82-a549-4aa1-9a1a-cc24584097d0" containerID="d362d93ba2837d432d824ec1f840d3c768c52bd3eff4fa08169e0d1427c9d1d0" exitCode=0 Nov 28 14:13:23 crc kubenswrapper[4631]: I1128 14:13:23.281379 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-rrfmx/crc-debug-vfgdh" event={"ID":"7c5efc82-a549-4aa1-9a1a-cc24584097d0","Type":"ContainerDied","Data":"d362d93ba2837d432d824ec1f840d3c768c52bd3eff4fa08169e0d1427c9d1d0"} Nov 28 14:13:23 crc kubenswrapper[4631]: I1128 14:13:23.281497 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-rrfmx/crc-debug-vfgdh" event={"ID":"7c5efc82-a549-4aa1-9a1a-cc24584097d0","Type":"ContainerStarted","Data":"09b091c831cd12cb6a85d3e4f4513d30d9db34e90c0ca0b0895dbb9b9649ff6e"} Nov 28 14:13:23 crc kubenswrapper[4631]: I1128 14:13:23.330185 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-rrfmx/crc-debug-vfgdh"] Nov 28 14:13:23 crc kubenswrapper[4631]: I1128 14:13:23.347456 4631 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-rrfmx/crc-debug-vfgdh"] Nov 28 14:13:23 crc kubenswrapper[4631]: I1128 14:13:23.536163 4631 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7c903619-e943-4d4e-a0ef-5c0fd47233e6" path="/var/lib/kubelet/pods/7c903619-e943-4d4e-a0ef-5c0fd47233e6/volumes" Nov 28 14:13:24 crc kubenswrapper[4631]: I1128 14:13:24.421393 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-rrfmx/crc-debug-vfgdh" Nov 28 14:13:24 crc kubenswrapper[4631]: I1128 14:13:24.579802 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g6h4w\" (UniqueName: \"kubernetes.io/projected/7c5efc82-a549-4aa1-9a1a-cc24584097d0-kube-api-access-g6h4w\") pod \"7c5efc82-a549-4aa1-9a1a-cc24584097d0\" (UID: \"7c5efc82-a549-4aa1-9a1a-cc24584097d0\") " Nov 28 14:13:24 crc kubenswrapper[4631]: I1128 14:13:24.579892 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/7c5efc82-a549-4aa1-9a1a-cc24584097d0-host\") pod \"7c5efc82-a549-4aa1-9a1a-cc24584097d0\" (UID: \"7c5efc82-a549-4aa1-9a1a-cc24584097d0\") " Nov 28 14:13:24 crc kubenswrapper[4631]: I1128 14:13:24.580005 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7c5efc82-a549-4aa1-9a1a-cc24584097d0-host" (OuterVolumeSpecName: "host") pod "7c5efc82-a549-4aa1-9a1a-cc24584097d0" (UID: "7c5efc82-a549-4aa1-9a1a-cc24584097d0"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 14:13:24 crc kubenswrapper[4631]: I1128 14:13:24.580455 4631 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/7c5efc82-a549-4aa1-9a1a-cc24584097d0-host\") on node \"crc\" DevicePath \"\"" Nov 28 14:13:24 crc kubenswrapper[4631]: I1128 14:13:24.594458 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7c5efc82-a549-4aa1-9a1a-cc24584097d0-kube-api-access-g6h4w" (OuterVolumeSpecName: "kube-api-access-g6h4w") pod "7c5efc82-a549-4aa1-9a1a-cc24584097d0" (UID: "7c5efc82-a549-4aa1-9a1a-cc24584097d0"). InnerVolumeSpecName "kube-api-access-g6h4w". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:13:24 crc kubenswrapper[4631]: I1128 14:13:24.683225 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g6h4w\" (UniqueName: \"kubernetes.io/projected/7c5efc82-a549-4aa1-9a1a-cc24584097d0-kube-api-access-g6h4w\") on node \"crc\" DevicePath \"\"" Nov 28 14:13:25 crc kubenswrapper[4631]: I1128 14:13:25.300754 4631 scope.go:117] "RemoveContainer" containerID="d362d93ba2837d432d824ec1f840d3c768c52bd3eff4fa08169e0d1427c9d1d0" Nov 28 14:13:25 crc kubenswrapper[4631]: I1128 14:13:25.300849 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-rrfmx/crc-debug-vfgdh" Nov 28 14:13:25 crc kubenswrapper[4631]: I1128 14:13:25.528579 4631 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7c5efc82-a549-4aa1-9a1a-cc24584097d0" path="/var/lib/kubelet/pods/7c5efc82-a549-4aa1-9a1a-cc24584097d0/volumes" Nov 28 14:13:32 crc kubenswrapper[4631]: I1128 14:13:32.512790 4631 scope.go:117] "RemoveContainer" containerID="d188114d99f65821ef24c4d713f07ed3cbe0cd10a5aaa23d6346374210035dc4" Nov 28 14:13:32 crc kubenswrapper[4631]: E1128 14:13:32.513511 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-47sxw_openshift-machine-config-operator(8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1)\"" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" podUID="8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1" Nov 28 14:13:43 crc kubenswrapper[4631]: I1128 14:13:43.689024 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-54885d4b84-55wlk_1a3cfa34-17ae-42d6-8ea7-46691bb10eaa/barbican-api/0.log" Nov 28 14:13:43 crc kubenswrapper[4631]: I1128 14:13:43.930636 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-54885d4b84-55wlk_1a3cfa34-17ae-42d6-8ea7-46691bb10eaa/barbican-api-log/0.log" Nov 28 14:13:44 crc kubenswrapper[4631]: I1128 14:13:44.114546 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-59c4c4dc74-qt87t_177f3e63-a6ce-4512-80b3-f229e3ace62a/barbican-keystone-listener/0.log" Nov 28 14:13:44 crc kubenswrapper[4631]: I1128 14:13:44.224505 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-59c4c4dc74-qt87t_177f3e63-a6ce-4512-80b3-f229e3ace62a/barbican-keystone-listener-log/0.log" Nov 28 14:13:44 crc kubenswrapper[4631]: I1128 14:13:44.311732 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-5b8f56c4bc-cntw9_1b8b4a3c-2e24-4a9b-b18d-0fa09f70a155/barbican-worker/0.log" Nov 28 14:13:44 crc kubenswrapper[4631]: I1128 14:13:44.464799 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-xz7jm_44ae011d-6420-4400-9c7c-21463e3c82b1/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Nov 28 14:13:44 crc kubenswrapper[4631]: I1128 14:13:44.474607 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-5b8f56c4bc-cntw9_1b8b4a3c-2e24-4a9b-b18d-0fa09f70a155/barbican-worker-log/0.log" Nov 28 14:13:44 crc kubenswrapper[4631]: I1128 14:13:44.773799 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_4cde0c05-4a6a-410c-b5c3-2f14289fc37a/ceilometer-central-agent/0.log" Nov 28 14:13:44 crc kubenswrapper[4631]: I1128 14:13:44.798032 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_4cde0c05-4a6a-410c-b5c3-2f14289fc37a/ceilometer-notification-agent/0.log" Nov 28 14:13:44 crc kubenswrapper[4631]: I1128 14:13:44.831493 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_4cde0c05-4a6a-410c-b5c3-2f14289fc37a/proxy-httpd/0.log" Nov 28 14:13:44 crc kubenswrapper[4631]: I1128 14:13:44.964268 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_4cde0c05-4a6a-410c-b5c3-2f14289fc37a/sg-core/0.log" Nov 28 14:13:45 crc kubenswrapper[4631]: I1128 14:13:45.106575 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_86f5dab6-6b28-494c-b9d6-52528c726875/cinder-api/0.log" Nov 28 14:13:45 crc kubenswrapper[4631]: I1128 14:13:45.180779 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_86f5dab6-6b28-494c-b9d6-52528c726875/cinder-api-log/0.log" Nov 28 14:13:45 crc kubenswrapper[4631]: I1128 14:13:45.307531 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_c4be14a0-f1a1-46f3-86ea-48e7dfb8da00/cinder-scheduler/0.log" Nov 28 14:13:45 crc kubenswrapper[4631]: I1128 14:13:45.363016 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_c4be14a0-f1a1-46f3-86ea-48e7dfb8da00/probe/0.log" Nov 28 14:13:45 crc kubenswrapper[4631]: I1128 14:13:45.517184 4631 scope.go:117] "RemoveContainer" containerID="d188114d99f65821ef24c4d713f07ed3cbe0cd10a5aaa23d6346374210035dc4" Nov 28 14:13:45 crc kubenswrapper[4631]: E1128 14:13:45.520613 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-47sxw_openshift-machine-config-operator(8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1)\"" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" podUID="8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1" Nov 28 14:13:45 crc kubenswrapper[4631]: I1128 14:13:45.942896 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-4lxnq_4d2450a3-e851-4bfb-83c1-115334898784/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Nov 28 14:13:45 crc kubenswrapper[4631]: I1128 14:13:45.974827 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-s9jts_8133b1ca-bdb2-46e1-8d39-2b45ab9284c8/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Nov 28 14:13:46 crc kubenswrapper[4631]: I1128 14:13:46.184112 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-6cd9bffc9-x8stz_6b64d173-0e5b-41e3-b841-68fef937dfb4/init/0.log" Nov 28 14:13:46 crc kubenswrapper[4631]: I1128 14:13:46.408235 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-6cd9bffc9-x8stz_6b64d173-0e5b-41e3-b841-68fef937dfb4/init/0.log" Nov 28 14:13:46 crc kubenswrapper[4631]: I1128 14:13:46.478279 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-6cd9bffc9-x8stz_6b64d173-0e5b-41e3-b841-68fef937dfb4/dnsmasq-dns/0.log" Nov 28 14:13:46 crc kubenswrapper[4631]: I1128 14:13:46.486805 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-2j6xr_2a23172b-a4cc-41e3-9464-2b3e7f7a51f8/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Nov 28 14:13:46 crc kubenswrapper[4631]: I1128 14:13:46.760540 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_aeea572b-1f4e-486b-a916-1b34f0b84d10/glance-log/0.log" Nov 28 14:13:46 crc kubenswrapper[4631]: I1128 14:13:46.770467 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_aeea572b-1f4e-486b-a916-1b34f0b84d10/glance-httpd/0.log" Nov 28 14:13:46 crc kubenswrapper[4631]: I1128 14:13:46.994909 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_b3727bc6-2743-41bf-9a94-45e4921eeb00/glance-httpd/0.log" Nov 28 14:13:47 crc kubenswrapper[4631]: I1128 14:13:47.000973 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_b3727bc6-2743-41bf-9a94-45e4921eeb00/glance-log/0.log" Nov 28 14:13:47 crc kubenswrapper[4631]: I1128 14:13:47.321861 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-544496c768-rzhsw_cd607dae-568a-4b81-af81-3310c0e95854/horizon/2.log" Nov 28 14:13:47 crc kubenswrapper[4631]: I1128 14:13:47.355250 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-544496c768-rzhsw_cd607dae-568a-4b81-af81-3310c0e95854/horizon/1.log" Nov 28 14:13:47 crc kubenswrapper[4631]: I1128 14:13:47.595160 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-544496c768-rzhsw_cd607dae-568a-4b81-af81-3310c0e95854/horizon-log/0.log" Nov 28 14:13:47 crc kubenswrapper[4631]: I1128 14:13:47.619160 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-c6hq5_4f7fda9a-fd1f-4485-a838-f5fee5a83b6e/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Nov 28 14:13:47 crc kubenswrapper[4631]: I1128 14:13:47.970913 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-gsp5r_5f971cb1-7c00-4dbe-abb1-e4eeb825fcee/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Nov 28 14:13:48 crc kubenswrapper[4631]: I1128 14:13:48.110518 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-6bd6cfc878-65bg5_2adae216-4db3-49d7-8d11-24bbd7d47f8f/keystone-api/0.log" Nov 28 14:13:48 crc kubenswrapper[4631]: I1128 14:13:48.296098 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29405641-ngxl6_2080b2bc-31ce-4061-abc1-6f044c10355b/keystone-cron/0.log" Nov 28 14:13:48 crc kubenswrapper[4631]: I1128 14:13:48.358823 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_c1f85a30-128c-487c-ac41-f1d7a64a52a1/kube-state-metrics/0.log" Nov 28 14:13:48 crc kubenswrapper[4631]: I1128 14:13:48.446819 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-pw8wp_86dddd7e-e065-4593-8d47-4d36a2cddb22/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Nov 28 14:13:48 crc kubenswrapper[4631]: I1128 14:13:48.722052 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-7f7bb9bf8f-42ssw_d05d0efc-381c-4222-8c8c-384c2c22ac9c/neutron-api/0.log" Nov 28 14:13:48 crc kubenswrapper[4631]: I1128 14:13:48.833936 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-7f7bb9bf8f-42ssw_d05d0efc-381c-4222-8c8c-384c2c22ac9c/neutron-httpd/0.log" Nov 28 14:13:48 crc kubenswrapper[4631]: I1128 14:13:48.995102 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-tl8jt_a73f7b13-ead0-46a0-8c62-d3c1e75bc2d8/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Nov 28 14:13:49 crc kubenswrapper[4631]: I1128 14:13:49.443177 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_9db3c70b-08ec-4b40-bdb9-94cdbc66fa1d/nova-api-log/0.log" Nov 28 14:13:49 crc kubenswrapper[4631]: I1128 14:13:49.514640 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_cc2a0580-660d-4a19-a755-70256a4e4abe/nova-cell0-conductor-conductor/0.log" Nov 28 14:13:49 crc kubenswrapper[4631]: I1128 14:13:49.614696 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_9db3c70b-08ec-4b40-bdb9-94cdbc66fa1d/nova-api-api/0.log" Nov 28 14:13:49 crc kubenswrapper[4631]: I1128 14:13:49.895485 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_0db994a3-5269-44c5-8a6f-8b7fdc798c3d/nova-cell1-conductor-conductor/0.log" Nov 28 14:13:50 crc kubenswrapper[4631]: I1128 14:13:50.046950 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_9c3a5a08-9590-4249-be26-8224d51da976/nova-cell1-novncproxy-novncproxy/0.log" Nov 28 14:13:50 crc kubenswrapper[4631]: I1128 14:13:50.300592 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-5hrhp_7692b0ac-e13d-4413-a2c0-e72062ec9110/nova-edpm-deployment-openstack-edpm-ipam/0.log" Nov 28 14:13:50 crc kubenswrapper[4631]: I1128 14:13:50.439745 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_52191ac4-b50d-4c9a-8a42-b3afe3cffe7f/nova-metadata-log/0.log" Nov 28 14:13:50 crc kubenswrapper[4631]: I1128 14:13:50.838785 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_83514cf7-b36a-4691-b5f9-559a1594044a/nova-scheduler-scheduler/0.log" Nov 28 14:13:50 crc kubenswrapper[4631]: I1128 14:13:50.911364 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_4bc20a10-95d5-4003-86d9-c6848bc1b188/mysql-bootstrap/0.log" Nov 28 14:13:51 crc kubenswrapper[4631]: I1128 14:13:51.273024 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_4bc20a10-95d5-4003-86d9-c6848bc1b188/galera/0.log" Nov 28 14:13:51 crc kubenswrapper[4631]: I1128 14:13:51.311723 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_4bc20a10-95d5-4003-86d9-c6848bc1b188/mysql-bootstrap/0.log" Nov 28 14:13:51 crc kubenswrapper[4631]: I1128 14:13:51.504155 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_52191ac4-b50d-4c9a-8a42-b3afe3cffe7f/nova-metadata-metadata/0.log" Nov 28 14:13:51 crc kubenswrapper[4631]: I1128 14:13:51.519438 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_1a711f62-44fc-4046-a3aa-b5ef77b45e62/mysql-bootstrap/0.log" Nov 28 14:13:51 crc kubenswrapper[4631]: I1128 14:13:51.764162 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_1a711f62-44fc-4046-a3aa-b5ef77b45e62/mysql-bootstrap/0.log" Nov 28 14:13:51 crc kubenswrapper[4631]: I1128 14:13:51.835001 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_1a711f62-44fc-4046-a3aa-b5ef77b45e62/galera/0.log" Nov 28 14:13:51 crc kubenswrapper[4631]: I1128 14:13:51.875936 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_ea063a3c-330a-4642-a815-c013ae6b80c7/openstackclient/0.log" Nov 28 14:13:52 crc kubenswrapper[4631]: I1128 14:13:52.141948 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-7jgdf_7fc5dcd5-fde5-4a72-8d4a-cda51785e77f/ovn-controller/0.log" Nov 28 14:13:52 crc kubenswrapper[4631]: I1128 14:13:52.239352 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-p7l62_6dfe4f1a-897a-4ece-a345-4a4f86901079/openstack-network-exporter/0.log" Nov 28 14:13:52 crc kubenswrapper[4631]: I1128 14:13:52.467487 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-lskqw_7fc0ee4e-435d-4297-94eb-9e8c6fa90004/ovsdb-server-init/0.log" Nov 28 14:13:52 crc kubenswrapper[4631]: I1128 14:13:52.733062 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-lskqw_7fc0ee4e-435d-4297-94eb-9e8c6fa90004/ovsdb-server-init/0.log" Nov 28 14:13:52 crc kubenswrapper[4631]: I1128 14:13:52.743196 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-lskqw_7fc0ee4e-435d-4297-94eb-9e8c6fa90004/ovs-vswitchd/0.log" Nov 28 14:13:52 crc kubenswrapper[4631]: I1128 14:13:52.925535 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-lskqw_7fc0ee4e-435d-4297-94eb-9e8c6fa90004/ovsdb-server/0.log" Nov 28 14:13:52 crc kubenswrapper[4631]: I1128 14:13:52.980631 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-2sls9_c09a85dc-b780-41c9-8122-653d4703aa7e/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Nov 28 14:13:53 crc kubenswrapper[4631]: I1128 14:13:53.150984 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_1418c5fe-bc14-4aeb-9449-30c85e26aed2/ovn-northd/0.log" Nov 28 14:13:53 crc kubenswrapper[4631]: I1128 14:13:53.259007 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_1418c5fe-bc14-4aeb-9449-30c85e26aed2/openstack-network-exporter/0.log" Nov 28 14:13:53 crc kubenswrapper[4631]: I1128 14:13:53.493185 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_ca11061d-b0a1-464d-9e1d-5572fc439580/openstack-network-exporter/0.log" Nov 28 14:13:53 crc kubenswrapper[4631]: I1128 14:13:53.501669 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_ca11061d-b0a1-464d-9e1d-5572fc439580/ovsdbserver-nb/0.log" Nov 28 14:13:53 crc kubenswrapper[4631]: I1128 14:13:53.610154 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_1d3ba54c-cb26-4e04-8310-34e6b4206119/openstack-network-exporter/0.log" Nov 28 14:13:53 crc kubenswrapper[4631]: I1128 14:13:53.756731 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_1d3ba54c-cb26-4e04-8310-34e6b4206119/ovsdbserver-sb/0.log" Nov 28 14:13:53 crc kubenswrapper[4631]: I1128 14:13:53.935864 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-769746444-8zhp8_176c2dec-4940-449b-ab1a-95cb11cdfeff/placement-api/0.log" Nov 28 14:13:54 crc kubenswrapper[4631]: I1128 14:13:54.104229 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-769746444-8zhp8_176c2dec-4940-449b-ab1a-95cb11cdfeff/placement-log/0.log" Nov 28 14:13:54 crc kubenswrapper[4631]: I1128 14:13:54.213859 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_014519bc-68f0-4132-9a46-6b0e5c538c7e/setup-container/0.log" Nov 28 14:13:54 crc kubenswrapper[4631]: I1128 14:13:54.502822 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_fc5706c2-8b7d-439a-95b3-981907360c92/setup-container/0.log" Nov 28 14:13:54 crc kubenswrapper[4631]: I1128 14:13:54.524201 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_014519bc-68f0-4132-9a46-6b0e5c538c7e/setup-container/0.log" Nov 28 14:13:54 crc kubenswrapper[4631]: I1128 14:13:54.571589 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_014519bc-68f0-4132-9a46-6b0e5c538c7e/rabbitmq/0.log" Nov 28 14:13:54 crc kubenswrapper[4631]: I1128 14:13:54.821204 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_fc5706c2-8b7d-439a-95b3-981907360c92/rabbitmq/0.log" Nov 28 14:13:54 crc kubenswrapper[4631]: I1128 14:13:54.856591 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-r2pq6_ec45d53a-f54c-422a-8a43-5b9d4b4ee0c6/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Nov 28 14:13:54 crc kubenswrapper[4631]: I1128 14:13:54.927525 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_fc5706c2-8b7d-439a-95b3-981907360c92/setup-container/0.log" Nov 28 14:13:55 crc kubenswrapper[4631]: I1128 14:13:55.257796 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_redhat-edpm-deployment-openstack-edpm-ipam-s9hz7_2b15b142-3dc9-4d33-96e7-5ba3e357ee59/redhat-edpm-deployment-openstack-edpm-ipam/0.log" Nov 28 14:13:55 crc kubenswrapper[4631]: I1128 14:13:55.338108 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-slcpp_0527a324-2259-4aac-a64b-33747a2e24ec/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Nov 28 14:13:55 crc kubenswrapper[4631]: I1128 14:13:55.583438 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-zmp56_8bdc8b41-a7c5-4c60-86ce-6ac5eb65b82f/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Nov 28 14:13:55 crc kubenswrapper[4631]: I1128 14:13:55.656745 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-m2k99_b1c516b6-6c37-4bfd-a0ee-f28200e19170/ssh-known-hosts-edpm-deployment/0.log" Nov 28 14:13:55 crc kubenswrapper[4631]: I1128 14:13:55.967385 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-68757c9bcf-d8gj4_7303e352-744c-4c93-ac72-0c434384f090/proxy-server/0.log" Nov 28 14:13:56 crc kubenswrapper[4631]: I1128 14:13:56.103054 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-68757c9bcf-d8gj4_7303e352-744c-4c93-ac72-0c434384f090/proxy-httpd/0.log" Nov 28 14:13:56 crc kubenswrapper[4631]: I1128 14:13:56.531417 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-wh7th_780a47b4-0078-4f97-a5f5-ad4e1b2015b4/swift-ring-rebalance/0.log" Nov 28 14:13:56 crc kubenswrapper[4631]: I1128 14:13:56.589689 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c3e1f6fb-1845-4a39-8eca-1262da1ff281/account-reaper/0.log" Nov 28 14:13:56 crc kubenswrapper[4631]: I1128 14:13:56.650697 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c3e1f6fb-1845-4a39-8eca-1262da1ff281/account-auditor/0.log" Nov 28 14:13:56 crc kubenswrapper[4631]: I1128 14:13:56.869131 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c3e1f6fb-1845-4a39-8eca-1262da1ff281/account-replicator/0.log" Nov 28 14:13:56 crc kubenswrapper[4631]: I1128 14:13:56.977212 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c3e1f6fb-1845-4a39-8eca-1262da1ff281/account-server/0.log" Nov 28 14:13:57 crc kubenswrapper[4631]: I1128 14:13:57.015783 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c3e1f6fb-1845-4a39-8eca-1262da1ff281/container-auditor/0.log" Nov 28 14:13:57 crc kubenswrapper[4631]: I1128 14:13:57.044351 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c3e1f6fb-1845-4a39-8eca-1262da1ff281/container-replicator/0.log" Nov 28 14:13:57 crc kubenswrapper[4631]: I1128 14:13:57.205520 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c3e1f6fb-1845-4a39-8eca-1262da1ff281/container-server/0.log" Nov 28 14:13:57 crc kubenswrapper[4631]: I1128 14:13:57.272613 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c3e1f6fb-1845-4a39-8eca-1262da1ff281/container-updater/0.log" Nov 28 14:13:57 crc kubenswrapper[4631]: I1128 14:13:57.309608 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c3e1f6fb-1845-4a39-8eca-1262da1ff281/object-auditor/0.log" Nov 28 14:13:57 crc kubenswrapper[4631]: I1128 14:13:57.403250 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c3e1f6fb-1845-4a39-8eca-1262da1ff281/object-expirer/0.log" Nov 28 14:13:57 crc kubenswrapper[4631]: I1128 14:13:57.512923 4631 scope.go:117] "RemoveContainer" containerID="d188114d99f65821ef24c4d713f07ed3cbe0cd10a5aaa23d6346374210035dc4" Nov 28 14:13:57 crc kubenswrapper[4631]: E1128 14:13:57.513359 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-47sxw_openshift-machine-config-operator(8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1)\"" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" podUID="8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1" Nov 28 14:13:57 crc kubenswrapper[4631]: I1128 14:13:57.605435 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c3e1f6fb-1845-4a39-8eca-1262da1ff281/object-updater/0.log" Nov 28 14:13:57 crc kubenswrapper[4631]: I1128 14:13:57.616317 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c3e1f6fb-1845-4a39-8eca-1262da1ff281/object-replicator/0.log" Nov 28 14:13:57 crc kubenswrapper[4631]: I1128 14:13:57.659552 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c3e1f6fb-1845-4a39-8eca-1262da1ff281/object-server/0.log" Nov 28 14:13:57 crc kubenswrapper[4631]: I1128 14:13:57.694028 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c3e1f6fb-1845-4a39-8eca-1262da1ff281/rsync/0.log" Nov 28 14:13:57 crc kubenswrapper[4631]: I1128 14:13:57.981104 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-x9bvm_dab7280d-4cb8-4a43-b056-82296e560317/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Nov 28 14:13:58 crc kubenswrapper[4631]: I1128 14:13:58.026269 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c3e1f6fb-1845-4a39-8eca-1262da1ff281/swift-recon-cron/0.log" Nov 28 14:13:58 crc kubenswrapper[4631]: I1128 14:13:58.244609 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_183d2a7b-cd4e-423c-82c0-a72e2ebec301/tempest-tests-tempest-tests-runner/0.log" Nov 28 14:13:58 crc kubenswrapper[4631]: I1128 14:13:58.357374 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_7fffbcdc-ae30-4841-9161-7a0f06bb7f67/test-operator-logs-container/0.log" Nov 28 14:13:58 crc kubenswrapper[4631]: I1128 14:13:58.542383 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-pwfzg_53cec30e-969e-4ccf-8ae2-76852297457a/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Nov 28 14:14:05 crc kubenswrapper[4631]: I1128 14:14:05.784762 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_d35228a7-a0fe-496d-9d62-5f1d5a44c72f/memcached/0.log" Nov 28 14:14:08 crc kubenswrapper[4631]: I1128 14:14:08.513406 4631 scope.go:117] "RemoveContainer" containerID="d188114d99f65821ef24c4d713f07ed3cbe0cd10a5aaa23d6346374210035dc4" Nov 28 14:14:08 crc kubenswrapper[4631]: E1128 14:14:08.515157 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-47sxw_openshift-machine-config-operator(8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1)\"" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" podUID="8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1" Nov 28 14:14:21 crc kubenswrapper[4631]: I1128 14:14:21.513744 4631 scope.go:117] "RemoveContainer" containerID="d188114d99f65821ef24c4d713f07ed3cbe0cd10a5aaa23d6346374210035dc4" Nov 28 14:14:21 crc kubenswrapper[4631]: E1128 14:14:21.514572 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-47sxw_openshift-machine-config-operator(8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1)\"" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" podUID="8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1" Nov 28 14:14:27 crc kubenswrapper[4631]: I1128 14:14:27.744208 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7b64f4fb85-hdr72_00a81f80-35b9-4eb4-a508-3f104ca426be/manager/0.log" Nov 28 14:14:27 crc kubenswrapper[4631]: I1128 14:14:27.759769 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7b64f4fb85-hdr72_00a81f80-35b9-4eb4-a508-3f104ca426be/kube-rbac-proxy/0.log" Nov 28 14:14:28 crc kubenswrapper[4631]: I1128 14:14:28.009144 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-6b7f75547b-rmlgq_8028d28d-2414-4ffe-9c5d-4854f77e980f/kube-rbac-proxy/0.log" Nov 28 14:14:28 crc kubenswrapper[4631]: I1128 14:14:28.049158 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-6b7f75547b-rmlgq_8028d28d-2414-4ffe-9c5d-4854f77e980f/manager/0.log" Nov 28 14:14:28 crc kubenswrapper[4631]: I1128 14:14:28.234134 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_dee8d58050df5e5c4d52f5078d12ac93857da536112fc471b2066b915e9wbcx_57ea9be9-7467-4c95-b00b-14f7df44eae4/util/0.log" Nov 28 14:14:28 crc kubenswrapper[4631]: I1128 14:14:28.409798 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_dee8d58050df5e5c4d52f5078d12ac93857da536112fc471b2066b915e9wbcx_57ea9be9-7467-4c95-b00b-14f7df44eae4/util/0.log" Nov 28 14:14:28 crc kubenswrapper[4631]: I1128 14:14:28.464303 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_dee8d58050df5e5c4d52f5078d12ac93857da536112fc471b2066b915e9wbcx_57ea9be9-7467-4c95-b00b-14f7df44eae4/pull/0.log" Nov 28 14:14:28 crc kubenswrapper[4631]: I1128 14:14:28.514174 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_dee8d58050df5e5c4d52f5078d12ac93857da536112fc471b2066b915e9wbcx_57ea9be9-7467-4c95-b00b-14f7df44eae4/pull/0.log" Nov 28 14:14:28 crc kubenswrapper[4631]: I1128 14:14:28.636991 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_dee8d58050df5e5c4d52f5078d12ac93857da536112fc471b2066b915e9wbcx_57ea9be9-7467-4c95-b00b-14f7df44eae4/pull/0.log" Nov 28 14:14:28 crc kubenswrapper[4631]: I1128 14:14:28.649595 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_dee8d58050df5e5c4d52f5078d12ac93857da536112fc471b2066b915e9wbcx_57ea9be9-7467-4c95-b00b-14f7df44eae4/util/0.log" Nov 28 14:14:28 crc kubenswrapper[4631]: I1128 14:14:28.696623 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_dee8d58050df5e5c4d52f5078d12ac93857da536112fc471b2066b915e9wbcx_57ea9be9-7467-4c95-b00b-14f7df44eae4/extract/0.log" Nov 28 14:14:28 crc kubenswrapper[4631]: I1128 14:14:28.861373 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-955677c94-6bt2f_06994d92-3dcb-461c-83ac-c7e65d4bd640/manager/0.log" Nov 28 14:14:28 crc kubenswrapper[4631]: I1128 14:14:28.888448 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-955677c94-6bt2f_06994d92-3dcb-461c-83ac-c7e65d4bd640/kube-rbac-proxy/0.log" Nov 28 14:14:29 crc kubenswrapper[4631]: I1128 14:14:29.035195 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-589cbd6b5b-npdtc_59ee25ad-f169-47b9-9d5c-f7a7a253a591/kube-rbac-proxy/0.log" Nov 28 14:14:29 crc kubenswrapper[4631]: I1128 14:14:29.189658 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-589cbd6b5b-npdtc_59ee25ad-f169-47b9-9d5c-f7a7a253a591/manager/0.log" Nov 28 14:14:29 crc kubenswrapper[4631]: I1128 14:14:29.282653 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5b77f656f-8pvq9_9cc92c5c-285e-4604-bc60-7637938a78c8/kube-rbac-proxy/0.log" Nov 28 14:14:29 crc kubenswrapper[4631]: I1128 14:14:29.424529 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5b77f656f-8pvq9_9cc92c5c-285e-4604-bc60-7637938a78c8/manager/0.log" Nov 28 14:14:29 crc kubenswrapper[4631]: I1128 14:14:29.455653 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-5d494799bf-2mq7v_aafff68d-b989-4edc-ab1c-b9d33edb8ee2/kube-rbac-proxy/0.log" Nov 28 14:14:29 crc kubenswrapper[4631]: I1128 14:14:29.541122 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-5d494799bf-2mq7v_aafff68d-b989-4edc-ab1c-b9d33edb8ee2/manager/0.log" Nov 28 14:14:29 crc kubenswrapper[4631]: I1128 14:14:29.661785 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-57548d458d-h7npd_0079ab2d-c5a8-45eb-b6e8-9574158de32f/kube-rbac-proxy/0.log" Nov 28 14:14:29 crc kubenswrapper[4631]: I1128 14:14:29.901087 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-67cb4dc6d4-db8kg_ed17d71f-b72c-4f12-b5f7-d5cb3bb130dc/kube-rbac-proxy/0.log" Nov 28 14:14:29 crc kubenswrapper[4631]: I1128 14:14:29.902543 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-57548d458d-h7npd_0079ab2d-c5a8-45eb-b6e8-9574158de32f/manager/0.log" Nov 28 14:14:29 crc kubenswrapper[4631]: I1128 14:14:29.989661 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-67cb4dc6d4-db8kg_ed17d71f-b72c-4f12-b5f7-d5cb3bb130dc/manager/0.log" Nov 28 14:14:30 crc kubenswrapper[4631]: I1128 14:14:30.247253 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-6b49bfffcd-5ks5f_f3e12436-8a74-4443-94a4-3b9717105b05/manager/0.log" Nov 28 14:14:30 crc kubenswrapper[4631]: I1128 14:14:30.261688 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-6b49bfffcd-5ks5f_f3e12436-8a74-4443-94a4-3b9717105b05/kube-rbac-proxy/0.log" Nov 28 14:14:30 crc kubenswrapper[4631]: I1128 14:14:30.414893 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-5d499bf58b-dck2d_dbbf77e7-3c97-480f-83b1-0eb8bb8f386d/kube-rbac-proxy/0.log" Nov 28 14:14:30 crc kubenswrapper[4631]: I1128 14:14:30.486542 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-5d499bf58b-dck2d_dbbf77e7-3c97-480f-83b1-0eb8bb8f386d/manager/0.log" Nov 28 14:14:30 crc kubenswrapper[4631]: I1128 14:14:30.535680 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-66f4dd4bc7-m9t96_cefc6050-9552-4dda-9864-39ac05626785/kube-rbac-proxy/0.log" Nov 28 14:14:30 crc kubenswrapper[4631]: I1128 14:14:30.710866 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-66f4dd4bc7-m9t96_cefc6050-9552-4dda-9864-39ac05626785/manager/0.log" Nov 28 14:14:30 crc kubenswrapper[4631]: I1128 14:14:30.723969 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-6fdcddb789-msc6n_2dab52af-5e82-4119-b4d5-56fb31abcb4d/kube-rbac-proxy/0.log" Nov 28 14:14:30 crc kubenswrapper[4631]: I1128 14:14:30.811100 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-6fdcddb789-msc6n_2dab52af-5e82-4119-b4d5-56fb31abcb4d/manager/0.log" Nov 28 14:14:31 crc kubenswrapper[4631]: I1128 14:14:31.014925 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-79556f57fc-tpbhk_bb908ace-bb62-4f55-8131-90e48da7585b/kube-rbac-proxy/0.log" Nov 28 14:14:31 crc kubenswrapper[4631]: I1128 14:14:31.116551 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-79556f57fc-tpbhk_bb908ace-bb62-4f55-8131-90e48da7585b/manager/0.log" Nov 28 14:14:31 crc kubenswrapper[4631]: I1128 14:14:31.290329 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-64cdc6ff96-4qzz7_ecfd67ad-5136-4e6e-9e41-3d933d10ea2c/kube-rbac-proxy/0.log" Nov 28 14:14:31 crc kubenswrapper[4631]: I1128 14:14:31.301496 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-64cdc6ff96-4qzz7_ecfd67ad-5136-4e6e-9e41-3d933d10ea2c/manager/0.log" Nov 28 14:14:31 crc kubenswrapper[4631]: I1128 14:14:31.370263 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-5fcdb54b6b2f9vf_ca18d816-8a63-4dcd-ac27-b1e7ad9ce4f6/kube-rbac-proxy/0.log" Nov 28 14:14:31 crc kubenswrapper[4631]: I1128 14:14:31.609052 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-5fcdb54b6b2f9vf_ca18d816-8a63-4dcd-ac27-b1e7ad9ce4f6/manager/0.log" Nov 28 14:14:32 crc kubenswrapper[4631]: I1128 14:14:32.101448 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-qclsp_8c8c3b22-0767-4d9f-9ec4-bf42ddcd8433/registry-server/0.log" Nov 28 14:14:32 crc kubenswrapper[4631]: I1128 14:14:32.278903 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-5bcf64b564-zsrq9_1104dd12-c822-4ffa-9634-d82898f8f575/operator/0.log" Nov 28 14:14:32 crc kubenswrapper[4631]: I1128 14:14:32.550537 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-56897c768d-wkpql_d09fac54-dd24-47df-83a1-23cae3102e7a/kube-rbac-proxy/0.log" Nov 28 14:14:32 crc kubenswrapper[4631]: I1128 14:14:32.693038 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-56897c768d-wkpql_d09fac54-dd24-47df-83a1-23cae3102e7a/manager/0.log" Nov 28 14:14:32 crc kubenswrapper[4631]: I1128 14:14:32.785192 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-57988cc5b5-wqpdp_05a6dcba-27b4-4b20-9275-f434b8943be9/kube-rbac-proxy/0.log" Nov 28 14:14:32 crc kubenswrapper[4631]: I1128 14:14:32.883992 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-854b9f67c6-hfb26_4bc38647-eefa-4ad6-bef5-6a7ed535b794/manager/0.log" Nov 28 14:14:32 crc kubenswrapper[4631]: I1128 14:14:32.934024 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-57988cc5b5-wqpdp_05a6dcba-27b4-4b20-9275-f434b8943be9/manager/0.log" Nov 28 14:14:33 crc kubenswrapper[4631]: I1128 14:14:33.076326 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-nrpfx_10a95153-0898-4016-92a6-d56633555348/operator/0.log" Nov 28 14:14:33 crc kubenswrapper[4631]: I1128 14:14:33.154878 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-d77b94747-2g446_2d59108a-2b92-421b-beb8-157c2eca09ac/kube-rbac-proxy/0.log" Nov 28 14:14:33 crc kubenswrapper[4631]: I1128 14:14:33.231540 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-d77b94747-2g446_2d59108a-2b92-421b-beb8-157c2eca09ac/manager/0.log" Nov 28 14:14:33 crc kubenswrapper[4631]: I1128 14:14:33.386954 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-76cc84c6bb-r94wn_f2aea9a5-85cf-464f-8eba-382b3db8d71d/kube-rbac-proxy/0.log" Nov 28 14:14:33 crc kubenswrapper[4631]: I1128 14:14:33.484748 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-76cc84c6bb-r94wn_f2aea9a5-85cf-464f-8eba-382b3db8d71d/manager/0.log" Nov 28 14:14:33 crc kubenswrapper[4631]: I1128 14:14:33.610937 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5cd6c7f4c8-jwd25_fa36877c-8fc2-4074-a75d-135c2006807a/kube-rbac-proxy/0.log" Nov 28 14:14:33 crc kubenswrapper[4631]: I1128 14:14:33.660193 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5cd6c7f4c8-jwd25_fa36877c-8fc2-4074-a75d-135c2006807a/manager/0.log" Nov 28 14:14:33 crc kubenswrapper[4631]: I1128 14:14:33.825773 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-656dcb59d4-wh9v9_826f27ac-a938-4292-a357-ff54c4117986/kube-rbac-proxy/0.log" Nov 28 14:14:33 crc kubenswrapper[4631]: I1128 14:14:33.902209 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-656dcb59d4-wh9v9_826f27ac-a938-4292-a357-ff54c4117986/manager/0.log" Nov 28 14:14:34 crc kubenswrapper[4631]: I1128 14:14:34.513859 4631 scope.go:117] "RemoveContainer" containerID="d188114d99f65821ef24c4d713f07ed3cbe0cd10a5aaa23d6346374210035dc4" Nov 28 14:14:34 crc kubenswrapper[4631]: E1128 14:14:34.514450 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-47sxw_openshift-machine-config-operator(8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1)\"" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" podUID="8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1" Nov 28 14:14:48 crc kubenswrapper[4631]: I1128 14:14:48.513112 4631 scope.go:117] "RemoveContainer" containerID="d188114d99f65821ef24c4d713f07ed3cbe0cd10a5aaa23d6346374210035dc4" Nov 28 14:14:48 crc kubenswrapper[4631]: E1128 14:14:48.513924 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-47sxw_openshift-machine-config-operator(8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1)\"" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" podUID="8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1" Nov 28 14:14:56 crc kubenswrapper[4631]: I1128 14:14:56.966809 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-2v5d6_2a2f7727-f05e-4deb-8eba-cb3ec3561621/control-plane-machine-set-operator/0.log" Nov 28 14:14:57 crc kubenswrapper[4631]: I1128 14:14:57.154767 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-2ch2b_889c117d-ea46-4f02-a2e0-42a47c6e4683/kube-rbac-proxy/0.log" Nov 28 14:14:57 crc kubenswrapper[4631]: I1128 14:14:57.175818 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-2ch2b_889c117d-ea46-4f02-a2e0-42a47c6e4683/machine-api-operator/0.log" Nov 28 14:15:00 crc kubenswrapper[4631]: I1128 14:15:00.164915 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29405655-lsd6b"] Nov 28 14:15:00 crc kubenswrapper[4631]: E1128 14:15:00.166073 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7c5efc82-a549-4aa1-9a1a-cc24584097d0" containerName="container-00" Nov 28 14:15:00 crc kubenswrapper[4631]: I1128 14:15:00.166093 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="7c5efc82-a549-4aa1-9a1a-cc24584097d0" containerName="container-00" Nov 28 14:15:00 crc kubenswrapper[4631]: I1128 14:15:00.166480 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="7c5efc82-a549-4aa1-9a1a-cc24584097d0" containerName="container-00" Nov 28 14:15:00 crc kubenswrapper[4631]: I1128 14:15:00.167395 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29405655-lsd6b" Nov 28 14:15:00 crc kubenswrapper[4631]: I1128 14:15:00.170174 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 28 14:15:00 crc kubenswrapper[4631]: I1128 14:15:00.170776 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 28 14:15:00 crc kubenswrapper[4631]: I1128 14:15:00.188412 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29405655-lsd6b"] Nov 28 14:15:00 crc kubenswrapper[4631]: I1128 14:15:00.214552 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/45746f49-c18f-4c80-aa8a-c8906d4aa93c-secret-volume\") pod \"collect-profiles-29405655-lsd6b\" (UID: \"45746f49-c18f-4c80-aa8a-c8906d4aa93c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405655-lsd6b" Nov 28 14:15:00 crc kubenswrapper[4631]: I1128 14:15:00.214649 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x5kn6\" (UniqueName: \"kubernetes.io/projected/45746f49-c18f-4c80-aa8a-c8906d4aa93c-kube-api-access-x5kn6\") pod \"collect-profiles-29405655-lsd6b\" (UID: \"45746f49-c18f-4c80-aa8a-c8906d4aa93c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405655-lsd6b" Nov 28 14:15:00 crc kubenswrapper[4631]: I1128 14:15:00.215095 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/45746f49-c18f-4c80-aa8a-c8906d4aa93c-config-volume\") pod \"collect-profiles-29405655-lsd6b\" (UID: \"45746f49-c18f-4c80-aa8a-c8906d4aa93c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405655-lsd6b" Nov 28 14:15:00 crc kubenswrapper[4631]: I1128 14:15:00.319667 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x5kn6\" (UniqueName: \"kubernetes.io/projected/45746f49-c18f-4c80-aa8a-c8906d4aa93c-kube-api-access-x5kn6\") pod \"collect-profiles-29405655-lsd6b\" (UID: \"45746f49-c18f-4c80-aa8a-c8906d4aa93c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405655-lsd6b" Nov 28 14:15:00 crc kubenswrapper[4631]: I1128 14:15:00.319798 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/45746f49-c18f-4c80-aa8a-c8906d4aa93c-config-volume\") pod \"collect-profiles-29405655-lsd6b\" (UID: \"45746f49-c18f-4c80-aa8a-c8906d4aa93c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405655-lsd6b" Nov 28 14:15:00 crc kubenswrapper[4631]: I1128 14:15:00.319943 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/45746f49-c18f-4c80-aa8a-c8906d4aa93c-secret-volume\") pod \"collect-profiles-29405655-lsd6b\" (UID: \"45746f49-c18f-4c80-aa8a-c8906d4aa93c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405655-lsd6b" Nov 28 14:15:00 crc kubenswrapper[4631]: I1128 14:15:00.321129 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/45746f49-c18f-4c80-aa8a-c8906d4aa93c-config-volume\") pod \"collect-profiles-29405655-lsd6b\" (UID: \"45746f49-c18f-4c80-aa8a-c8906d4aa93c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405655-lsd6b" Nov 28 14:15:00 crc kubenswrapper[4631]: I1128 14:15:00.328043 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/45746f49-c18f-4c80-aa8a-c8906d4aa93c-secret-volume\") pod \"collect-profiles-29405655-lsd6b\" (UID: \"45746f49-c18f-4c80-aa8a-c8906d4aa93c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405655-lsd6b" Nov 28 14:15:00 crc kubenswrapper[4631]: I1128 14:15:00.338735 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x5kn6\" (UniqueName: \"kubernetes.io/projected/45746f49-c18f-4c80-aa8a-c8906d4aa93c-kube-api-access-x5kn6\") pod \"collect-profiles-29405655-lsd6b\" (UID: \"45746f49-c18f-4c80-aa8a-c8906d4aa93c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405655-lsd6b" Nov 28 14:15:00 crc kubenswrapper[4631]: I1128 14:15:00.489062 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29405655-lsd6b" Nov 28 14:15:01 crc kubenswrapper[4631]: W1128 14:15:01.074997 4631 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod45746f49_c18f_4c80_aa8a_c8906d4aa93c.slice/crio-c038cfb35ea72d2f2d98ae403ab91e31d29074c01c01577fe8b8630089aa7de8 WatchSource:0}: Error finding container c038cfb35ea72d2f2d98ae403ab91e31d29074c01c01577fe8b8630089aa7de8: Status 404 returned error can't find the container with id c038cfb35ea72d2f2d98ae403ab91e31d29074c01c01577fe8b8630089aa7de8 Nov 28 14:15:01 crc kubenswrapper[4631]: I1128 14:15:01.081427 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29405655-lsd6b"] Nov 28 14:15:01 crc kubenswrapper[4631]: I1128 14:15:01.271655 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29405655-lsd6b" event={"ID":"45746f49-c18f-4c80-aa8a-c8906d4aa93c","Type":"ContainerStarted","Data":"6be524071c1e2f6aed7abac9cd13b1719ee0bb2abd9e848fea07f8d4eee38091"} Nov 28 14:15:01 crc kubenswrapper[4631]: I1128 14:15:01.271698 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29405655-lsd6b" event={"ID":"45746f49-c18f-4c80-aa8a-c8906d4aa93c","Type":"ContainerStarted","Data":"c038cfb35ea72d2f2d98ae403ab91e31d29074c01c01577fe8b8630089aa7de8"} Nov 28 14:15:01 crc kubenswrapper[4631]: I1128 14:15:01.290827 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29405655-lsd6b" podStartSLOduration=1.2908052749999999 podStartE2EDuration="1.290805275s" podCreationTimestamp="2025-11-28 14:15:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 14:15:01.287477484 +0000 UTC m=+3258.094780828" watchObservedRunningTime="2025-11-28 14:15:01.290805275 +0000 UTC m=+3258.098108619" Nov 28 14:15:02 crc kubenswrapper[4631]: I1128 14:15:02.284026 4631 generic.go:334] "Generic (PLEG): container finished" podID="45746f49-c18f-4c80-aa8a-c8906d4aa93c" containerID="6be524071c1e2f6aed7abac9cd13b1719ee0bb2abd9e848fea07f8d4eee38091" exitCode=0 Nov 28 14:15:02 crc kubenswrapper[4631]: I1128 14:15:02.284361 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29405655-lsd6b" event={"ID":"45746f49-c18f-4c80-aa8a-c8906d4aa93c","Type":"ContainerDied","Data":"6be524071c1e2f6aed7abac9cd13b1719ee0bb2abd9e848fea07f8d4eee38091"} Nov 28 14:15:02 crc kubenswrapper[4631]: I1128 14:15:02.513411 4631 scope.go:117] "RemoveContainer" containerID="d188114d99f65821ef24c4d713f07ed3cbe0cd10a5aaa23d6346374210035dc4" Nov 28 14:15:02 crc kubenswrapper[4631]: E1128 14:15:02.513658 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-47sxw_openshift-machine-config-operator(8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1)\"" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" podUID="8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1" Nov 28 14:15:03 crc kubenswrapper[4631]: I1128 14:15:03.731204 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29405655-lsd6b" Nov 28 14:15:03 crc kubenswrapper[4631]: I1128 14:15:03.783791 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/45746f49-c18f-4c80-aa8a-c8906d4aa93c-config-volume\") pod \"45746f49-c18f-4c80-aa8a-c8906d4aa93c\" (UID: \"45746f49-c18f-4c80-aa8a-c8906d4aa93c\") " Nov 28 14:15:03 crc kubenswrapper[4631]: I1128 14:15:03.783903 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x5kn6\" (UniqueName: \"kubernetes.io/projected/45746f49-c18f-4c80-aa8a-c8906d4aa93c-kube-api-access-x5kn6\") pod \"45746f49-c18f-4c80-aa8a-c8906d4aa93c\" (UID: \"45746f49-c18f-4c80-aa8a-c8906d4aa93c\") " Nov 28 14:15:03 crc kubenswrapper[4631]: I1128 14:15:03.783990 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/45746f49-c18f-4c80-aa8a-c8906d4aa93c-secret-volume\") pod \"45746f49-c18f-4c80-aa8a-c8906d4aa93c\" (UID: \"45746f49-c18f-4c80-aa8a-c8906d4aa93c\") " Nov 28 14:15:03 crc kubenswrapper[4631]: I1128 14:15:03.785540 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/45746f49-c18f-4c80-aa8a-c8906d4aa93c-config-volume" (OuterVolumeSpecName: "config-volume") pod "45746f49-c18f-4c80-aa8a-c8906d4aa93c" (UID: "45746f49-c18f-4c80-aa8a-c8906d4aa93c"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:15:03 crc kubenswrapper[4631]: I1128 14:15:03.789732 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/45746f49-c18f-4c80-aa8a-c8906d4aa93c-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "45746f49-c18f-4c80-aa8a-c8906d4aa93c" (UID: "45746f49-c18f-4c80-aa8a-c8906d4aa93c"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:15:03 crc kubenswrapper[4631]: I1128 14:15:03.802231 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/45746f49-c18f-4c80-aa8a-c8906d4aa93c-kube-api-access-x5kn6" (OuterVolumeSpecName: "kube-api-access-x5kn6") pod "45746f49-c18f-4c80-aa8a-c8906d4aa93c" (UID: "45746f49-c18f-4c80-aa8a-c8906d4aa93c"). InnerVolumeSpecName "kube-api-access-x5kn6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:15:03 crc kubenswrapper[4631]: I1128 14:15:03.885516 4631 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/45746f49-c18f-4c80-aa8a-c8906d4aa93c-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 28 14:15:03 crc kubenswrapper[4631]: I1128 14:15:03.885558 4631 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/45746f49-c18f-4c80-aa8a-c8906d4aa93c-config-volume\") on node \"crc\" DevicePath \"\"" Nov 28 14:15:03 crc kubenswrapper[4631]: I1128 14:15:03.885572 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x5kn6\" (UniqueName: \"kubernetes.io/projected/45746f49-c18f-4c80-aa8a-c8906d4aa93c-kube-api-access-x5kn6\") on node \"crc\" DevicePath \"\"" Nov 28 14:15:04 crc kubenswrapper[4631]: I1128 14:15:04.303162 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29405655-lsd6b" event={"ID":"45746f49-c18f-4c80-aa8a-c8906d4aa93c","Type":"ContainerDied","Data":"c038cfb35ea72d2f2d98ae403ab91e31d29074c01c01577fe8b8630089aa7de8"} Nov 28 14:15:04 crc kubenswrapper[4631]: I1128 14:15:04.303529 4631 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c038cfb35ea72d2f2d98ae403ab91e31d29074c01c01577fe8b8630089aa7de8" Nov 28 14:15:04 crc kubenswrapper[4631]: I1128 14:15:04.303230 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29405655-lsd6b" Nov 28 14:15:04 crc kubenswrapper[4631]: I1128 14:15:04.372969 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29405610-wbcgx"] Nov 28 14:15:04 crc kubenswrapper[4631]: I1128 14:15:04.383319 4631 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29405610-wbcgx"] Nov 28 14:15:05 crc kubenswrapper[4631]: I1128 14:15:05.525795 4631 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aed94628-7bd5-4f59-add9-81a80b78a133" path="/var/lib/kubelet/pods/aed94628-7bd5-4f59-add9-81a80b78a133/volumes" Nov 28 14:15:06 crc kubenswrapper[4631]: I1128 14:15:06.477108 4631 scope.go:117] "RemoveContainer" containerID="8645aad4175aa8cba8b2690f5c3464b2bccdd8a0134b163b25c95b78477ae78b" Nov 28 14:15:11 crc kubenswrapper[4631]: I1128 14:15:11.299859 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-kf7cv_4e46e751-ddac-4bc4-b463-d7cf9b772871/cert-manager-controller/0.log" Nov 28 14:15:11 crc kubenswrapper[4631]: I1128 14:15:11.485998 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-rslr5_6591b4ca-08f0-4346-aed6-8e9d1caf2b7f/cert-manager-cainjector/0.log" Nov 28 14:15:11 crc kubenswrapper[4631]: I1128 14:15:11.521593 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-l6tzn_11540749-0bb3-4c94-9d22-12a889ac10ce/cert-manager-webhook/0.log" Nov 28 14:15:13 crc kubenswrapper[4631]: I1128 14:15:13.524512 4631 scope.go:117] "RemoveContainer" containerID="d188114d99f65821ef24c4d713f07ed3cbe0cd10a5aaa23d6346374210035dc4" Nov 28 14:15:13 crc kubenswrapper[4631]: E1128 14:15:13.525390 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-47sxw_openshift-machine-config-operator(8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1)\"" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" podUID="8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1" Nov 28 14:15:25 crc kubenswrapper[4631]: I1128 14:15:25.374544 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-7fbb5f6569-k9dxt_61c15dca-2c0d-4c3f-b63c-6a6b7be6ad99/nmstate-console-plugin/0.log" Nov 28 14:15:25 crc kubenswrapper[4631]: I1128 14:15:25.481691 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-9t5sn_6dd14cda-0401-4840-8e08-aebcb536ffb9/nmstate-handler/0.log" Nov 28 14:15:25 crc kubenswrapper[4631]: I1128 14:15:25.722474 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-z7fvn_40f6a5ae-0fc9-4af6-a001-667ce456f18a/nmstate-metrics/0.log" Nov 28 14:15:25 crc kubenswrapper[4631]: I1128 14:15:25.788156 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-z7fvn_40f6a5ae-0fc9-4af6-a001-667ce456f18a/kube-rbac-proxy/0.log" Nov 28 14:15:25 crc kubenswrapper[4631]: I1128 14:15:25.896986 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-5b5b58f5c8-8swhv_9124a25d-f057-4244-bd4d-a612b7a4a01b/nmstate-operator/0.log" Nov 28 14:15:26 crc kubenswrapper[4631]: I1128 14:15:26.033586 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-5f6d4c5ccb-b2ssz_f4c0815f-e4ee-4be8-855a-9bf949bf91a0/nmstate-webhook/0.log" Nov 28 14:15:26 crc kubenswrapper[4631]: I1128 14:15:26.513543 4631 scope.go:117] "RemoveContainer" containerID="d188114d99f65821ef24c4d713f07ed3cbe0cd10a5aaa23d6346374210035dc4" Nov 28 14:15:26 crc kubenswrapper[4631]: E1128 14:15:26.513980 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-47sxw_openshift-machine-config-operator(8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1)\"" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" podUID="8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1" Nov 28 14:15:35 crc kubenswrapper[4631]: I1128 14:15:35.742444 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-vtcht"] Nov 28 14:15:35 crc kubenswrapper[4631]: E1128 14:15:35.744970 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="45746f49-c18f-4c80-aa8a-c8906d4aa93c" containerName="collect-profiles" Nov 28 14:15:35 crc kubenswrapper[4631]: I1128 14:15:35.745004 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="45746f49-c18f-4c80-aa8a-c8906d4aa93c" containerName="collect-profiles" Nov 28 14:15:35 crc kubenswrapper[4631]: I1128 14:15:35.745246 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="45746f49-c18f-4c80-aa8a-c8906d4aa93c" containerName="collect-profiles" Nov 28 14:15:35 crc kubenswrapper[4631]: I1128 14:15:35.747308 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vtcht" Nov 28 14:15:35 crc kubenswrapper[4631]: I1128 14:15:35.767445 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-vtcht"] Nov 28 14:15:35 crc kubenswrapper[4631]: I1128 14:15:35.904079 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/69cad66f-388f-4703-8ed5-07cdc939dbef-utilities\") pod \"redhat-operators-vtcht\" (UID: \"69cad66f-388f-4703-8ed5-07cdc939dbef\") " pod="openshift-marketplace/redhat-operators-vtcht" Nov 28 14:15:35 crc kubenswrapper[4631]: I1128 14:15:35.904167 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/69cad66f-388f-4703-8ed5-07cdc939dbef-catalog-content\") pod \"redhat-operators-vtcht\" (UID: \"69cad66f-388f-4703-8ed5-07cdc939dbef\") " pod="openshift-marketplace/redhat-operators-vtcht" Nov 28 14:15:35 crc kubenswrapper[4631]: I1128 14:15:35.904259 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hqdkc\" (UniqueName: \"kubernetes.io/projected/69cad66f-388f-4703-8ed5-07cdc939dbef-kube-api-access-hqdkc\") pod \"redhat-operators-vtcht\" (UID: \"69cad66f-388f-4703-8ed5-07cdc939dbef\") " pod="openshift-marketplace/redhat-operators-vtcht" Nov 28 14:15:36 crc kubenswrapper[4631]: I1128 14:15:36.007832 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/69cad66f-388f-4703-8ed5-07cdc939dbef-utilities\") pod \"redhat-operators-vtcht\" (UID: \"69cad66f-388f-4703-8ed5-07cdc939dbef\") " pod="openshift-marketplace/redhat-operators-vtcht" Nov 28 14:15:36 crc kubenswrapper[4631]: I1128 14:15:36.007996 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/69cad66f-388f-4703-8ed5-07cdc939dbef-catalog-content\") pod \"redhat-operators-vtcht\" (UID: \"69cad66f-388f-4703-8ed5-07cdc939dbef\") " pod="openshift-marketplace/redhat-operators-vtcht" Nov 28 14:15:36 crc kubenswrapper[4631]: I1128 14:15:36.008119 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hqdkc\" (UniqueName: \"kubernetes.io/projected/69cad66f-388f-4703-8ed5-07cdc939dbef-kube-api-access-hqdkc\") pod \"redhat-operators-vtcht\" (UID: \"69cad66f-388f-4703-8ed5-07cdc939dbef\") " pod="openshift-marketplace/redhat-operators-vtcht" Nov 28 14:15:36 crc kubenswrapper[4631]: I1128 14:15:36.008805 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/69cad66f-388f-4703-8ed5-07cdc939dbef-catalog-content\") pod \"redhat-operators-vtcht\" (UID: \"69cad66f-388f-4703-8ed5-07cdc939dbef\") " pod="openshift-marketplace/redhat-operators-vtcht" Nov 28 14:15:36 crc kubenswrapper[4631]: I1128 14:15:36.009458 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/69cad66f-388f-4703-8ed5-07cdc939dbef-utilities\") pod \"redhat-operators-vtcht\" (UID: \"69cad66f-388f-4703-8ed5-07cdc939dbef\") " pod="openshift-marketplace/redhat-operators-vtcht" Nov 28 14:15:36 crc kubenswrapper[4631]: I1128 14:15:36.037853 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hqdkc\" (UniqueName: \"kubernetes.io/projected/69cad66f-388f-4703-8ed5-07cdc939dbef-kube-api-access-hqdkc\") pod \"redhat-operators-vtcht\" (UID: \"69cad66f-388f-4703-8ed5-07cdc939dbef\") " pod="openshift-marketplace/redhat-operators-vtcht" Nov 28 14:15:36 crc kubenswrapper[4631]: I1128 14:15:36.076609 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vtcht" Nov 28 14:15:36 crc kubenswrapper[4631]: I1128 14:15:36.612483 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-vtcht"] Nov 28 14:15:37 crc kubenswrapper[4631]: I1128 14:15:37.638842 4631 generic.go:334] "Generic (PLEG): container finished" podID="69cad66f-388f-4703-8ed5-07cdc939dbef" containerID="11f55df99c87535f98013d4bda986e135dc15f20fe56a06b1c95207fd3df977e" exitCode=0 Nov 28 14:15:37 crc kubenswrapper[4631]: I1128 14:15:37.639008 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vtcht" event={"ID":"69cad66f-388f-4703-8ed5-07cdc939dbef","Type":"ContainerDied","Data":"11f55df99c87535f98013d4bda986e135dc15f20fe56a06b1c95207fd3df977e"} Nov 28 14:15:37 crc kubenswrapper[4631]: I1128 14:15:37.639145 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vtcht" event={"ID":"69cad66f-388f-4703-8ed5-07cdc939dbef","Type":"ContainerStarted","Data":"e26cc68e97ad8cdf48e92dabbbd48a4075fb906767a6099484a9b6c720fdb05c"} Nov 28 14:15:37 crc kubenswrapper[4631]: I1128 14:15:37.641614 4631 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 28 14:15:39 crc kubenswrapper[4631]: I1128 14:15:39.662512 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vtcht" event={"ID":"69cad66f-388f-4703-8ed5-07cdc939dbef","Type":"ContainerStarted","Data":"f5daa50d4d93003620f76a0a4d52abc43ce0cf9be2d0cc6de2484049b33eedec"} Nov 28 14:15:41 crc kubenswrapper[4631]: I1128 14:15:41.513707 4631 scope.go:117] "RemoveContainer" containerID="d188114d99f65821ef24c4d713f07ed3cbe0cd10a5aaa23d6346374210035dc4" Nov 28 14:15:41 crc kubenswrapper[4631]: E1128 14:15:41.514280 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-47sxw_openshift-machine-config-operator(8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1)\"" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" podUID="8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1" Nov 28 14:15:42 crc kubenswrapper[4631]: I1128 14:15:42.575981 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-fpkw8_f9e35a8e-a1f5-40f9-b673-533ee2f1425f/kube-rbac-proxy/0.log" Nov 28 14:15:42 crc kubenswrapper[4631]: I1128 14:15:42.690791 4631 generic.go:334] "Generic (PLEG): container finished" podID="69cad66f-388f-4703-8ed5-07cdc939dbef" containerID="f5daa50d4d93003620f76a0a4d52abc43ce0cf9be2d0cc6de2484049b33eedec" exitCode=0 Nov 28 14:15:42 crc kubenswrapper[4631]: I1128 14:15:42.690853 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vtcht" event={"ID":"69cad66f-388f-4703-8ed5-07cdc939dbef","Type":"ContainerDied","Data":"f5daa50d4d93003620f76a0a4d52abc43ce0cf9be2d0cc6de2484049b33eedec"} Nov 28 14:15:42 crc kubenswrapper[4631]: I1128 14:15:42.783715 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-fpkw8_f9e35a8e-a1f5-40f9-b673-533ee2f1425f/controller/0.log" Nov 28 14:15:42 crc kubenswrapper[4631]: I1128 14:15:42.865998 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-5cpds_15c5a33d-b3c0-45aa-963f-b6d9aa639edd/cp-frr-files/0.log" Nov 28 14:15:43 crc kubenswrapper[4631]: I1128 14:15:43.114552 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-5cpds_15c5a33d-b3c0-45aa-963f-b6d9aa639edd/cp-reloader/0.log" Nov 28 14:15:43 crc kubenswrapper[4631]: I1128 14:15:43.126675 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-5cpds_15c5a33d-b3c0-45aa-963f-b6d9aa639edd/cp-frr-files/0.log" Nov 28 14:15:43 crc kubenswrapper[4631]: I1128 14:15:43.175552 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-5cpds_15c5a33d-b3c0-45aa-963f-b6d9aa639edd/cp-reloader/0.log" Nov 28 14:15:43 crc kubenswrapper[4631]: I1128 14:15:43.213326 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-5cpds_15c5a33d-b3c0-45aa-963f-b6d9aa639edd/cp-metrics/0.log" Nov 28 14:15:43 crc kubenswrapper[4631]: I1128 14:15:43.467930 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-5cpds_15c5a33d-b3c0-45aa-963f-b6d9aa639edd/cp-frr-files/0.log" Nov 28 14:15:43 crc kubenswrapper[4631]: I1128 14:15:43.583076 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-5cpds_15c5a33d-b3c0-45aa-963f-b6d9aa639edd/cp-metrics/0.log" Nov 28 14:15:43 crc kubenswrapper[4631]: I1128 14:15:43.648648 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-5cpds_15c5a33d-b3c0-45aa-963f-b6d9aa639edd/cp-metrics/0.log" Nov 28 14:15:43 crc kubenswrapper[4631]: I1128 14:15:43.649760 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-5cpds_15c5a33d-b3c0-45aa-963f-b6d9aa639edd/cp-reloader/0.log" Nov 28 14:15:43 crc kubenswrapper[4631]: I1128 14:15:43.926151 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-5cpds_15c5a33d-b3c0-45aa-963f-b6d9aa639edd/cp-reloader/0.log" Nov 28 14:15:43 crc kubenswrapper[4631]: I1128 14:15:43.988591 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-5cpds_15c5a33d-b3c0-45aa-963f-b6d9aa639edd/cp-frr-files/0.log" Nov 28 14:15:43 crc kubenswrapper[4631]: I1128 14:15:43.999507 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-5cpds_15c5a33d-b3c0-45aa-963f-b6d9aa639edd/cp-metrics/0.log" Nov 28 14:15:44 crc kubenswrapper[4631]: I1128 14:15:44.335960 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-5cpds_15c5a33d-b3c0-45aa-963f-b6d9aa639edd/controller/0.log" Nov 28 14:15:44 crc kubenswrapper[4631]: I1128 14:15:44.431172 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-5cpds_15c5a33d-b3c0-45aa-963f-b6d9aa639edd/kube-rbac-proxy/0.log" Nov 28 14:15:44 crc kubenswrapper[4631]: I1128 14:15:44.926511 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-5cpds_15c5a33d-b3c0-45aa-963f-b6d9aa639edd/frr-metrics/0.log" Nov 28 14:15:45 crc kubenswrapper[4631]: I1128 14:15:45.202224 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-5cpds_15c5a33d-b3c0-45aa-963f-b6d9aa639edd/reloader/0.log" Nov 28 14:15:45 crc kubenswrapper[4631]: I1128 14:15:45.249639 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-5cpds_15c5a33d-b3c0-45aa-963f-b6d9aa639edd/kube-rbac-proxy-frr/0.log" Nov 28 14:15:45 crc kubenswrapper[4631]: I1128 14:15:45.573406 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-7fcb986d4-9vxk7_020cecb2-380e-4c2a-8741-eee44e230abf/frr-k8s-webhook-server/0.log" Nov 28 14:15:45 crc kubenswrapper[4631]: I1128 14:15:45.752643 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vtcht" event={"ID":"69cad66f-388f-4703-8ed5-07cdc939dbef","Type":"ContainerStarted","Data":"3a4f6fd9f8cdf0f6f03cd61f2ca69af053811febae0f88166c745610d8bf3db9"} Nov 28 14:15:45 crc kubenswrapper[4631]: I1128 14:15:45.786263 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-vtcht" podStartSLOduration=3.71712766 podStartE2EDuration="10.786241538s" podCreationTimestamp="2025-11-28 14:15:35 +0000 UTC" firstStartedPulling="2025-11-28 14:15:37.64133354 +0000 UTC m=+3294.448636884" lastFinishedPulling="2025-11-28 14:15:44.710447418 +0000 UTC m=+3301.517750762" observedRunningTime="2025-11-28 14:15:45.779976884 +0000 UTC m=+3302.587280248" watchObservedRunningTime="2025-11-28 14:15:45.786241538 +0000 UTC m=+3302.593544882" Nov 28 14:15:46 crc kubenswrapper[4631]: I1128 14:15:46.000666 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-7c98947894-pw5mq_299f5184-243e-4fe0-b60e-2d40eee5a3f6/manager/0.log" Nov 28 14:15:46 crc kubenswrapper[4631]: I1128 14:15:46.078621 4631 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-vtcht" Nov 28 14:15:46 crc kubenswrapper[4631]: I1128 14:15:46.078692 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-vtcht" Nov 28 14:15:46 crc kubenswrapper[4631]: I1128 14:15:46.248368 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-5cpds_15c5a33d-b3c0-45aa-963f-b6d9aa639edd/frr/0.log" Nov 28 14:15:46 crc kubenswrapper[4631]: I1128 14:15:46.296372 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-fc587766c-d92lf_dc71ed9c-9a56-449a-afab-1a28a746fb2e/webhook-server/0.log" Nov 28 14:15:46 crc kubenswrapper[4631]: I1128 14:15:46.578526 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-d24rl_255213e3-20a3-4b0b-a165-c598ecce4766/kube-rbac-proxy/0.log" Nov 28 14:15:46 crc kubenswrapper[4631]: I1128 14:15:46.924081 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-d24rl_255213e3-20a3-4b0b-a165-c598ecce4766/speaker/0.log" Nov 28 14:15:47 crc kubenswrapper[4631]: I1128 14:15:47.170060 4631 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-vtcht" podUID="69cad66f-388f-4703-8ed5-07cdc939dbef" containerName="registry-server" probeResult="failure" output=< Nov 28 14:15:47 crc kubenswrapper[4631]: timeout: failed to connect service ":50051" within 1s Nov 28 14:15:47 crc kubenswrapper[4631]: > Nov 28 14:15:50 crc kubenswrapper[4631]: I1128 14:15:50.080390 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-4dwsx"] Nov 28 14:15:50 crc kubenswrapper[4631]: I1128 14:15:50.084123 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4dwsx" Nov 28 14:15:50 crc kubenswrapper[4631]: I1128 14:15:50.090462 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-4dwsx"] Nov 28 14:15:50 crc kubenswrapper[4631]: I1128 14:15:50.148424 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eb54d719-ef50-4f73-88e6-7645a7cce9fd-catalog-content\") pod \"community-operators-4dwsx\" (UID: \"eb54d719-ef50-4f73-88e6-7645a7cce9fd\") " pod="openshift-marketplace/community-operators-4dwsx" Nov 28 14:15:50 crc kubenswrapper[4631]: I1128 14:15:50.148612 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eb54d719-ef50-4f73-88e6-7645a7cce9fd-utilities\") pod \"community-operators-4dwsx\" (UID: \"eb54d719-ef50-4f73-88e6-7645a7cce9fd\") " pod="openshift-marketplace/community-operators-4dwsx" Nov 28 14:15:50 crc kubenswrapper[4631]: I1128 14:15:50.148809 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mcfnk\" (UniqueName: \"kubernetes.io/projected/eb54d719-ef50-4f73-88e6-7645a7cce9fd-kube-api-access-mcfnk\") pod \"community-operators-4dwsx\" (UID: \"eb54d719-ef50-4f73-88e6-7645a7cce9fd\") " pod="openshift-marketplace/community-operators-4dwsx" Nov 28 14:15:50 crc kubenswrapper[4631]: I1128 14:15:50.250875 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eb54d719-ef50-4f73-88e6-7645a7cce9fd-utilities\") pod \"community-operators-4dwsx\" (UID: \"eb54d719-ef50-4f73-88e6-7645a7cce9fd\") " pod="openshift-marketplace/community-operators-4dwsx" Nov 28 14:15:50 crc kubenswrapper[4631]: I1128 14:15:50.250984 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mcfnk\" (UniqueName: \"kubernetes.io/projected/eb54d719-ef50-4f73-88e6-7645a7cce9fd-kube-api-access-mcfnk\") pod \"community-operators-4dwsx\" (UID: \"eb54d719-ef50-4f73-88e6-7645a7cce9fd\") " pod="openshift-marketplace/community-operators-4dwsx" Nov 28 14:15:50 crc kubenswrapper[4631]: I1128 14:15:50.251027 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eb54d719-ef50-4f73-88e6-7645a7cce9fd-catalog-content\") pod \"community-operators-4dwsx\" (UID: \"eb54d719-ef50-4f73-88e6-7645a7cce9fd\") " pod="openshift-marketplace/community-operators-4dwsx" Nov 28 14:15:50 crc kubenswrapper[4631]: I1128 14:15:50.251659 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eb54d719-ef50-4f73-88e6-7645a7cce9fd-catalog-content\") pod \"community-operators-4dwsx\" (UID: \"eb54d719-ef50-4f73-88e6-7645a7cce9fd\") " pod="openshift-marketplace/community-operators-4dwsx" Nov 28 14:15:50 crc kubenswrapper[4631]: I1128 14:15:50.251906 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eb54d719-ef50-4f73-88e6-7645a7cce9fd-utilities\") pod \"community-operators-4dwsx\" (UID: \"eb54d719-ef50-4f73-88e6-7645a7cce9fd\") " pod="openshift-marketplace/community-operators-4dwsx" Nov 28 14:15:50 crc kubenswrapper[4631]: I1128 14:15:50.272516 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mcfnk\" (UniqueName: \"kubernetes.io/projected/eb54d719-ef50-4f73-88e6-7645a7cce9fd-kube-api-access-mcfnk\") pod \"community-operators-4dwsx\" (UID: \"eb54d719-ef50-4f73-88e6-7645a7cce9fd\") " pod="openshift-marketplace/community-operators-4dwsx" Nov 28 14:15:50 crc kubenswrapper[4631]: I1128 14:15:50.402228 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4dwsx" Nov 28 14:15:50 crc kubenswrapper[4631]: I1128 14:15:50.981317 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-4dwsx"] Nov 28 14:15:51 crc kubenswrapper[4631]: I1128 14:15:51.802604 4631 generic.go:334] "Generic (PLEG): container finished" podID="eb54d719-ef50-4f73-88e6-7645a7cce9fd" containerID="7d7113caf899ea27e1dc4f453dfc8bdb21d60096444e783466907fe4c6824d62" exitCode=0 Nov 28 14:15:51 crc kubenswrapper[4631]: I1128 14:15:51.802747 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4dwsx" event={"ID":"eb54d719-ef50-4f73-88e6-7645a7cce9fd","Type":"ContainerDied","Data":"7d7113caf899ea27e1dc4f453dfc8bdb21d60096444e783466907fe4c6824d62"} Nov 28 14:15:51 crc kubenswrapper[4631]: I1128 14:15:51.802887 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4dwsx" event={"ID":"eb54d719-ef50-4f73-88e6-7645a7cce9fd","Type":"ContainerStarted","Data":"f63920d7eacb5ba140497b1ef74a1943bc6d99e205d37f653332440adbb286c3"} Nov 28 14:15:52 crc kubenswrapper[4631]: I1128 14:15:52.818807 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4dwsx" event={"ID":"eb54d719-ef50-4f73-88e6-7645a7cce9fd","Type":"ContainerStarted","Data":"fe12fd977a8f84b1b6f77d108852bf44e9c220958ed5bc21a2005d673b875646"} Nov 28 14:15:53 crc kubenswrapper[4631]: I1128 14:15:53.521823 4631 scope.go:117] "RemoveContainer" containerID="d188114d99f65821ef24c4d713f07ed3cbe0cd10a5aaa23d6346374210035dc4" Nov 28 14:15:53 crc kubenswrapper[4631]: E1128 14:15:53.522139 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-47sxw_openshift-machine-config-operator(8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1)\"" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" podUID="8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1" Nov 28 14:15:53 crc kubenswrapper[4631]: I1128 14:15:53.829246 4631 generic.go:334] "Generic (PLEG): container finished" podID="eb54d719-ef50-4f73-88e6-7645a7cce9fd" containerID="fe12fd977a8f84b1b6f77d108852bf44e9c220958ed5bc21a2005d673b875646" exitCode=0 Nov 28 14:15:53 crc kubenswrapper[4631]: I1128 14:15:53.829309 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4dwsx" event={"ID":"eb54d719-ef50-4f73-88e6-7645a7cce9fd","Type":"ContainerDied","Data":"fe12fd977a8f84b1b6f77d108852bf44e9c220958ed5bc21a2005d673b875646"} Nov 28 14:15:54 crc kubenswrapper[4631]: I1128 14:15:54.845116 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4dwsx" event={"ID":"eb54d719-ef50-4f73-88e6-7645a7cce9fd","Type":"ContainerStarted","Data":"72d1a956cd4f2c02329558be2819307816ddd3261a337f52d09261220040f8ab"} Nov 28 14:15:54 crc kubenswrapper[4631]: I1128 14:15:54.883942 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-4dwsx" podStartSLOduration=2.250389745 podStartE2EDuration="4.883920947s" podCreationTimestamp="2025-11-28 14:15:50 +0000 UTC" firstStartedPulling="2025-11-28 14:15:51.804472114 +0000 UTC m=+3308.611775458" lastFinishedPulling="2025-11-28 14:15:54.438003316 +0000 UTC m=+3311.245306660" observedRunningTime="2025-11-28 14:15:54.873036761 +0000 UTC m=+3311.680340115" watchObservedRunningTime="2025-11-28 14:15:54.883920947 +0000 UTC m=+3311.691224291" Nov 28 14:15:57 crc kubenswrapper[4631]: I1128 14:15:57.126014 4631 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-vtcht" podUID="69cad66f-388f-4703-8ed5-07cdc939dbef" containerName="registry-server" probeResult="failure" output=< Nov 28 14:15:57 crc kubenswrapper[4631]: timeout: failed to connect service ":50051" within 1s Nov 28 14:15:57 crc kubenswrapper[4631]: > Nov 28 14:16:00 crc kubenswrapper[4631]: I1128 14:16:00.403132 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-4dwsx" Nov 28 14:16:00 crc kubenswrapper[4631]: I1128 14:16:00.403759 4631 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-4dwsx" Nov 28 14:16:00 crc kubenswrapper[4631]: I1128 14:16:00.454746 4631 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-4dwsx" Nov 28 14:16:00 crc kubenswrapper[4631]: I1128 14:16:00.949965 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-4dwsx" Nov 28 14:16:01 crc kubenswrapper[4631]: I1128 14:16:01.005967 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-4dwsx"] Nov 28 14:16:02 crc kubenswrapper[4631]: I1128 14:16:02.906592 4631 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-4dwsx" podUID="eb54d719-ef50-4f73-88e6-7645a7cce9fd" containerName="registry-server" containerID="cri-o://72d1a956cd4f2c02329558be2819307816ddd3261a337f52d09261220040f8ab" gracePeriod=2 Nov 28 14:16:03 crc kubenswrapper[4631]: I1128 14:16:03.926800 4631 generic.go:334] "Generic (PLEG): container finished" podID="eb54d719-ef50-4f73-88e6-7645a7cce9fd" containerID="72d1a956cd4f2c02329558be2819307816ddd3261a337f52d09261220040f8ab" exitCode=0 Nov 28 14:16:03 crc kubenswrapper[4631]: I1128 14:16:03.928220 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4dwsx" event={"ID":"eb54d719-ef50-4f73-88e6-7645a7cce9fd","Type":"ContainerDied","Data":"72d1a956cd4f2c02329558be2819307816ddd3261a337f52d09261220040f8ab"} Nov 28 14:16:04 crc kubenswrapper[4631]: I1128 14:16:04.035673 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4dwsx" Nov 28 14:16:04 crc kubenswrapper[4631]: I1128 14:16:04.103226 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mcfnk\" (UniqueName: \"kubernetes.io/projected/eb54d719-ef50-4f73-88e6-7645a7cce9fd-kube-api-access-mcfnk\") pod \"eb54d719-ef50-4f73-88e6-7645a7cce9fd\" (UID: \"eb54d719-ef50-4f73-88e6-7645a7cce9fd\") " Nov 28 14:16:04 crc kubenswrapper[4631]: I1128 14:16:04.103404 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eb54d719-ef50-4f73-88e6-7645a7cce9fd-utilities\") pod \"eb54d719-ef50-4f73-88e6-7645a7cce9fd\" (UID: \"eb54d719-ef50-4f73-88e6-7645a7cce9fd\") " Nov 28 14:16:04 crc kubenswrapper[4631]: I1128 14:16:04.103432 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eb54d719-ef50-4f73-88e6-7645a7cce9fd-catalog-content\") pod \"eb54d719-ef50-4f73-88e6-7645a7cce9fd\" (UID: \"eb54d719-ef50-4f73-88e6-7645a7cce9fd\") " Nov 28 14:16:04 crc kubenswrapper[4631]: I1128 14:16:04.107796 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/eb54d719-ef50-4f73-88e6-7645a7cce9fd-utilities" (OuterVolumeSpecName: "utilities") pod "eb54d719-ef50-4f73-88e6-7645a7cce9fd" (UID: "eb54d719-ef50-4f73-88e6-7645a7cce9fd"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 14:16:04 crc kubenswrapper[4631]: I1128 14:16:04.132537 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eb54d719-ef50-4f73-88e6-7645a7cce9fd-kube-api-access-mcfnk" (OuterVolumeSpecName: "kube-api-access-mcfnk") pod "eb54d719-ef50-4f73-88e6-7645a7cce9fd" (UID: "eb54d719-ef50-4f73-88e6-7645a7cce9fd"). InnerVolumeSpecName "kube-api-access-mcfnk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:16:04 crc kubenswrapper[4631]: I1128 14:16:04.163991 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/eb54d719-ef50-4f73-88e6-7645a7cce9fd-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "eb54d719-ef50-4f73-88e6-7645a7cce9fd" (UID: "eb54d719-ef50-4f73-88e6-7645a7cce9fd"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 14:16:04 crc kubenswrapper[4631]: I1128 14:16:04.205802 4631 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eb54d719-ef50-4f73-88e6-7645a7cce9fd-utilities\") on node \"crc\" DevicePath \"\"" Nov 28 14:16:04 crc kubenswrapper[4631]: I1128 14:16:04.206092 4631 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eb54d719-ef50-4f73-88e6-7645a7cce9fd-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 28 14:16:04 crc kubenswrapper[4631]: I1128 14:16:04.206110 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mcfnk\" (UniqueName: \"kubernetes.io/projected/eb54d719-ef50-4f73-88e6-7645a7cce9fd-kube-api-access-mcfnk\") on node \"crc\" DevicePath \"\"" Nov 28 14:16:04 crc kubenswrapper[4631]: I1128 14:16:04.278273 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f6xhhl_86f665f5-dfec-46bd-91aa-b07aa5803661/util/0.log" Nov 28 14:16:04 crc kubenswrapper[4631]: I1128 14:16:04.513158 4631 scope.go:117] "RemoveContainer" containerID="d188114d99f65821ef24c4d713f07ed3cbe0cd10a5aaa23d6346374210035dc4" Nov 28 14:16:04 crc kubenswrapper[4631]: E1128 14:16:04.513564 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-47sxw_openshift-machine-config-operator(8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1)\"" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" podUID="8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1" Nov 28 14:16:04 crc kubenswrapper[4631]: I1128 14:16:04.517535 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f6xhhl_86f665f5-dfec-46bd-91aa-b07aa5803661/util/0.log" Nov 28 14:16:04 crc kubenswrapper[4631]: I1128 14:16:04.574064 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f6xhhl_86f665f5-dfec-46bd-91aa-b07aa5803661/pull/0.log" Nov 28 14:16:04 crc kubenswrapper[4631]: I1128 14:16:04.657562 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f6xhhl_86f665f5-dfec-46bd-91aa-b07aa5803661/pull/0.log" Nov 28 14:16:04 crc kubenswrapper[4631]: I1128 14:16:04.903083 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f6xhhl_86f665f5-dfec-46bd-91aa-b07aa5803661/pull/0.log" Nov 28 14:16:04 crc kubenswrapper[4631]: I1128 14:16:04.909883 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f6xhhl_86f665f5-dfec-46bd-91aa-b07aa5803661/util/0.log" Nov 28 14:16:04 crc kubenswrapper[4631]: I1128 14:16:04.939432 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4dwsx" event={"ID":"eb54d719-ef50-4f73-88e6-7645a7cce9fd","Type":"ContainerDied","Data":"f63920d7eacb5ba140497b1ef74a1943bc6d99e205d37f653332440adbb286c3"} Nov 28 14:16:04 crc kubenswrapper[4631]: I1128 14:16:04.939480 4631 scope.go:117] "RemoveContainer" containerID="72d1a956cd4f2c02329558be2819307816ddd3261a337f52d09261220040f8ab" Nov 28 14:16:04 crc kubenswrapper[4631]: I1128 14:16:04.939519 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4dwsx" Nov 28 14:16:04 crc kubenswrapper[4631]: I1128 14:16:04.963258 4631 scope.go:117] "RemoveContainer" containerID="fe12fd977a8f84b1b6f77d108852bf44e9c220958ed5bc21a2005d673b875646" Nov 28 14:16:04 crc kubenswrapper[4631]: I1128 14:16:04.982891 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-4dwsx"] Nov 28 14:16:04 crc kubenswrapper[4631]: I1128 14:16:04.983053 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f6xhhl_86f665f5-dfec-46bd-91aa-b07aa5803661/extract/0.log" Nov 28 14:16:04 crc kubenswrapper[4631]: I1128 14:16:04.992008 4631 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-4dwsx"] Nov 28 14:16:05 crc kubenswrapper[4631]: I1128 14:16:05.002060 4631 scope.go:117] "RemoveContainer" containerID="7d7113caf899ea27e1dc4f453dfc8bdb21d60096444e783466907fe4c6824d62" Nov 28 14:16:05 crc kubenswrapper[4631]: I1128 14:16:05.213998 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83rfhlr_ddb177f8-0125-45bc-bbc4-130d689e10c9/util/0.log" Nov 28 14:16:05 crc kubenswrapper[4631]: I1128 14:16:05.456550 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83rfhlr_ddb177f8-0125-45bc-bbc4-130d689e10c9/util/0.log" Nov 28 14:16:05 crc kubenswrapper[4631]: I1128 14:16:05.492649 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83rfhlr_ddb177f8-0125-45bc-bbc4-130d689e10c9/pull/0.log" Nov 28 14:16:05 crc kubenswrapper[4631]: I1128 14:16:05.527491 4631 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eb54d719-ef50-4f73-88e6-7645a7cce9fd" path="/var/lib/kubelet/pods/eb54d719-ef50-4f73-88e6-7645a7cce9fd/volumes" Nov 28 14:16:05 crc kubenswrapper[4631]: I1128 14:16:05.536577 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83rfhlr_ddb177f8-0125-45bc-bbc4-130d689e10c9/pull/0.log" Nov 28 14:16:05 crc kubenswrapper[4631]: I1128 14:16:05.776515 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83rfhlr_ddb177f8-0125-45bc-bbc4-130d689e10c9/pull/0.log" Nov 28 14:16:05 crc kubenswrapper[4631]: I1128 14:16:05.825996 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83rfhlr_ddb177f8-0125-45bc-bbc4-130d689e10c9/extract/0.log" Nov 28 14:16:05 crc kubenswrapper[4631]: I1128 14:16:05.881187 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83rfhlr_ddb177f8-0125-45bc-bbc4-130d689e10c9/util/0.log" Nov 28 14:16:06 crc kubenswrapper[4631]: I1128 14:16:06.017367 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-j6d86_e55d89ce-6920-4cb5-951f-5b36cb73b6df/extract-utilities/0.log" Nov 28 14:16:06 crc kubenswrapper[4631]: I1128 14:16:06.134774 4631 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-vtcht" Nov 28 14:16:06 crc kubenswrapper[4631]: I1128 14:16:06.192193 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-vtcht" Nov 28 14:16:06 crc kubenswrapper[4631]: I1128 14:16:06.320399 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-j6d86_e55d89ce-6920-4cb5-951f-5b36cb73b6df/extract-content/0.log" Nov 28 14:16:06 crc kubenswrapper[4631]: I1128 14:16:06.363853 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-j6d86_e55d89ce-6920-4cb5-951f-5b36cb73b6df/extract-utilities/0.log" Nov 28 14:16:06 crc kubenswrapper[4631]: I1128 14:16:06.380098 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-j6d86_e55d89ce-6920-4cb5-951f-5b36cb73b6df/extract-content/0.log" Nov 28 14:16:06 crc kubenswrapper[4631]: I1128 14:16:06.524366 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-j6d86_e55d89ce-6920-4cb5-951f-5b36cb73b6df/extract-utilities/0.log" Nov 28 14:16:06 crc kubenswrapper[4631]: I1128 14:16:06.558536 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-j6d86_e55d89ce-6920-4cb5-951f-5b36cb73b6df/extract-content/0.log" Nov 28 14:16:07 crc kubenswrapper[4631]: I1128 14:16:07.018148 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-j6d86_e55d89ce-6920-4cb5-951f-5b36cb73b6df/registry-server/0.log" Nov 28 14:16:07 crc kubenswrapper[4631]: I1128 14:16:07.280058 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-vtcht"] Nov 28 14:16:07 crc kubenswrapper[4631]: I1128 14:16:07.624988 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-7hvgh_bbd0356c-2316-430e-81d3-c3c505c86e9e/extract-utilities/0.log" Nov 28 14:16:07 crc kubenswrapper[4631]: I1128 14:16:07.864404 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-7hvgh_bbd0356c-2316-430e-81d3-c3c505c86e9e/extract-utilities/0.log" Nov 28 14:16:07 crc kubenswrapper[4631]: I1128 14:16:07.913779 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-7hvgh_bbd0356c-2316-430e-81d3-c3c505c86e9e/extract-content/0.log" Nov 28 14:16:07 crc kubenswrapper[4631]: I1128 14:16:07.948516 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-7hvgh_bbd0356c-2316-430e-81d3-c3c505c86e9e/extract-content/0.log" Nov 28 14:16:07 crc kubenswrapper[4631]: I1128 14:16:07.981192 4631 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-vtcht" podUID="69cad66f-388f-4703-8ed5-07cdc939dbef" containerName="registry-server" containerID="cri-o://3a4f6fd9f8cdf0f6f03cd61f2ca69af053811febae0f88166c745610d8bf3db9" gracePeriod=2 Nov 28 14:16:08 crc kubenswrapper[4631]: I1128 14:16:08.241549 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-7hvgh_bbd0356c-2316-430e-81d3-c3c505c86e9e/extract-utilities/0.log" Nov 28 14:16:08 crc kubenswrapper[4631]: I1128 14:16:08.291399 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-7hvgh_bbd0356c-2316-430e-81d3-c3c505c86e9e/extract-content/0.log" Nov 28 14:16:08 crc kubenswrapper[4631]: I1128 14:16:08.619449 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-gsrzv_4eb86ab5-b771-49dc-8bfb-49ce46c0c712/marketplace-operator/0.log" Nov 28 14:16:08 crc kubenswrapper[4631]: I1128 14:16:08.796935 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-fsz47_abdc286b-6b18-47b2-b974-410a0f9614b9/extract-utilities/0.log" Nov 28 14:16:08 crc kubenswrapper[4631]: I1128 14:16:08.813895 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-7hvgh_bbd0356c-2316-430e-81d3-c3c505c86e9e/registry-server/0.log" Nov 28 14:16:09 crc kubenswrapper[4631]: I1128 14:16:09.013839 4631 generic.go:334] "Generic (PLEG): container finished" podID="69cad66f-388f-4703-8ed5-07cdc939dbef" containerID="3a4f6fd9f8cdf0f6f03cd61f2ca69af053811febae0f88166c745610d8bf3db9" exitCode=0 Nov 28 14:16:09 crc kubenswrapper[4631]: I1128 14:16:09.014213 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vtcht" event={"ID":"69cad66f-388f-4703-8ed5-07cdc939dbef","Type":"ContainerDied","Data":"3a4f6fd9f8cdf0f6f03cd61f2ca69af053811febae0f88166c745610d8bf3db9"} Nov 28 14:16:09 crc kubenswrapper[4631]: I1128 14:16:09.014329 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vtcht" event={"ID":"69cad66f-388f-4703-8ed5-07cdc939dbef","Type":"ContainerDied","Data":"e26cc68e97ad8cdf48e92dabbbd48a4075fb906767a6099484a9b6c720fdb05c"} Nov 28 14:16:09 crc kubenswrapper[4631]: I1128 14:16:09.014415 4631 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e26cc68e97ad8cdf48e92dabbbd48a4075fb906767a6099484a9b6c720fdb05c" Nov 28 14:16:09 crc kubenswrapper[4631]: I1128 14:16:09.053576 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-fsz47_abdc286b-6b18-47b2-b974-410a0f9614b9/extract-utilities/0.log" Nov 28 14:16:09 crc kubenswrapper[4631]: I1128 14:16:09.077540 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vtcht" Nov 28 14:16:09 crc kubenswrapper[4631]: I1128 14:16:09.154773 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-fsz47_abdc286b-6b18-47b2-b974-410a0f9614b9/extract-content/0.log" Nov 28 14:16:09 crc kubenswrapper[4631]: I1128 14:16:09.182019 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-fsz47_abdc286b-6b18-47b2-b974-410a0f9614b9/extract-content/0.log" Nov 28 14:16:09 crc kubenswrapper[4631]: I1128 14:16:09.201233 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/69cad66f-388f-4703-8ed5-07cdc939dbef-utilities\") pod \"69cad66f-388f-4703-8ed5-07cdc939dbef\" (UID: \"69cad66f-388f-4703-8ed5-07cdc939dbef\") " Nov 28 14:16:09 crc kubenswrapper[4631]: I1128 14:16:09.201307 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hqdkc\" (UniqueName: \"kubernetes.io/projected/69cad66f-388f-4703-8ed5-07cdc939dbef-kube-api-access-hqdkc\") pod \"69cad66f-388f-4703-8ed5-07cdc939dbef\" (UID: \"69cad66f-388f-4703-8ed5-07cdc939dbef\") " Nov 28 14:16:09 crc kubenswrapper[4631]: I1128 14:16:09.201451 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/69cad66f-388f-4703-8ed5-07cdc939dbef-catalog-content\") pod \"69cad66f-388f-4703-8ed5-07cdc939dbef\" (UID: \"69cad66f-388f-4703-8ed5-07cdc939dbef\") " Nov 28 14:16:09 crc kubenswrapper[4631]: I1128 14:16:09.201859 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/69cad66f-388f-4703-8ed5-07cdc939dbef-utilities" (OuterVolumeSpecName: "utilities") pod "69cad66f-388f-4703-8ed5-07cdc939dbef" (UID: "69cad66f-388f-4703-8ed5-07cdc939dbef"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 14:16:09 crc kubenswrapper[4631]: I1128 14:16:09.210602 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/69cad66f-388f-4703-8ed5-07cdc939dbef-kube-api-access-hqdkc" (OuterVolumeSpecName: "kube-api-access-hqdkc") pod "69cad66f-388f-4703-8ed5-07cdc939dbef" (UID: "69cad66f-388f-4703-8ed5-07cdc939dbef"). InnerVolumeSpecName "kube-api-access-hqdkc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:16:09 crc kubenswrapper[4631]: I1128 14:16:09.293014 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/69cad66f-388f-4703-8ed5-07cdc939dbef-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "69cad66f-388f-4703-8ed5-07cdc939dbef" (UID: "69cad66f-388f-4703-8ed5-07cdc939dbef"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 14:16:09 crc kubenswrapper[4631]: I1128 14:16:09.304072 4631 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/69cad66f-388f-4703-8ed5-07cdc939dbef-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 28 14:16:09 crc kubenswrapper[4631]: I1128 14:16:09.304133 4631 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/69cad66f-388f-4703-8ed5-07cdc939dbef-utilities\") on node \"crc\" DevicePath \"\"" Nov 28 14:16:09 crc kubenswrapper[4631]: I1128 14:16:09.304144 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hqdkc\" (UniqueName: \"kubernetes.io/projected/69cad66f-388f-4703-8ed5-07cdc939dbef-kube-api-access-hqdkc\") on node \"crc\" DevicePath \"\"" Nov 28 14:16:09 crc kubenswrapper[4631]: I1128 14:16:09.403072 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-fsz47_abdc286b-6b18-47b2-b974-410a0f9614b9/extract-utilities/0.log" Nov 28 14:16:09 crc kubenswrapper[4631]: I1128 14:16:09.421626 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-fsz47_abdc286b-6b18-47b2-b974-410a0f9614b9/extract-content/0.log" Nov 28 14:16:09 crc kubenswrapper[4631]: I1128 14:16:09.557263 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-fsz47_abdc286b-6b18-47b2-b974-410a0f9614b9/registry-server/0.log" Nov 28 14:16:09 crc kubenswrapper[4631]: I1128 14:16:09.623750 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-mzsr9_d2115a7c-dc8e-4f02-bded-25e080c7de51/extract-utilities/0.log" Nov 28 14:16:09 crc kubenswrapper[4631]: I1128 14:16:09.821770 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-mzsr9_d2115a7c-dc8e-4f02-bded-25e080c7de51/extract-utilities/0.log" Nov 28 14:16:09 crc kubenswrapper[4631]: I1128 14:16:09.834802 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-mzsr9_d2115a7c-dc8e-4f02-bded-25e080c7de51/extract-content/0.log" Nov 28 14:16:09 crc kubenswrapper[4631]: I1128 14:16:09.915787 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-mzsr9_d2115a7c-dc8e-4f02-bded-25e080c7de51/extract-content/0.log" Nov 28 14:16:10 crc kubenswrapper[4631]: I1128 14:16:10.022570 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vtcht" Nov 28 14:16:10 crc kubenswrapper[4631]: I1128 14:16:10.045268 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-vtcht"] Nov 28 14:16:10 crc kubenswrapper[4631]: I1128 14:16:10.053924 4631 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-vtcht"] Nov 28 14:16:10 crc kubenswrapper[4631]: I1128 14:16:10.115135 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-mzsr9_d2115a7c-dc8e-4f02-bded-25e080c7de51/extract-content/0.log" Nov 28 14:16:10 crc kubenswrapper[4631]: I1128 14:16:10.180260 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-mzsr9_d2115a7c-dc8e-4f02-bded-25e080c7de51/extract-utilities/0.log" Nov 28 14:16:10 crc kubenswrapper[4631]: I1128 14:16:10.500097 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-mzsr9_d2115a7c-dc8e-4f02-bded-25e080c7de51/registry-server/0.log" Nov 28 14:16:11 crc kubenswrapper[4631]: I1128 14:16:11.523280 4631 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="69cad66f-388f-4703-8ed5-07cdc939dbef" path="/var/lib/kubelet/pods/69cad66f-388f-4703-8ed5-07cdc939dbef/volumes" Nov 28 14:16:16 crc kubenswrapper[4631]: I1128 14:16:16.513057 4631 scope.go:117] "RemoveContainer" containerID="d188114d99f65821ef24c4d713f07ed3cbe0cd10a5aaa23d6346374210035dc4" Nov 28 14:16:16 crc kubenswrapper[4631]: E1128 14:16:16.513793 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-47sxw_openshift-machine-config-operator(8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1)\"" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" podUID="8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1" Nov 28 14:16:30 crc kubenswrapper[4631]: I1128 14:16:30.512985 4631 scope.go:117] "RemoveContainer" containerID="d188114d99f65821ef24c4d713f07ed3cbe0cd10a5aaa23d6346374210035dc4" Nov 28 14:16:31 crc kubenswrapper[4631]: I1128 14:16:31.578629 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" event={"ID":"8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1","Type":"ContainerStarted","Data":"a4c0198e8942036cb2b3c39a8f93d0e5314f141d5b92505f69b7cefcad77bdfc"} Nov 28 14:18:18 crc kubenswrapper[4631]: I1128 14:18:18.671982 4631 generic.go:334] "Generic (PLEG): container finished" podID="5796c446-01db-4372-aabb-dc74c82f687f" containerID="cc7e4d2bf296df0e3c26dab3ddf3ac2ff265c052cc5a9fe3a50621fcc68bc68d" exitCode=0 Nov 28 14:18:18 crc kubenswrapper[4631]: I1128 14:18:18.672069 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-rrfmx/must-gather-8h6ns" event={"ID":"5796c446-01db-4372-aabb-dc74c82f687f","Type":"ContainerDied","Data":"cc7e4d2bf296df0e3c26dab3ddf3ac2ff265c052cc5a9fe3a50621fcc68bc68d"} Nov 28 14:18:18 crc kubenswrapper[4631]: I1128 14:18:18.673217 4631 scope.go:117] "RemoveContainer" containerID="cc7e4d2bf296df0e3c26dab3ddf3ac2ff265c052cc5a9fe3a50621fcc68bc68d" Nov 28 14:18:19 crc kubenswrapper[4631]: I1128 14:18:19.377380 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-rrfmx_must-gather-8h6ns_5796c446-01db-4372-aabb-dc74c82f687f/gather/0.log" Nov 28 14:18:27 crc kubenswrapper[4631]: I1128 14:18:27.149186 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-rrfmx/must-gather-8h6ns"] Nov 28 14:18:27 crc kubenswrapper[4631]: I1128 14:18:27.150066 4631 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-rrfmx/must-gather-8h6ns" podUID="5796c446-01db-4372-aabb-dc74c82f687f" containerName="copy" containerID="cri-o://d648d8aa6f3d6d2803aa909de93b07f2613d8f0882da4a7ea4ec2ed32fd50ddc" gracePeriod=2 Nov 28 14:18:27 crc kubenswrapper[4631]: I1128 14:18:27.159943 4631 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-rrfmx/must-gather-8h6ns"] Nov 28 14:18:27 crc kubenswrapper[4631]: I1128 14:18:27.775136 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-rrfmx_must-gather-8h6ns_5796c446-01db-4372-aabb-dc74c82f687f/copy/0.log" Nov 28 14:18:27 crc kubenswrapper[4631]: I1128 14:18:27.776213 4631 generic.go:334] "Generic (PLEG): container finished" podID="5796c446-01db-4372-aabb-dc74c82f687f" containerID="d648d8aa6f3d6d2803aa909de93b07f2613d8f0882da4a7ea4ec2ed32fd50ddc" exitCode=143 Nov 28 14:18:28 crc kubenswrapper[4631]: I1128 14:18:28.173788 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-rrfmx_must-gather-8h6ns_5796c446-01db-4372-aabb-dc74c82f687f/copy/0.log" Nov 28 14:18:28 crc kubenswrapper[4631]: I1128 14:18:28.174532 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-rrfmx/must-gather-8h6ns" Nov 28 14:18:28 crc kubenswrapper[4631]: I1128 14:18:28.195831 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/5796c446-01db-4372-aabb-dc74c82f687f-must-gather-output\") pod \"5796c446-01db-4372-aabb-dc74c82f687f\" (UID: \"5796c446-01db-4372-aabb-dc74c82f687f\") " Nov 28 14:18:28 crc kubenswrapper[4631]: I1128 14:18:28.195879 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mh457\" (UniqueName: \"kubernetes.io/projected/5796c446-01db-4372-aabb-dc74c82f687f-kube-api-access-mh457\") pod \"5796c446-01db-4372-aabb-dc74c82f687f\" (UID: \"5796c446-01db-4372-aabb-dc74c82f687f\") " Nov 28 14:18:28 crc kubenswrapper[4631]: I1128 14:18:28.203553 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5796c446-01db-4372-aabb-dc74c82f687f-kube-api-access-mh457" (OuterVolumeSpecName: "kube-api-access-mh457") pod "5796c446-01db-4372-aabb-dc74c82f687f" (UID: "5796c446-01db-4372-aabb-dc74c82f687f"). InnerVolumeSpecName "kube-api-access-mh457". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:18:28 crc kubenswrapper[4631]: I1128 14:18:28.299117 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mh457\" (UniqueName: \"kubernetes.io/projected/5796c446-01db-4372-aabb-dc74c82f687f-kube-api-access-mh457\") on node \"crc\" DevicePath \"\"" Nov 28 14:18:28 crc kubenswrapper[4631]: I1128 14:18:28.377200 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5796c446-01db-4372-aabb-dc74c82f687f-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "5796c446-01db-4372-aabb-dc74c82f687f" (UID: "5796c446-01db-4372-aabb-dc74c82f687f"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 14:18:28 crc kubenswrapper[4631]: I1128 14:18:28.400832 4631 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/5796c446-01db-4372-aabb-dc74c82f687f-must-gather-output\") on node \"crc\" DevicePath \"\"" Nov 28 14:18:28 crc kubenswrapper[4631]: I1128 14:18:28.786716 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-rrfmx_must-gather-8h6ns_5796c446-01db-4372-aabb-dc74c82f687f/copy/0.log" Nov 28 14:18:28 crc kubenswrapper[4631]: I1128 14:18:28.787607 4631 scope.go:117] "RemoveContainer" containerID="d648d8aa6f3d6d2803aa909de93b07f2613d8f0882da4a7ea4ec2ed32fd50ddc" Nov 28 14:18:28 crc kubenswrapper[4631]: I1128 14:18:28.787694 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-rrfmx/must-gather-8h6ns" Nov 28 14:18:28 crc kubenswrapper[4631]: I1128 14:18:28.811606 4631 scope.go:117] "RemoveContainer" containerID="cc7e4d2bf296df0e3c26dab3ddf3ac2ff265c052cc5a9fe3a50621fcc68bc68d" Nov 28 14:18:29 crc kubenswrapper[4631]: I1128 14:18:29.526815 4631 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5796c446-01db-4372-aabb-dc74c82f687f" path="/var/lib/kubelet/pods/5796c446-01db-4372-aabb-dc74c82f687f/volumes" Nov 28 14:18:49 crc kubenswrapper[4631]: I1128 14:18:49.634471 4631 patch_prober.go:28] interesting pod/machine-config-daemon-47sxw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 28 14:18:49 crc kubenswrapper[4631]: I1128 14:18:49.635092 4631 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" podUID="8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 28 14:19:06 crc kubenswrapper[4631]: I1128 14:19:06.624994 4631 scope.go:117] "RemoveContainer" containerID="133fee652109f1922b9d4b283f025441889404feb95688bffb7daaa74d116c12" Nov 28 14:19:06 crc kubenswrapper[4631]: I1128 14:19:06.666961 4631 scope.go:117] "RemoveContainer" containerID="f2a2ea3d1162c39900b83bebf7e8625c87c5679080f92468ee16eb9d80d5da89" Nov 28 14:19:06 crc kubenswrapper[4631]: I1128 14:19:06.694198 4631 scope.go:117] "RemoveContainer" containerID="c9d1ab70b920fefed3d3887f6768848e237034125e663fbe672553a6ce20a9d2" Nov 28 14:19:19 crc kubenswrapper[4631]: I1128 14:19:19.635269 4631 patch_prober.go:28] interesting pod/machine-config-daemon-47sxw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 28 14:19:19 crc kubenswrapper[4631]: I1128 14:19:19.635908 4631 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" podUID="8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 28 14:19:49 crc kubenswrapper[4631]: I1128 14:19:49.634898 4631 patch_prober.go:28] interesting pod/machine-config-daemon-47sxw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 28 14:19:49 crc kubenswrapper[4631]: I1128 14:19:49.635554 4631 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" podUID="8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 28 14:19:49 crc kubenswrapper[4631]: I1128 14:19:49.635620 4631 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" Nov 28 14:19:49 crc kubenswrapper[4631]: I1128 14:19:49.636430 4631 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"a4c0198e8942036cb2b3c39a8f93d0e5314f141d5b92505f69b7cefcad77bdfc"} pod="openshift-machine-config-operator/machine-config-daemon-47sxw" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 28 14:19:49 crc kubenswrapper[4631]: I1128 14:19:49.636523 4631 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" podUID="8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1" containerName="machine-config-daemon" containerID="cri-o://a4c0198e8942036cb2b3c39a8f93d0e5314f141d5b92505f69b7cefcad77bdfc" gracePeriod=600 Nov 28 14:19:50 crc kubenswrapper[4631]: I1128 14:19:50.587962 4631 generic.go:334] "Generic (PLEG): container finished" podID="8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1" containerID="a4c0198e8942036cb2b3c39a8f93d0e5314f141d5b92505f69b7cefcad77bdfc" exitCode=0 Nov 28 14:19:50 crc kubenswrapper[4631]: I1128 14:19:50.588032 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" event={"ID":"8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1","Type":"ContainerDied","Data":"a4c0198e8942036cb2b3c39a8f93d0e5314f141d5b92505f69b7cefcad77bdfc"} Nov 28 14:19:50 crc kubenswrapper[4631]: I1128 14:19:50.588252 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" event={"ID":"8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1","Type":"ContainerStarted","Data":"b2f5e32c4eeb164aad43a3d1940c5d648cc8679d4f599da50345a14cc66f0979"} Nov 28 14:19:50 crc kubenswrapper[4631]: I1128 14:19:50.588282 4631 scope.go:117] "RemoveContainer" containerID="d188114d99f65821ef24c4d713f07ed3cbe0cd10a5aaa23d6346374210035dc4" Nov 28 14:20:57 crc kubenswrapper[4631]: I1128 14:20:57.741325 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-95xfg/must-gather-8gjjm"] Nov 28 14:20:57 crc kubenswrapper[4631]: E1128 14:20:57.742191 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5796c446-01db-4372-aabb-dc74c82f687f" containerName="gather" Nov 28 14:20:57 crc kubenswrapper[4631]: I1128 14:20:57.742206 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="5796c446-01db-4372-aabb-dc74c82f687f" containerName="gather" Nov 28 14:20:57 crc kubenswrapper[4631]: E1128 14:20:57.742215 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="69cad66f-388f-4703-8ed5-07cdc939dbef" containerName="extract-utilities" Nov 28 14:20:57 crc kubenswrapper[4631]: I1128 14:20:57.742221 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="69cad66f-388f-4703-8ed5-07cdc939dbef" containerName="extract-utilities" Nov 28 14:20:57 crc kubenswrapper[4631]: E1128 14:20:57.742242 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="69cad66f-388f-4703-8ed5-07cdc939dbef" containerName="extract-content" Nov 28 14:20:57 crc kubenswrapper[4631]: I1128 14:20:57.742248 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="69cad66f-388f-4703-8ed5-07cdc939dbef" containerName="extract-content" Nov 28 14:20:57 crc kubenswrapper[4631]: E1128 14:20:57.742257 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="69cad66f-388f-4703-8ed5-07cdc939dbef" containerName="registry-server" Nov 28 14:20:57 crc kubenswrapper[4631]: I1128 14:20:57.742262 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="69cad66f-388f-4703-8ed5-07cdc939dbef" containerName="registry-server" Nov 28 14:20:57 crc kubenswrapper[4631]: E1128 14:20:57.742275 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eb54d719-ef50-4f73-88e6-7645a7cce9fd" containerName="extract-utilities" Nov 28 14:20:57 crc kubenswrapper[4631]: I1128 14:20:57.742280 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="eb54d719-ef50-4f73-88e6-7645a7cce9fd" containerName="extract-utilities" Nov 28 14:20:57 crc kubenswrapper[4631]: E1128 14:20:57.742306 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eb54d719-ef50-4f73-88e6-7645a7cce9fd" containerName="registry-server" Nov 28 14:20:57 crc kubenswrapper[4631]: I1128 14:20:57.742312 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="eb54d719-ef50-4f73-88e6-7645a7cce9fd" containerName="registry-server" Nov 28 14:20:57 crc kubenswrapper[4631]: E1128 14:20:57.742331 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eb54d719-ef50-4f73-88e6-7645a7cce9fd" containerName="extract-content" Nov 28 14:20:57 crc kubenswrapper[4631]: I1128 14:20:57.742338 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="eb54d719-ef50-4f73-88e6-7645a7cce9fd" containerName="extract-content" Nov 28 14:20:57 crc kubenswrapper[4631]: E1128 14:20:57.742351 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5796c446-01db-4372-aabb-dc74c82f687f" containerName="copy" Nov 28 14:20:57 crc kubenswrapper[4631]: I1128 14:20:57.742357 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="5796c446-01db-4372-aabb-dc74c82f687f" containerName="copy" Nov 28 14:20:57 crc kubenswrapper[4631]: I1128 14:20:57.742539 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="eb54d719-ef50-4f73-88e6-7645a7cce9fd" containerName="registry-server" Nov 28 14:20:57 crc kubenswrapper[4631]: I1128 14:20:57.742558 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="5796c446-01db-4372-aabb-dc74c82f687f" containerName="copy" Nov 28 14:20:57 crc kubenswrapper[4631]: I1128 14:20:57.742573 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="5796c446-01db-4372-aabb-dc74c82f687f" containerName="gather" Nov 28 14:20:57 crc kubenswrapper[4631]: I1128 14:20:57.742586 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="69cad66f-388f-4703-8ed5-07cdc939dbef" containerName="registry-server" Nov 28 14:20:57 crc kubenswrapper[4631]: I1128 14:20:57.743562 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-95xfg/must-gather-8gjjm" Nov 28 14:20:57 crc kubenswrapper[4631]: I1128 14:20:57.745058 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-95xfg"/"kube-root-ca.crt" Nov 28 14:20:57 crc kubenswrapper[4631]: I1128 14:20:57.748306 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-95xfg"/"openshift-service-ca.crt" Nov 28 14:20:57 crc kubenswrapper[4631]: I1128 14:20:57.761250 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-95xfg/must-gather-8gjjm"] Nov 28 14:20:57 crc kubenswrapper[4631]: I1128 14:20:57.838200 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tq9qx\" (UniqueName: \"kubernetes.io/projected/a5418d8d-9e54-4825-b19e-22eed3133a61-kube-api-access-tq9qx\") pod \"must-gather-8gjjm\" (UID: \"a5418d8d-9e54-4825-b19e-22eed3133a61\") " pod="openshift-must-gather-95xfg/must-gather-8gjjm" Nov 28 14:20:57 crc kubenswrapper[4631]: I1128 14:20:57.838381 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/a5418d8d-9e54-4825-b19e-22eed3133a61-must-gather-output\") pod \"must-gather-8gjjm\" (UID: \"a5418d8d-9e54-4825-b19e-22eed3133a61\") " pod="openshift-must-gather-95xfg/must-gather-8gjjm" Nov 28 14:20:57 crc kubenswrapper[4631]: I1128 14:20:57.940109 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tq9qx\" (UniqueName: \"kubernetes.io/projected/a5418d8d-9e54-4825-b19e-22eed3133a61-kube-api-access-tq9qx\") pod \"must-gather-8gjjm\" (UID: \"a5418d8d-9e54-4825-b19e-22eed3133a61\") " pod="openshift-must-gather-95xfg/must-gather-8gjjm" Nov 28 14:20:57 crc kubenswrapper[4631]: I1128 14:20:57.940222 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/a5418d8d-9e54-4825-b19e-22eed3133a61-must-gather-output\") pod \"must-gather-8gjjm\" (UID: \"a5418d8d-9e54-4825-b19e-22eed3133a61\") " pod="openshift-must-gather-95xfg/must-gather-8gjjm" Nov 28 14:20:57 crc kubenswrapper[4631]: I1128 14:20:57.940642 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/a5418d8d-9e54-4825-b19e-22eed3133a61-must-gather-output\") pod \"must-gather-8gjjm\" (UID: \"a5418d8d-9e54-4825-b19e-22eed3133a61\") " pod="openshift-must-gather-95xfg/must-gather-8gjjm" Nov 28 14:20:57 crc kubenswrapper[4631]: I1128 14:20:57.979454 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tq9qx\" (UniqueName: \"kubernetes.io/projected/a5418d8d-9e54-4825-b19e-22eed3133a61-kube-api-access-tq9qx\") pod \"must-gather-8gjjm\" (UID: \"a5418d8d-9e54-4825-b19e-22eed3133a61\") " pod="openshift-must-gather-95xfg/must-gather-8gjjm" Nov 28 14:20:58 crc kubenswrapper[4631]: I1128 14:20:58.065583 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-95xfg/must-gather-8gjjm" Nov 28 14:20:58 crc kubenswrapper[4631]: I1128 14:20:58.618546 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-95xfg/must-gather-8gjjm"] Nov 28 14:20:59 crc kubenswrapper[4631]: I1128 14:20:59.255176 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-95xfg/must-gather-8gjjm" event={"ID":"a5418d8d-9e54-4825-b19e-22eed3133a61","Type":"ContainerStarted","Data":"7f767ef445a32dd22acba63789c626676072dc383b0e4417c654f46e21f680ef"} Nov 28 14:20:59 crc kubenswrapper[4631]: I1128 14:20:59.255533 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-95xfg/must-gather-8gjjm" event={"ID":"a5418d8d-9e54-4825-b19e-22eed3133a61","Type":"ContainerStarted","Data":"1afa63395df40354a7ce29c0fba3b0bc410af13cdc4292c6f37a5479f29e656e"} Nov 28 14:20:59 crc kubenswrapper[4631]: I1128 14:20:59.255554 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-95xfg/must-gather-8gjjm" event={"ID":"a5418d8d-9e54-4825-b19e-22eed3133a61","Type":"ContainerStarted","Data":"edec10c77c23d6bf60f0c5760d7a771e527786b3be12e81d679a05d17955b690"} Nov 28 14:20:59 crc kubenswrapper[4631]: I1128 14:20:59.298542 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-95xfg/must-gather-8gjjm" podStartSLOduration=2.298520075 podStartE2EDuration="2.298520075s" podCreationTimestamp="2025-11-28 14:20:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 14:20:59.290612881 +0000 UTC m=+3616.097916235" watchObservedRunningTime="2025-11-28 14:20:59.298520075 +0000 UTC m=+3616.105823419" Nov 28 14:21:02 crc kubenswrapper[4631]: I1128 14:21:02.731832 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-95xfg/crc-debug-mlt99"] Nov 28 14:21:02 crc kubenswrapper[4631]: I1128 14:21:02.733721 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-95xfg/crc-debug-mlt99" Nov 28 14:21:02 crc kubenswrapper[4631]: I1128 14:21:02.737243 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-95xfg"/"default-dockercfg-v99zh" Nov 28 14:21:02 crc kubenswrapper[4631]: I1128 14:21:02.865946 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/31a26cdc-f95f-4644-bef9-703e25a95e99-host\") pod \"crc-debug-mlt99\" (UID: \"31a26cdc-f95f-4644-bef9-703e25a95e99\") " pod="openshift-must-gather-95xfg/crc-debug-mlt99" Nov 28 14:21:02 crc kubenswrapper[4631]: I1128 14:21:02.866581 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5r7lh\" (UniqueName: \"kubernetes.io/projected/31a26cdc-f95f-4644-bef9-703e25a95e99-kube-api-access-5r7lh\") pod \"crc-debug-mlt99\" (UID: \"31a26cdc-f95f-4644-bef9-703e25a95e99\") " pod="openshift-must-gather-95xfg/crc-debug-mlt99" Nov 28 14:21:02 crc kubenswrapper[4631]: I1128 14:21:02.995237 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5r7lh\" (UniqueName: \"kubernetes.io/projected/31a26cdc-f95f-4644-bef9-703e25a95e99-kube-api-access-5r7lh\") pod \"crc-debug-mlt99\" (UID: \"31a26cdc-f95f-4644-bef9-703e25a95e99\") " pod="openshift-must-gather-95xfg/crc-debug-mlt99" Nov 28 14:21:02 crc kubenswrapper[4631]: I1128 14:21:02.995442 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/31a26cdc-f95f-4644-bef9-703e25a95e99-host\") pod \"crc-debug-mlt99\" (UID: \"31a26cdc-f95f-4644-bef9-703e25a95e99\") " pod="openshift-must-gather-95xfg/crc-debug-mlt99" Nov 28 14:21:02 crc kubenswrapper[4631]: I1128 14:21:02.995844 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/31a26cdc-f95f-4644-bef9-703e25a95e99-host\") pod \"crc-debug-mlt99\" (UID: \"31a26cdc-f95f-4644-bef9-703e25a95e99\") " pod="openshift-must-gather-95xfg/crc-debug-mlt99" Nov 28 14:21:03 crc kubenswrapper[4631]: I1128 14:21:03.016268 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5r7lh\" (UniqueName: \"kubernetes.io/projected/31a26cdc-f95f-4644-bef9-703e25a95e99-kube-api-access-5r7lh\") pod \"crc-debug-mlt99\" (UID: \"31a26cdc-f95f-4644-bef9-703e25a95e99\") " pod="openshift-must-gather-95xfg/crc-debug-mlt99" Nov 28 14:21:03 crc kubenswrapper[4631]: I1128 14:21:03.061073 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-95xfg/crc-debug-mlt99" Nov 28 14:21:03 crc kubenswrapper[4631]: W1128 14:21:03.101050 4631 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod31a26cdc_f95f_4644_bef9_703e25a95e99.slice/crio-7d7207cf0bfa6b800e379efa73b456d89d68d30a95ae028c2cacbd1fffb75dd9 WatchSource:0}: Error finding container 7d7207cf0bfa6b800e379efa73b456d89d68d30a95ae028c2cacbd1fffb75dd9: Status 404 returned error can't find the container with id 7d7207cf0bfa6b800e379efa73b456d89d68d30a95ae028c2cacbd1fffb75dd9 Nov 28 14:21:03 crc kubenswrapper[4631]: I1128 14:21:03.292982 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-95xfg/crc-debug-mlt99" event={"ID":"31a26cdc-f95f-4644-bef9-703e25a95e99","Type":"ContainerStarted","Data":"7d7207cf0bfa6b800e379efa73b456d89d68d30a95ae028c2cacbd1fffb75dd9"} Nov 28 14:21:04 crc kubenswrapper[4631]: I1128 14:21:04.301635 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-95xfg/crc-debug-mlt99" event={"ID":"31a26cdc-f95f-4644-bef9-703e25a95e99","Type":"ContainerStarted","Data":"24f6c67a028a1e168937286b3a24cfa1cde8b0d45198cc361e109ac977ce98f1"} Nov 28 14:21:38 crc kubenswrapper[4631]: I1128 14:21:38.611711 4631 generic.go:334] "Generic (PLEG): container finished" podID="31a26cdc-f95f-4644-bef9-703e25a95e99" containerID="24f6c67a028a1e168937286b3a24cfa1cde8b0d45198cc361e109ac977ce98f1" exitCode=0 Nov 28 14:21:38 crc kubenswrapper[4631]: I1128 14:21:38.611914 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-95xfg/crc-debug-mlt99" event={"ID":"31a26cdc-f95f-4644-bef9-703e25a95e99","Type":"ContainerDied","Data":"24f6c67a028a1e168937286b3a24cfa1cde8b0d45198cc361e109ac977ce98f1"} Nov 28 14:21:39 crc kubenswrapper[4631]: I1128 14:21:39.718401 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-95xfg/crc-debug-mlt99" Nov 28 14:21:39 crc kubenswrapper[4631]: I1128 14:21:39.753824 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-95xfg/crc-debug-mlt99"] Nov 28 14:21:39 crc kubenswrapper[4631]: I1128 14:21:39.763253 4631 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-95xfg/crc-debug-mlt99"] Nov 28 14:21:39 crc kubenswrapper[4631]: I1128 14:21:39.861004 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/31a26cdc-f95f-4644-bef9-703e25a95e99-host\") pod \"31a26cdc-f95f-4644-bef9-703e25a95e99\" (UID: \"31a26cdc-f95f-4644-bef9-703e25a95e99\") " Nov 28 14:21:39 crc kubenswrapper[4631]: I1128 14:21:39.861119 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/31a26cdc-f95f-4644-bef9-703e25a95e99-host" (OuterVolumeSpecName: "host") pod "31a26cdc-f95f-4644-bef9-703e25a95e99" (UID: "31a26cdc-f95f-4644-bef9-703e25a95e99"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 14:21:39 crc kubenswrapper[4631]: I1128 14:21:39.861143 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5r7lh\" (UniqueName: \"kubernetes.io/projected/31a26cdc-f95f-4644-bef9-703e25a95e99-kube-api-access-5r7lh\") pod \"31a26cdc-f95f-4644-bef9-703e25a95e99\" (UID: \"31a26cdc-f95f-4644-bef9-703e25a95e99\") " Nov 28 14:21:39 crc kubenswrapper[4631]: I1128 14:21:39.861470 4631 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/31a26cdc-f95f-4644-bef9-703e25a95e99-host\") on node \"crc\" DevicePath \"\"" Nov 28 14:21:39 crc kubenswrapper[4631]: I1128 14:21:39.876646 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31a26cdc-f95f-4644-bef9-703e25a95e99-kube-api-access-5r7lh" (OuterVolumeSpecName: "kube-api-access-5r7lh") pod "31a26cdc-f95f-4644-bef9-703e25a95e99" (UID: "31a26cdc-f95f-4644-bef9-703e25a95e99"). InnerVolumeSpecName "kube-api-access-5r7lh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:21:39 crc kubenswrapper[4631]: I1128 14:21:39.963389 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5r7lh\" (UniqueName: \"kubernetes.io/projected/31a26cdc-f95f-4644-bef9-703e25a95e99-kube-api-access-5r7lh\") on node \"crc\" DevicePath \"\"" Nov 28 14:21:40 crc kubenswrapper[4631]: I1128 14:21:40.631640 4631 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7d7207cf0bfa6b800e379efa73b456d89d68d30a95ae028c2cacbd1fffb75dd9" Nov 28 14:21:40 crc kubenswrapper[4631]: I1128 14:21:40.631724 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-95xfg/crc-debug-mlt99" Nov 28 14:21:41 crc kubenswrapper[4631]: I1128 14:21:41.000130 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-95xfg/crc-debug-tmjjj"] Nov 28 14:21:41 crc kubenswrapper[4631]: E1128 14:21:41.000859 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="31a26cdc-f95f-4644-bef9-703e25a95e99" containerName="container-00" Nov 28 14:21:41 crc kubenswrapper[4631]: I1128 14:21:41.000874 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="31a26cdc-f95f-4644-bef9-703e25a95e99" containerName="container-00" Nov 28 14:21:41 crc kubenswrapper[4631]: I1128 14:21:41.001045 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="31a26cdc-f95f-4644-bef9-703e25a95e99" containerName="container-00" Nov 28 14:21:41 crc kubenswrapper[4631]: I1128 14:21:41.001688 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-95xfg/crc-debug-tmjjj" Nov 28 14:21:41 crc kubenswrapper[4631]: I1128 14:21:41.004096 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-95xfg"/"default-dockercfg-v99zh" Nov 28 14:21:41 crc kubenswrapper[4631]: I1128 14:21:41.183865 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/7ee03796-fefb-4735-b9c8-ec4e3366d471-host\") pod \"crc-debug-tmjjj\" (UID: \"7ee03796-fefb-4735-b9c8-ec4e3366d471\") " pod="openshift-must-gather-95xfg/crc-debug-tmjjj" Nov 28 14:21:41 crc kubenswrapper[4631]: I1128 14:21:41.184273 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v5h5k\" (UniqueName: \"kubernetes.io/projected/7ee03796-fefb-4735-b9c8-ec4e3366d471-kube-api-access-v5h5k\") pod \"crc-debug-tmjjj\" (UID: \"7ee03796-fefb-4735-b9c8-ec4e3366d471\") " pod="openshift-must-gather-95xfg/crc-debug-tmjjj" Nov 28 14:21:41 crc kubenswrapper[4631]: I1128 14:21:41.285833 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v5h5k\" (UniqueName: \"kubernetes.io/projected/7ee03796-fefb-4735-b9c8-ec4e3366d471-kube-api-access-v5h5k\") pod \"crc-debug-tmjjj\" (UID: \"7ee03796-fefb-4735-b9c8-ec4e3366d471\") " pod="openshift-must-gather-95xfg/crc-debug-tmjjj" Nov 28 14:21:41 crc kubenswrapper[4631]: I1128 14:21:41.285964 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/7ee03796-fefb-4735-b9c8-ec4e3366d471-host\") pod \"crc-debug-tmjjj\" (UID: \"7ee03796-fefb-4735-b9c8-ec4e3366d471\") " pod="openshift-must-gather-95xfg/crc-debug-tmjjj" Nov 28 14:21:41 crc kubenswrapper[4631]: I1128 14:21:41.286238 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/7ee03796-fefb-4735-b9c8-ec4e3366d471-host\") pod \"crc-debug-tmjjj\" (UID: \"7ee03796-fefb-4735-b9c8-ec4e3366d471\") " pod="openshift-must-gather-95xfg/crc-debug-tmjjj" Nov 28 14:21:41 crc kubenswrapper[4631]: I1128 14:21:41.305925 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v5h5k\" (UniqueName: \"kubernetes.io/projected/7ee03796-fefb-4735-b9c8-ec4e3366d471-kube-api-access-v5h5k\") pod \"crc-debug-tmjjj\" (UID: \"7ee03796-fefb-4735-b9c8-ec4e3366d471\") " pod="openshift-must-gather-95xfg/crc-debug-tmjjj" Nov 28 14:21:41 crc kubenswrapper[4631]: I1128 14:21:41.317112 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-95xfg/crc-debug-tmjjj" Nov 28 14:21:41 crc kubenswrapper[4631]: I1128 14:21:41.523323 4631 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31a26cdc-f95f-4644-bef9-703e25a95e99" path="/var/lib/kubelet/pods/31a26cdc-f95f-4644-bef9-703e25a95e99/volumes" Nov 28 14:21:41 crc kubenswrapper[4631]: I1128 14:21:41.655760 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-95xfg/crc-debug-tmjjj" event={"ID":"7ee03796-fefb-4735-b9c8-ec4e3366d471","Type":"ContainerStarted","Data":"d126fbf9cc509f6cb7769088270e6e5c2c314b925751b64c35bd2f1307c962fa"} Nov 28 14:21:41 crc kubenswrapper[4631]: I1128 14:21:41.655803 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-95xfg/crc-debug-tmjjj" event={"ID":"7ee03796-fefb-4735-b9c8-ec4e3366d471","Type":"ContainerStarted","Data":"99444c4a096af39cce6c5fb3788d5a69712596f874661b9f15511640ac3d560a"} Nov 28 14:21:42 crc kubenswrapper[4631]: I1128 14:21:42.051874 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-95xfg/crc-debug-tmjjj"] Nov 28 14:21:42 crc kubenswrapper[4631]: I1128 14:21:42.066302 4631 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-95xfg/crc-debug-tmjjj"] Nov 28 14:21:42 crc kubenswrapper[4631]: I1128 14:21:42.665342 4631 generic.go:334] "Generic (PLEG): container finished" podID="7ee03796-fefb-4735-b9c8-ec4e3366d471" containerID="d126fbf9cc509f6cb7769088270e6e5c2c314b925751b64c35bd2f1307c962fa" exitCode=0 Nov 28 14:21:42 crc kubenswrapper[4631]: I1128 14:21:42.761141 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-95xfg/crc-debug-tmjjj" Nov 28 14:21:42 crc kubenswrapper[4631]: I1128 14:21:42.910349 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v5h5k\" (UniqueName: \"kubernetes.io/projected/7ee03796-fefb-4735-b9c8-ec4e3366d471-kube-api-access-v5h5k\") pod \"7ee03796-fefb-4735-b9c8-ec4e3366d471\" (UID: \"7ee03796-fefb-4735-b9c8-ec4e3366d471\") " Nov 28 14:21:42 crc kubenswrapper[4631]: I1128 14:21:42.910426 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/7ee03796-fefb-4735-b9c8-ec4e3366d471-host\") pod \"7ee03796-fefb-4735-b9c8-ec4e3366d471\" (UID: \"7ee03796-fefb-4735-b9c8-ec4e3366d471\") " Nov 28 14:21:42 crc kubenswrapper[4631]: I1128 14:21:42.910556 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7ee03796-fefb-4735-b9c8-ec4e3366d471-host" (OuterVolumeSpecName: "host") pod "7ee03796-fefb-4735-b9c8-ec4e3366d471" (UID: "7ee03796-fefb-4735-b9c8-ec4e3366d471"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 14:21:42 crc kubenswrapper[4631]: I1128 14:21:42.910966 4631 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/7ee03796-fefb-4735-b9c8-ec4e3366d471-host\") on node \"crc\" DevicePath \"\"" Nov 28 14:21:42 crc kubenswrapper[4631]: I1128 14:21:42.928659 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7ee03796-fefb-4735-b9c8-ec4e3366d471-kube-api-access-v5h5k" (OuterVolumeSpecName: "kube-api-access-v5h5k") pod "7ee03796-fefb-4735-b9c8-ec4e3366d471" (UID: "7ee03796-fefb-4735-b9c8-ec4e3366d471"). InnerVolumeSpecName "kube-api-access-v5h5k". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:21:43 crc kubenswrapper[4631]: I1128 14:21:43.012961 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v5h5k\" (UniqueName: \"kubernetes.io/projected/7ee03796-fefb-4735-b9c8-ec4e3366d471-kube-api-access-v5h5k\") on node \"crc\" DevicePath \"\"" Nov 28 14:21:43 crc kubenswrapper[4631]: I1128 14:21:43.464358 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-95xfg/crc-debug-djxgk"] Nov 28 14:21:43 crc kubenswrapper[4631]: E1128 14:21:43.464787 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ee03796-fefb-4735-b9c8-ec4e3366d471" containerName="container-00" Nov 28 14:21:43 crc kubenswrapper[4631]: I1128 14:21:43.464807 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ee03796-fefb-4735-b9c8-ec4e3366d471" containerName="container-00" Nov 28 14:21:43 crc kubenswrapper[4631]: I1128 14:21:43.464987 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="7ee03796-fefb-4735-b9c8-ec4e3366d471" containerName="container-00" Nov 28 14:21:43 crc kubenswrapper[4631]: I1128 14:21:43.465619 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-95xfg/crc-debug-djxgk" Nov 28 14:21:43 crc kubenswrapper[4631]: I1128 14:21:43.526536 4631 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7ee03796-fefb-4735-b9c8-ec4e3366d471" path="/var/lib/kubelet/pods/7ee03796-fefb-4735-b9c8-ec4e3366d471/volumes" Nov 28 14:21:43 crc kubenswrapper[4631]: I1128 14:21:43.622366 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b893f4f2-8903-4aaa-962a-a5a4532f781e-host\") pod \"crc-debug-djxgk\" (UID: \"b893f4f2-8903-4aaa-962a-a5a4532f781e\") " pod="openshift-must-gather-95xfg/crc-debug-djxgk" Nov 28 14:21:43 crc kubenswrapper[4631]: I1128 14:21:43.622665 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v9fkt\" (UniqueName: \"kubernetes.io/projected/b893f4f2-8903-4aaa-962a-a5a4532f781e-kube-api-access-v9fkt\") pod \"crc-debug-djxgk\" (UID: \"b893f4f2-8903-4aaa-962a-a5a4532f781e\") " pod="openshift-must-gather-95xfg/crc-debug-djxgk" Nov 28 14:21:43 crc kubenswrapper[4631]: I1128 14:21:43.674366 4631 scope.go:117] "RemoveContainer" containerID="d126fbf9cc509f6cb7769088270e6e5c2c314b925751b64c35bd2f1307c962fa" Nov 28 14:21:43 crc kubenswrapper[4631]: I1128 14:21:43.674519 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-95xfg/crc-debug-tmjjj" Nov 28 14:21:43 crc kubenswrapper[4631]: I1128 14:21:43.724213 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v9fkt\" (UniqueName: \"kubernetes.io/projected/b893f4f2-8903-4aaa-962a-a5a4532f781e-kube-api-access-v9fkt\") pod \"crc-debug-djxgk\" (UID: \"b893f4f2-8903-4aaa-962a-a5a4532f781e\") " pod="openshift-must-gather-95xfg/crc-debug-djxgk" Nov 28 14:21:43 crc kubenswrapper[4631]: I1128 14:21:43.724350 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b893f4f2-8903-4aaa-962a-a5a4532f781e-host\") pod \"crc-debug-djxgk\" (UID: \"b893f4f2-8903-4aaa-962a-a5a4532f781e\") " pod="openshift-must-gather-95xfg/crc-debug-djxgk" Nov 28 14:21:43 crc kubenswrapper[4631]: I1128 14:21:43.724482 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b893f4f2-8903-4aaa-962a-a5a4532f781e-host\") pod \"crc-debug-djxgk\" (UID: \"b893f4f2-8903-4aaa-962a-a5a4532f781e\") " pod="openshift-must-gather-95xfg/crc-debug-djxgk" Nov 28 14:21:43 crc kubenswrapper[4631]: I1128 14:21:43.741591 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v9fkt\" (UniqueName: \"kubernetes.io/projected/b893f4f2-8903-4aaa-962a-a5a4532f781e-kube-api-access-v9fkt\") pod \"crc-debug-djxgk\" (UID: \"b893f4f2-8903-4aaa-962a-a5a4532f781e\") " pod="openshift-must-gather-95xfg/crc-debug-djxgk" Nov 28 14:21:43 crc kubenswrapper[4631]: I1128 14:21:43.779870 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-95xfg/crc-debug-djxgk" Nov 28 14:21:43 crc kubenswrapper[4631]: W1128 14:21:43.838432 4631 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb893f4f2_8903_4aaa_962a_a5a4532f781e.slice/crio-798dc755ec235273d210f74a4f8894d0e9cc18416388a56eb4de4fdff01bc4c5 WatchSource:0}: Error finding container 798dc755ec235273d210f74a4f8894d0e9cc18416388a56eb4de4fdff01bc4c5: Status 404 returned error can't find the container with id 798dc755ec235273d210f74a4f8894d0e9cc18416388a56eb4de4fdff01bc4c5 Nov 28 14:21:44 crc kubenswrapper[4631]: E1128 14:21:44.353999 4631 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb893f4f2_8903_4aaa_962a_a5a4532f781e.slice/crio-conmon-7cb3e8d59173b433caaefde9ffee610ce323073748c19b4cd00134c9aeb5f473.scope\": RecentStats: unable to find data in memory cache]" Nov 28 14:21:44 crc kubenswrapper[4631]: I1128 14:21:44.686240 4631 generic.go:334] "Generic (PLEG): container finished" podID="b893f4f2-8903-4aaa-962a-a5a4532f781e" containerID="7cb3e8d59173b433caaefde9ffee610ce323073748c19b4cd00134c9aeb5f473" exitCode=0 Nov 28 14:21:44 crc kubenswrapper[4631]: I1128 14:21:44.686443 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-95xfg/crc-debug-djxgk" event={"ID":"b893f4f2-8903-4aaa-962a-a5a4532f781e","Type":"ContainerDied","Data":"7cb3e8d59173b433caaefde9ffee610ce323073748c19b4cd00134c9aeb5f473"} Nov 28 14:21:44 crc kubenswrapper[4631]: I1128 14:21:44.686592 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-95xfg/crc-debug-djxgk" event={"ID":"b893f4f2-8903-4aaa-962a-a5a4532f781e","Type":"ContainerStarted","Data":"798dc755ec235273d210f74a4f8894d0e9cc18416388a56eb4de4fdff01bc4c5"} Nov 28 14:21:44 crc kubenswrapper[4631]: I1128 14:21:44.733879 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-95xfg/crc-debug-djxgk"] Nov 28 14:21:44 crc kubenswrapper[4631]: I1128 14:21:44.741740 4631 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-95xfg/crc-debug-djxgk"] Nov 28 14:21:45 crc kubenswrapper[4631]: I1128 14:21:45.816405 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-95xfg/crc-debug-djxgk" Nov 28 14:21:45 crc kubenswrapper[4631]: I1128 14:21:45.961255 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v9fkt\" (UniqueName: \"kubernetes.io/projected/b893f4f2-8903-4aaa-962a-a5a4532f781e-kube-api-access-v9fkt\") pod \"b893f4f2-8903-4aaa-962a-a5a4532f781e\" (UID: \"b893f4f2-8903-4aaa-962a-a5a4532f781e\") " Nov 28 14:21:45 crc kubenswrapper[4631]: I1128 14:21:45.961650 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b893f4f2-8903-4aaa-962a-a5a4532f781e-host\") pod \"b893f4f2-8903-4aaa-962a-a5a4532f781e\" (UID: \"b893f4f2-8903-4aaa-962a-a5a4532f781e\") " Nov 28 14:21:45 crc kubenswrapper[4631]: I1128 14:21:45.961768 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b893f4f2-8903-4aaa-962a-a5a4532f781e-host" (OuterVolumeSpecName: "host") pod "b893f4f2-8903-4aaa-962a-a5a4532f781e" (UID: "b893f4f2-8903-4aaa-962a-a5a4532f781e"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 14:21:45 crc kubenswrapper[4631]: I1128 14:21:45.962610 4631 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b893f4f2-8903-4aaa-962a-a5a4532f781e-host\") on node \"crc\" DevicePath \"\"" Nov 28 14:21:45 crc kubenswrapper[4631]: I1128 14:21:45.967038 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b893f4f2-8903-4aaa-962a-a5a4532f781e-kube-api-access-v9fkt" (OuterVolumeSpecName: "kube-api-access-v9fkt") pod "b893f4f2-8903-4aaa-962a-a5a4532f781e" (UID: "b893f4f2-8903-4aaa-962a-a5a4532f781e"). InnerVolumeSpecName "kube-api-access-v9fkt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:21:46 crc kubenswrapper[4631]: I1128 14:21:46.064412 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v9fkt\" (UniqueName: \"kubernetes.io/projected/b893f4f2-8903-4aaa-962a-a5a4532f781e-kube-api-access-v9fkt\") on node \"crc\" DevicePath \"\"" Nov 28 14:21:46 crc kubenswrapper[4631]: I1128 14:21:46.714562 4631 scope.go:117] "RemoveContainer" containerID="7cb3e8d59173b433caaefde9ffee610ce323073748c19b4cd00134c9aeb5f473" Nov 28 14:21:46 crc kubenswrapper[4631]: I1128 14:21:46.714642 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-95xfg/crc-debug-djxgk" Nov 28 14:21:47 crc kubenswrapper[4631]: I1128 14:21:47.526525 4631 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b893f4f2-8903-4aaa-962a-a5a4532f781e" path="/var/lib/kubelet/pods/b893f4f2-8903-4aaa-962a-a5a4532f781e/volumes" Nov 28 14:22:06 crc kubenswrapper[4631]: I1128 14:22:06.862566 4631 scope.go:117] "RemoveContainer" containerID="3a4f6fd9f8cdf0f6f03cd61f2ca69af053811febae0f88166c745610d8bf3db9" Nov 28 14:22:06 crc kubenswrapper[4631]: I1128 14:22:06.881710 4631 scope.go:117] "RemoveContainer" containerID="11f55df99c87535f98013d4bda986e135dc15f20fe56a06b1c95207fd3df977e" Nov 28 14:22:06 crc kubenswrapper[4631]: I1128 14:22:06.904981 4631 scope.go:117] "RemoveContainer" containerID="f5daa50d4d93003620f76a0a4d52abc43ce0cf9be2d0cc6de2484049b33eedec" Nov 28 14:22:19 crc kubenswrapper[4631]: I1128 14:22:19.634222 4631 patch_prober.go:28] interesting pod/machine-config-daemon-47sxw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 28 14:22:19 crc kubenswrapper[4631]: I1128 14:22:19.634838 4631 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" podUID="8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 28 14:22:22 crc kubenswrapper[4631]: I1128 14:22:22.368875 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-54885d4b84-55wlk_1a3cfa34-17ae-42d6-8ea7-46691bb10eaa/barbican-api/0.log" Nov 28 14:22:22 crc kubenswrapper[4631]: I1128 14:22:22.459999 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-54885d4b84-55wlk_1a3cfa34-17ae-42d6-8ea7-46691bb10eaa/barbican-api-log/0.log" Nov 28 14:22:22 crc kubenswrapper[4631]: I1128 14:22:22.648514 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-59c4c4dc74-qt87t_177f3e63-a6ce-4512-80b3-f229e3ace62a/barbican-keystone-listener/0.log" Nov 28 14:22:22 crc kubenswrapper[4631]: I1128 14:22:22.783966 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-5b8f56c4bc-cntw9_1b8b4a3c-2e24-4a9b-b18d-0fa09f70a155/barbican-worker/0.log" Nov 28 14:22:22 crc kubenswrapper[4631]: I1128 14:22:22.889221 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-59c4c4dc74-qt87t_177f3e63-a6ce-4512-80b3-f229e3ace62a/barbican-keystone-listener-log/0.log" Nov 28 14:22:22 crc kubenswrapper[4631]: I1128 14:22:22.919255 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-5b8f56c4bc-cntw9_1b8b4a3c-2e24-4a9b-b18d-0fa09f70a155/barbican-worker-log/0.log" Nov 28 14:22:23 crc kubenswrapper[4631]: I1128 14:22:23.155802 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-xz7jm_44ae011d-6420-4400-9c7c-21463e3c82b1/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Nov 28 14:22:23 crc kubenswrapper[4631]: I1128 14:22:23.198609 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_4cde0c05-4a6a-410c-b5c3-2f14289fc37a/ceilometer-central-agent/0.log" Nov 28 14:22:23 crc kubenswrapper[4631]: I1128 14:22:23.373305 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_4cde0c05-4a6a-410c-b5c3-2f14289fc37a/ceilometer-notification-agent/0.log" Nov 28 14:22:23 crc kubenswrapper[4631]: I1128 14:22:23.429060 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_4cde0c05-4a6a-410c-b5c3-2f14289fc37a/proxy-httpd/0.log" Nov 28 14:22:23 crc kubenswrapper[4631]: I1128 14:22:23.457408 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_4cde0c05-4a6a-410c-b5c3-2f14289fc37a/sg-core/0.log" Nov 28 14:22:23 crc kubenswrapper[4631]: I1128 14:22:23.980947 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_86f5dab6-6b28-494c-b9d6-52528c726875/cinder-api-log/0.log" Nov 28 14:22:24 crc kubenswrapper[4631]: I1128 14:22:24.064089 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_86f5dab6-6b28-494c-b9d6-52528c726875/cinder-api/0.log" Nov 28 14:22:24 crc kubenswrapper[4631]: I1128 14:22:24.124769 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_c4be14a0-f1a1-46f3-86ea-48e7dfb8da00/cinder-scheduler/0.log" Nov 28 14:22:24 crc kubenswrapper[4631]: I1128 14:22:24.367883 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_c4be14a0-f1a1-46f3-86ea-48e7dfb8da00/probe/0.log" Nov 28 14:22:24 crc kubenswrapper[4631]: I1128 14:22:24.478592 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-4lxnq_4d2450a3-e851-4bfb-83c1-115334898784/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Nov 28 14:22:24 crc kubenswrapper[4631]: I1128 14:22:24.696989 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-s9jts_8133b1ca-bdb2-46e1-8d39-2b45ab9284c8/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Nov 28 14:22:24 crc kubenswrapper[4631]: I1128 14:22:24.810043 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-6cd9bffc9-x8stz_6b64d173-0e5b-41e3-b841-68fef937dfb4/init/0.log" Nov 28 14:22:24 crc kubenswrapper[4631]: I1128 14:22:24.965125 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-6cd9bffc9-x8stz_6b64d173-0e5b-41e3-b841-68fef937dfb4/init/0.log" Nov 28 14:22:25 crc kubenswrapper[4631]: I1128 14:22:25.052413 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-2j6xr_2a23172b-a4cc-41e3-9464-2b3e7f7a51f8/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Nov 28 14:22:25 crc kubenswrapper[4631]: I1128 14:22:25.146479 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-6cd9bffc9-x8stz_6b64d173-0e5b-41e3-b841-68fef937dfb4/dnsmasq-dns/0.log" Nov 28 14:22:25 crc kubenswrapper[4631]: I1128 14:22:25.299503 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_aeea572b-1f4e-486b-a916-1b34f0b84d10/glance-log/0.log" Nov 28 14:22:25 crc kubenswrapper[4631]: I1128 14:22:25.450843 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_aeea572b-1f4e-486b-a916-1b34f0b84d10/glance-httpd/0.log" Nov 28 14:22:25 crc kubenswrapper[4631]: I1128 14:22:25.598002 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_b3727bc6-2743-41bf-9a94-45e4921eeb00/glance-httpd/0.log" Nov 28 14:22:25 crc kubenswrapper[4631]: I1128 14:22:25.676731 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_b3727bc6-2743-41bf-9a94-45e4921eeb00/glance-log/0.log" Nov 28 14:22:25 crc kubenswrapper[4631]: I1128 14:22:25.857391 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-544496c768-rzhsw_cd607dae-568a-4b81-af81-3310c0e95854/horizon/2.log" Nov 28 14:22:26 crc kubenswrapper[4631]: I1128 14:22:26.073073 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-544496c768-rzhsw_cd607dae-568a-4b81-af81-3310c0e95854/horizon/1.log" Nov 28 14:22:26 crc kubenswrapper[4631]: I1128 14:22:26.178849 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-c6hq5_4f7fda9a-fd1f-4485-a838-f5fee5a83b6e/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Nov 28 14:22:26 crc kubenswrapper[4631]: I1128 14:22:26.298030 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-544496c768-rzhsw_cd607dae-568a-4b81-af81-3310c0e95854/horizon-log/0.log" Nov 28 14:22:26 crc kubenswrapper[4631]: I1128 14:22:26.390735 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-gsp5r_5f971cb1-7c00-4dbe-abb1-e4eeb825fcee/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Nov 28 14:22:26 crc kubenswrapper[4631]: I1128 14:22:26.642143 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29405641-ngxl6_2080b2bc-31ce-4061-abc1-6f044c10355b/keystone-cron/0.log" Nov 28 14:22:26 crc kubenswrapper[4631]: I1128 14:22:26.808936 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-6bd6cfc878-65bg5_2adae216-4db3-49d7-8d11-24bbd7d47f8f/keystone-api/0.log" Nov 28 14:22:26 crc kubenswrapper[4631]: I1128 14:22:26.907466 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_c1f85a30-128c-487c-ac41-f1d7a64a52a1/kube-state-metrics/0.log" Nov 28 14:22:26 crc kubenswrapper[4631]: I1128 14:22:26.993937 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-pw8wp_86dddd7e-e065-4593-8d47-4d36a2cddb22/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Nov 28 14:22:27 crc kubenswrapper[4631]: I1128 14:22:27.294701 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-7f7bb9bf8f-42ssw_d05d0efc-381c-4222-8c8c-384c2c22ac9c/neutron-httpd/0.log" Nov 28 14:22:27 crc kubenswrapper[4631]: I1128 14:22:27.308766 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-7f7bb9bf8f-42ssw_d05d0efc-381c-4222-8c8c-384c2c22ac9c/neutron-api/0.log" Nov 28 14:22:27 crc kubenswrapper[4631]: I1128 14:22:27.651534 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-tl8jt_a73f7b13-ead0-46a0-8c62-d3c1e75bc2d8/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Nov 28 14:22:27 crc kubenswrapper[4631]: I1128 14:22:27.869998 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_9db3c70b-08ec-4b40-bdb9-94cdbc66fa1d/nova-api-log/0.log" Nov 28 14:22:28 crc kubenswrapper[4631]: I1128 14:22:28.134974 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_cc2a0580-660d-4a19-a755-70256a4e4abe/nova-cell0-conductor-conductor/0.log" Nov 28 14:22:28 crc kubenswrapper[4631]: I1128 14:22:28.182725 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_9db3c70b-08ec-4b40-bdb9-94cdbc66fa1d/nova-api-api/0.log" Nov 28 14:22:28 crc kubenswrapper[4631]: I1128 14:22:28.377747 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_0db994a3-5269-44c5-8a6f-8b7fdc798c3d/nova-cell1-conductor-conductor/0.log" Nov 28 14:22:28 crc kubenswrapper[4631]: I1128 14:22:28.667078 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_9c3a5a08-9590-4249-be26-8224d51da976/nova-cell1-novncproxy-novncproxy/0.log" Nov 28 14:22:28 crc kubenswrapper[4631]: I1128 14:22:28.775799 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-5hrhp_7692b0ac-e13d-4413-a2c0-e72062ec9110/nova-edpm-deployment-openstack-edpm-ipam/0.log" Nov 28 14:22:29 crc kubenswrapper[4631]: I1128 14:22:29.010351 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_52191ac4-b50d-4c9a-8a42-b3afe3cffe7f/nova-metadata-log/0.log" Nov 28 14:22:29 crc kubenswrapper[4631]: I1128 14:22:29.361202 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_83514cf7-b36a-4691-b5f9-559a1594044a/nova-scheduler-scheduler/0.log" Nov 28 14:22:29 crc kubenswrapper[4631]: I1128 14:22:29.554006 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_4bc20a10-95d5-4003-86d9-c6848bc1b188/mysql-bootstrap/0.log" Nov 28 14:22:29 crc kubenswrapper[4631]: I1128 14:22:29.832917 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_4bc20a10-95d5-4003-86d9-c6848bc1b188/mysql-bootstrap/0.log" Nov 28 14:22:29 crc kubenswrapper[4631]: I1128 14:22:29.839663 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_4bc20a10-95d5-4003-86d9-c6848bc1b188/galera/0.log" Nov 28 14:22:30 crc kubenswrapper[4631]: I1128 14:22:30.176031 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_1a711f62-44fc-4046-a3aa-b5ef77b45e62/mysql-bootstrap/0.log" Nov 28 14:22:30 crc kubenswrapper[4631]: I1128 14:22:30.348650 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_1a711f62-44fc-4046-a3aa-b5ef77b45e62/mysql-bootstrap/0.log" Nov 28 14:22:30 crc kubenswrapper[4631]: I1128 14:22:30.398460 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_1a711f62-44fc-4046-a3aa-b5ef77b45e62/galera/0.log" Nov 28 14:22:30 crc kubenswrapper[4631]: I1128 14:22:30.416407 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_52191ac4-b50d-4c9a-8a42-b3afe3cffe7f/nova-metadata-metadata/0.log" Nov 28 14:22:30 crc kubenswrapper[4631]: I1128 14:22:30.649601 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_ea063a3c-330a-4642-a815-c013ae6b80c7/openstackclient/0.log" Nov 28 14:22:30 crc kubenswrapper[4631]: I1128 14:22:30.683477 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-7jgdf_7fc5dcd5-fde5-4a72-8d4a-cda51785e77f/ovn-controller/0.log" Nov 28 14:22:30 crc kubenswrapper[4631]: I1128 14:22:30.957834 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-p7l62_6dfe4f1a-897a-4ece-a345-4a4f86901079/openstack-network-exporter/0.log" Nov 28 14:22:30 crc kubenswrapper[4631]: I1128 14:22:30.975027 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-lskqw_7fc0ee4e-435d-4297-94eb-9e8c6fa90004/ovsdb-server-init/0.log" Nov 28 14:22:31 crc kubenswrapper[4631]: I1128 14:22:31.275078 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-lskqw_7fc0ee4e-435d-4297-94eb-9e8c6fa90004/ovsdb-server-init/0.log" Nov 28 14:22:31 crc kubenswrapper[4631]: I1128 14:22:31.297155 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-lskqw_7fc0ee4e-435d-4297-94eb-9e8c6fa90004/ovsdb-server/0.log" Nov 28 14:22:31 crc kubenswrapper[4631]: I1128 14:22:31.401269 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-lskqw_7fc0ee4e-435d-4297-94eb-9e8c6fa90004/ovs-vswitchd/0.log" Nov 28 14:22:31 crc kubenswrapper[4631]: I1128 14:22:31.562406 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-2sls9_c09a85dc-b780-41c9-8122-653d4703aa7e/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Nov 28 14:22:31 crc kubenswrapper[4631]: I1128 14:22:31.716849 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_1418c5fe-bc14-4aeb-9449-30c85e26aed2/openstack-network-exporter/0.log" Nov 28 14:22:31 crc kubenswrapper[4631]: I1128 14:22:31.888319 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_ca11061d-b0a1-464d-9e1d-5572fc439580/openstack-network-exporter/0.log" Nov 28 14:22:31 crc kubenswrapper[4631]: I1128 14:22:31.950126 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_1418c5fe-bc14-4aeb-9449-30c85e26aed2/ovn-northd/0.log" Nov 28 14:22:32 crc kubenswrapper[4631]: I1128 14:22:32.108567 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_ca11061d-b0a1-464d-9e1d-5572fc439580/ovsdbserver-nb/0.log" Nov 28 14:22:32 crc kubenswrapper[4631]: I1128 14:22:32.211935 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_1d3ba54c-cb26-4e04-8310-34e6b4206119/openstack-network-exporter/0.log" Nov 28 14:22:32 crc kubenswrapper[4631]: I1128 14:22:32.248615 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_1d3ba54c-cb26-4e04-8310-34e6b4206119/ovsdbserver-sb/0.log" Nov 28 14:22:32 crc kubenswrapper[4631]: I1128 14:22:32.648912 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-769746444-8zhp8_176c2dec-4940-449b-ab1a-95cb11cdfeff/placement-log/0.log" Nov 28 14:22:32 crc kubenswrapper[4631]: I1128 14:22:32.719412 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-769746444-8zhp8_176c2dec-4940-449b-ab1a-95cb11cdfeff/placement-api/0.log" Nov 28 14:22:32 crc kubenswrapper[4631]: I1128 14:22:32.852176 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_014519bc-68f0-4132-9a46-6b0e5c538c7e/setup-container/0.log" Nov 28 14:22:33 crc kubenswrapper[4631]: I1128 14:22:33.084106 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_014519bc-68f0-4132-9a46-6b0e5c538c7e/setup-container/0.log" Nov 28 14:22:33 crc kubenswrapper[4631]: I1128 14:22:33.096781 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_014519bc-68f0-4132-9a46-6b0e5c538c7e/rabbitmq/0.log" Nov 28 14:22:33 crc kubenswrapper[4631]: I1128 14:22:33.276914 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_fc5706c2-8b7d-439a-95b3-981907360c92/setup-container/0.log" Nov 28 14:22:33 crc kubenswrapper[4631]: I1128 14:22:33.571914 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_fc5706c2-8b7d-439a-95b3-981907360c92/rabbitmq/0.log" Nov 28 14:22:33 crc kubenswrapper[4631]: I1128 14:22:33.614655 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_fc5706c2-8b7d-439a-95b3-981907360c92/setup-container/0.log" Nov 28 14:22:33 crc kubenswrapper[4631]: I1128 14:22:33.675039 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-r2pq6_ec45d53a-f54c-422a-8a43-5b9d4b4ee0c6/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Nov 28 14:22:33 crc kubenswrapper[4631]: I1128 14:22:33.879471 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_redhat-edpm-deployment-openstack-edpm-ipam-s9hz7_2b15b142-3dc9-4d33-96e7-5ba3e357ee59/redhat-edpm-deployment-openstack-edpm-ipam/0.log" Nov 28 14:22:34 crc kubenswrapper[4631]: I1128 14:22:34.028339 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-slcpp_0527a324-2259-4aac-a64b-33747a2e24ec/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Nov 28 14:22:34 crc kubenswrapper[4631]: I1128 14:22:34.250088 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-zmp56_8bdc8b41-a7c5-4c60-86ce-6ac5eb65b82f/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Nov 28 14:22:34 crc kubenswrapper[4631]: I1128 14:22:34.639480 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-m2k99_b1c516b6-6c37-4bfd-a0ee-f28200e19170/ssh-known-hosts-edpm-deployment/0.log" Nov 28 14:22:34 crc kubenswrapper[4631]: I1128 14:22:34.657901 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-68757c9bcf-d8gj4_7303e352-744c-4c93-ac72-0c434384f090/proxy-httpd/0.log" Nov 28 14:22:35 crc kubenswrapper[4631]: I1128 14:22:35.071196 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-68757c9bcf-d8gj4_7303e352-744c-4c93-ac72-0c434384f090/proxy-server/0.log" Nov 28 14:22:35 crc kubenswrapper[4631]: I1128 14:22:35.124185 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-wh7th_780a47b4-0078-4f97-a5f5-ad4e1b2015b4/swift-ring-rebalance/0.log" Nov 28 14:22:35 crc kubenswrapper[4631]: I1128 14:22:35.382701 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c3e1f6fb-1845-4a39-8eca-1262da1ff281/account-auditor/0.log" Nov 28 14:22:35 crc kubenswrapper[4631]: I1128 14:22:35.418202 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c3e1f6fb-1845-4a39-8eca-1262da1ff281/account-reaper/0.log" Nov 28 14:22:35 crc kubenswrapper[4631]: I1128 14:22:35.554525 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c3e1f6fb-1845-4a39-8eca-1262da1ff281/account-replicator/0.log" Nov 28 14:22:35 crc kubenswrapper[4631]: I1128 14:22:35.559472 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c3e1f6fb-1845-4a39-8eca-1262da1ff281/account-server/0.log" Nov 28 14:22:35 crc kubenswrapper[4631]: I1128 14:22:35.777670 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c3e1f6fb-1845-4a39-8eca-1262da1ff281/container-auditor/0.log" Nov 28 14:22:35 crc kubenswrapper[4631]: I1128 14:22:35.822917 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c3e1f6fb-1845-4a39-8eca-1262da1ff281/container-replicator/0.log" Nov 28 14:22:35 crc kubenswrapper[4631]: I1128 14:22:35.886295 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c3e1f6fb-1845-4a39-8eca-1262da1ff281/container-server/0.log" Nov 28 14:22:35 crc kubenswrapper[4631]: I1128 14:22:35.999700 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c3e1f6fb-1845-4a39-8eca-1262da1ff281/container-updater/0.log" Nov 28 14:22:36 crc kubenswrapper[4631]: I1128 14:22:36.069401 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c3e1f6fb-1845-4a39-8eca-1262da1ff281/object-auditor/0.log" Nov 28 14:22:36 crc kubenswrapper[4631]: I1128 14:22:36.146098 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c3e1f6fb-1845-4a39-8eca-1262da1ff281/object-expirer/0.log" Nov 28 14:22:36 crc kubenswrapper[4631]: I1128 14:22:36.218658 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c3e1f6fb-1845-4a39-8eca-1262da1ff281/object-replicator/0.log" Nov 28 14:22:36 crc kubenswrapper[4631]: I1128 14:22:36.341356 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c3e1f6fb-1845-4a39-8eca-1262da1ff281/object-updater/0.log" Nov 28 14:22:36 crc kubenswrapper[4631]: I1128 14:22:36.379757 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c3e1f6fb-1845-4a39-8eca-1262da1ff281/object-server/0.log" Nov 28 14:22:36 crc kubenswrapper[4631]: I1128 14:22:36.471226 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c3e1f6fb-1845-4a39-8eca-1262da1ff281/rsync/0.log" Nov 28 14:22:36 crc kubenswrapper[4631]: I1128 14:22:36.539054 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c3e1f6fb-1845-4a39-8eca-1262da1ff281/swift-recon-cron/0.log" Nov 28 14:22:36 crc kubenswrapper[4631]: I1128 14:22:36.788560 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-x9bvm_dab7280d-4cb8-4a43-b056-82296e560317/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Nov 28 14:22:36 crc kubenswrapper[4631]: I1128 14:22:36.809585 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-scmhh"] Nov 28 14:22:36 crc kubenswrapper[4631]: E1128 14:22:36.810027 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b893f4f2-8903-4aaa-962a-a5a4532f781e" containerName="container-00" Nov 28 14:22:36 crc kubenswrapper[4631]: I1128 14:22:36.810042 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="b893f4f2-8903-4aaa-962a-a5a4532f781e" containerName="container-00" Nov 28 14:22:36 crc kubenswrapper[4631]: I1128 14:22:36.810220 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="b893f4f2-8903-4aaa-962a-a5a4532f781e" containerName="container-00" Nov 28 14:22:36 crc kubenswrapper[4631]: I1128 14:22:36.812208 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-scmhh" Nov 28 14:22:36 crc kubenswrapper[4631]: I1128 14:22:36.844009 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-scmhh"] Nov 28 14:22:36 crc kubenswrapper[4631]: I1128 14:22:36.882671 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9cd4efd2-7af0-4384-b565-a15063290cdf-catalog-content\") pod \"redhat-marketplace-scmhh\" (UID: \"9cd4efd2-7af0-4384-b565-a15063290cdf\") " pod="openshift-marketplace/redhat-marketplace-scmhh" Nov 28 14:22:36 crc kubenswrapper[4631]: I1128 14:22:36.882837 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bwcql\" (UniqueName: \"kubernetes.io/projected/9cd4efd2-7af0-4384-b565-a15063290cdf-kube-api-access-bwcql\") pod \"redhat-marketplace-scmhh\" (UID: \"9cd4efd2-7af0-4384-b565-a15063290cdf\") " pod="openshift-marketplace/redhat-marketplace-scmhh" Nov 28 14:22:36 crc kubenswrapper[4631]: I1128 14:22:36.882863 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9cd4efd2-7af0-4384-b565-a15063290cdf-utilities\") pod \"redhat-marketplace-scmhh\" (UID: \"9cd4efd2-7af0-4384-b565-a15063290cdf\") " pod="openshift-marketplace/redhat-marketplace-scmhh" Nov 28 14:22:36 crc kubenswrapper[4631]: I1128 14:22:36.974883 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_183d2a7b-cd4e-423c-82c0-a72e2ebec301/tempest-tests-tempest-tests-runner/0.log" Nov 28 14:22:36 crc kubenswrapper[4631]: I1128 14:22:36.985038 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bwcql\" (UniqueName: \"kubernetes.io/projected/9cd4efd2-7af0-4384-b565-a15063290cdf-kube-api-access-bwcql\") pod \"redhat-marketplace-scmhh\" (UID: \"9cd4efd2-7af0-4384-b565-a15063290cdf\") " pod="openshift-marketplace/redhat-marketplace-scmhh" Nov 28 14:22:36 crc kubenswrapper[4631]: I1128 14:22:36.985087 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9cd4efd2-7af0-4384-b565-a15063290cdf-utilities\") pod \"redhat-marketplace-scmhh\" (UID: \"9cd4efd2-7af0-4384-b565-a15063290cdf\") " pod="openshift-marketplace/redhat-marketplace-scmhh" Nov 28 14:22:36 crc kubenswrapper[4631]: I1128 14:22:36.985136 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9cd4efd2-7af0-4384-b565-a15063290cdf-catalog-content\") pod \"redhat-marketplace-scmhh\" (UID: \"9cd4efd2-7af0-4384-b565-a15063290cdf\") " pod="openshift-marketplace/redhat-marketplace-scmhh" Nov 28 14:22:36 crc kubenswrapper[4631]: I1128 14:22:36.986076 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9cd4efd2-7af0-4384-b565-a15063290cdf-utilities\") pod \"redhat-marketplace-scmhh\" (UID: \"9cd4efd2-7af0-4384-b565-a15063290cdf\") " pod="openshift-marketplace/redhat-marketplace-scmhh" Nov 28 14:22:36 crc kubenswrapper[4631]: I1128 14:22:36.986315 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9cd4efd2-7af0-4384-b565-a15063290cdf-catalog-content\") pod \"redhat-marketplace-scmhh\" (UID: \"9cd4efd2-7af0-4384-b565-a15063290cdf\") " pod="openshift-marketplace/redhat-marketplace-scmhh" Nov 28 14:22:37 crc kubenswrapper[4631]: I1128 14:22:37.017306 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bwcql\" (UniqueName: \"kubernetes.io/projected/9cd4efd2-7af0-4384-b565-a15063290cdf-kube-api-access-bwcql\") pod \"redhat-marketplace-scmhh\" (UID: \"9cd4efd2-7af0-4384-b565-a15063290cdf\") " pod="openshift-marketplace/redhat-marketplace-scmhh" Nov 28 14:22:37 crc kubenswrapper[4631]: I1128 14:22:37.148773 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-scmhh" Nov 28 14:22:37 crc kubenswrapper[4631]: I1128 14:22:37.508236 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_7fffbcdc-ae30-4841-9161-7a0f06bb7f67/test-operator-logs-container/0.log" Nov 28 14:22:37 crc kubenswrapper[4631]: I1128 14:22:37.578845 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-pwfzg_53cec30e-969e-4ccf-8ae2-76852297457a/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Nov 28 14:22:37 crc kubenswrapper[4631]: I1128 14:22:37.785837 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-scmhh"] Nov 28 14:22:38 crc kubenswrapper[4631]: I1128 14:22:38.566334 4631 generic.go:334] "Generic (PLEG): container finished" podID="9cd4efd2-7af0-4384-b565-a15063290cdf" containerID="6ddb68775c5c349683be7c8546a2cab505f745fa8178dcdde2229605483c8633" exitCode=0 Nov 28 14:22:38 crc kubenswrapper[4631]: I1128 14:22:38.566912 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-scmhh" event={"ID":"9cd4efd2-7af0-4384-b565-a15063290cdf","Type":"ContainerDied","Data":"6ddb68775c5c349683be7c8546a2cab505f745fa8178dcdde2229605483c8633"} Nov 28 14:22:38 crc kubenswrapper[4631]: I1128 14:22:38.566953 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-scmhh" event={"ID":"9cd4efd2-7af0-4384-b565-a15063290cdf","Type":"ContainerStarted","Data":"52a12d12b820a6ecb133729b898379386d36938fa3619884bfaa6969196f3034"} Nov 28 14:22:38 crc kubenswrapper[4631]: I1128 14:22:38.569800 4631 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 28 14:22:39 crc kubenswrapper[4631]: I1128 14:22:39.582904 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-scmhh" event={"ID":"9cd4efd2-7af0-4384-b565-a15063290cdf","Type":"ContainerStarted","Data":"3f892e875e96c94d93e8d3f5337e64c0c0ff4340b5bfea02dad4094143fff442"} Nov 28 14:22:40 crc kubenswrapper[4631]: I1128 14:22:40.595516 4631 generic.go:334] "Generic (PLEG): container finished" podID="9cd4efd2-7af0-4384-b565-a15063290cdf" containerID="3f892e875e96c94d93e8d3f5337e64c0c0ff4340b5bfea02dad4094143fff442" exitCode=0 Nov 28 14:22:40 crc kubenswrapper[4631]: I1128 14:22:40.595888 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-scmhh" event={"ID":"9cd4efd2-7af0-4384-b565-a15063290cdf","Type":"ContainerDied","Data":"3f892e875e96c94d93e8d3f5337e64c0c0ff4340b5bfea02dad4094143fff442"} Nov 28 14:22:42 crc kubenswrapper[4631]: I1128 14:22:42.624949 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-scmhh" event={"ID":"9cd4efd2-7af0-4384-b565-a15063290cdf","Type":"ContainerStarted","Data":"94a30ecc3f299ed899b0a3638750ccac92167ac690de1131d129a882f3b6e70b"} Nov 28 14:22:42 crc kubenswrapper[4631]: I1128 14:22:42.650840 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-scmhh" podStartSLOduration=3.727242283 podStartE2EDuration="6.6508197s" podCreationTimestamp="2025-11-28 14:22:36 +0000 UTC" firstStartedPulling="2025-11-28 14:22:38.569557194 +0000 UTC m=+3715.376860538" lastFinishedPulling="2025-11-28 14:22:41.493134611 +0000 UTC m=+3718.300437955" observedRunningTime="2025-11-28 14:22:42.649310543 +0000 UTC m=+3719.456613897" watchObservedRunningTime="2025-11-28 14:22:42.6508197 +0000 UTC m=+3719.458123044" Nov 28 14:22:46 crc kubenswrapper[4631]: I1128 14:22:46.410806 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_d35228a7-a0fe-496d-9d62-5f1d5a44c72f/memcached/0.log" Nov 28 14:22:47 crc kubenswrapper[4631]: I1128 14:22:47.150000 4631 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-scmhh" Nov 28 14:22:47 crc kubenswrapper[4631]: I1128 14:22:47.150088 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-scmhh" Nov 28 14:22:47 crc kubenswrapper[4631]: I1128 14:22:47.206754 4631 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-scmhh" Nov 28 14:22:47 crc kubenswrapper[4631]: I1128 14:22:47.731884 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-scmhh" Nov 28 14:22:47 crc kubenswrapper[4631]: I1128 14:22:47.781813 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-scmhh"] Nov 28 14:22:49 crc kubenswrapper[4631]: I1128 14:22:49.634786 4631 patch_prober.go:28] interesting pod/machine-config-daemon-47sxw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 28 14:22:49 crc kubenswrapper[4631]: I1128 14:22:49.634876 4631 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" podUID="8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 28 14:22:49 crc kubenswrapper[4631]: I1128 14:22:49.707026 4631 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-scmhh" podUID="9cd4efd2-7af0-4384-b565-a15063290cdf" containerName="registry-server" containerID="cri-o://94a30ecc3f299ed899b0a3638750ccac92167ac690de1131d129a882f3b6e70b" gracePeriod=2 Nov 28 14:22:50 crc kubenswrapper[4631]: I1128 14:22:50.246283 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-scmhh" Nov 28 14:22:50 crc kubenswrapper[4631]: I1128 14:22:50.401202 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bwcql\" (UniqueName: \"kubernetes.io/projected/9cd4efd2-7af0-4384-b565-a15063290cdf-kube-api-access-bwcql\") pod \"9cd4efd2-7af0-4384-b565-a15063290cdf\" (UID: \"9cd4efd2-7af0-4384-b565-a15063290cdf\") " Nov 28 14:22:50 crc kubenswrapper[4631]: I1128 14:22:50.401363 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9cd4efd2-7af0-4384-b565-a15063290cdf-catalog-content\") pod \"9cd4efd2-7af0-4384-b565-a15063290cdf\" (UID: \"9cd4efd2-7af0-4384-b565-a15063290cdf\") " Nov 28 14:22:50 crc kubenswrapper[4631]: I1128 14:22:50.401446 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9cd4efd2-7af0-4384-b565-a15063290cdf-utilities\") pod \"9cd4efd2-7af0-4384-b565-a15063290cdf\" (UID: \"9cd4efd2-7af0-4384-b565-a15063290cdf\") " Nov 28 14:22:50 crc kubenswrapper[4631]: I1128 14:22:50.402327 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9cd4efd2-7af0-4384-b565-a15063290cdf-utilities" (OuterVolumeSpecName: "utilities") pod "9cd4efd2-7af0-4384-b565-a15063290cdf" (UID: "9cd4efd2-7af0-4384-b565-a15063290cdf"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 14:22:50 crc kubenswrapper[4631]: I1128 14:22:50.402901 4631 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9cd4efd2-7af0-4384-b565-a15063290cdf-utilities\") on node \"crc\" DevicePath \"\"" Nov 28 14:22:50 crc kubenswrapper[4631]: I1128 14:22:50.414057 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9cd4efd2-7af0-4384-b565-a15063290cdf-kube-api-access-bwcql" (OuterVolumeSpecName: "kube-api-access-bwcql") pod "9cd4efd2-7af0-4384-b565-a15063290cdf" (UID: "9cd4efd2-7af0-4384-b565-a15063290cdf"). InnerVolumeSpecName "kube-api-access-bwcql". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:22:50 crc kubenswrapper[4631]: I1128 14:22:50.429611 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9cd4efd2-7af0-4384-b565-a15063290cdf-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9cd4efd2-7af0-4384-b565-a15063290cdf" (UID: "9cd4efd2-7af0-4384-b565-a15063290cdf"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 14:22:50 crc kubenswrapper[4631]: I1128 14:22:50.504665 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bwcql\" (UniqueName: \"kubernetes.io/projected/9cd4efd2-7af0-4384-b565-a15063290cdf-kube-api-access-bwcql\") on node \"crc\" DevicePath \"\"" Nov 28 14:22:50 crc kubenswrapper[4631]: I1128 14:22:50.504698 4631 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9cd4efd2-7af0-4384-b565-a15063290cdf-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 28 14:22:50 crc kubenswrapper[4631]: I1128 14:22:50.720741 4631 generic.go:334] "Generic (PLEG): container finished" podID="9cd4efd2-7af0-4384-b565-a15063290cdf" containerID="94a30ecc3f299ed899b0a3638750ccac92167ac690de1131d129a882f3b6e70b" exitCode=0 Nov 28 14:22:50 crc kubenswrapper[4631]: I1128 14:22:50.720841 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-scmhh" Nov 28 14:22:50 crc kubenswrapper[4631]: I1128 14:22:50.720931 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-scmhh" event={"ID":"9cd4efd2-7af0-4384-b565-a15063290cdf","Type":"ContainerDied","Data":"94a30ecc3f299ed899b0a3638750ccac92167ac690de1131d129a882f3b6e70b"} Nov 28 14:22:50 crc kubenswrapper[4631]: I1128 14:22:50.721773 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-scmhh" event={"ID":"9cd4efd2-7af0-4384-b565-a15063290cdf","Type":"ContainerDied","Data":"52a12d12b820a6ecb133729b898379386d36938fa3619884bfaa6969196f3034"} Nov 28 14:22:50 crc kubenswrapper[4631]: I1128 14:22:50.721807 4631 scope.go:117] "RemoveContainer" containerID="94a30ecc3f299ed899b0a3638750ccac92167ac690de1131d129a882f3b6e70b" Nov 28 14:22:50 crc kubenswrapper[4631]: I1128 14:22:50.769475 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-scmhh"] Nov 28 14:22:50 crc kubenswrapper[4631]: I1128 14:22:50.772084 4631 scope.go:117] "RemoveContainer" containerID="3f892e875e96c94d93e8d3f5337e64c0c0ff4340b5bfea02dad4094143fff442" Nov 28 14:22:50 crc kubenswrapper[4631]: I1128 14:22:50.782817 4631 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-scmhh"] Nov 28 14:22:50 crc kubenswrapper[4631]: I1128 14:22:50.803986 4631 scope.go:117] "RemoveContainer" containerID="6ddb68775c5c349683be7c8546a2cab505f745fa8178dcdde2229605483c8633" Nov 28 14:22:50 crc kubenswrapper[4631]: I1128 14:22:50.850635 4631 scope.go:117] "RemoveContainer" containerID="94a30ecc3f299ed899b0a3638750ccac92167ac690de1131d129a882f3b6e70b" Nov 28 14:22:50 crc kubenswrapper[4631]: E1128 14:22:50.851164 4631 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"94a30ecc3f299ed899b0a3638750ccac92167ac690de1131d129a882f3b6e70b\": container with ID starting with 94a30ecc3f299ed899b0a3638750ccac92167ac690de1131d129a882f3b6e70b not found: ID does not exist" containerID="94a30ecc3f299ed899b0a3638750ccac92167ac690de1131d129a882f3b6e70b" Nov 28 14:22:50 crc kubenswrapper[4631]: I1128 14:22:50.851220 4631 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"94a30ecc3f299ed899b0a3638750ccac92167ac690de1131d129a882f3b6e70b"} err="failed to get container status \"94a30ecc3f299ed899b0a3638750ccac92167ac690de1131d129a882f3b6e70b\": rpc error: code = NotFound desc = could not find container \"94a30ecc3f299ed899b0a3638750ccac92167ac690de1131d129a882f3b6e70b\": container with ID starting with 94a30ecc3f299ed899b0a3638750ccac92167ac690de1131d129a882f3b6e70b not found: ID does not exist" Nov 28 14:22:50 crc kubenswrapper[4631]: I1128 14:22:50.851253 4631 scope.go:117] "RemoveContainer" containerID="3f892e875e96c94d93e8d3f5337e64c0c0ff4340b5bfea02dad4094143fff442" Nov 28 14:22:50 crc kubenswrapper[4631]: E1128 14:22:50.853189 4631 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3f892e875e96c94d93e8d3f5337e64c0c0ff4340b5bfea02dad4094143fff442\": container with ID starting with 3f892e875e96c94d93e8d3f5337e64c0c0ff4340b5bfea02dad4094143fff442 not found: ID does not exist" containerID="3f892e875e96c94d93e8d3f5337e64c0c0ff4340b5bfea02dad4094143fff442" Nov 28 14:22:50 crc kubenswrapper[4631]: I1128 14:22:50.853225 4631 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3f892e875e96c94d93e8d3f5337e64c0c0ff4340b5bfea02dad4094143fff442"} err="failed to get container status \"3f892e875e96c94d93e8d3f5337e64c0c0ff4340b5bfea02dad4094143fff442\": rpc error: code = NotFound desc = could not find container \"3f892e875e96c94d93e8d3f5337e64c0c0ff4340b5bfea02dad4094143fff442\": container with ID starting with 3f892e875e96c94d93e8d3f5337e64c0c0ff4340b5bfea02dad4094143fff442 not found: ID does not exist" Nov 28 14:22:50 crc kubenswrapper[4631]: I1128 14:22:50.853251 4631 scope.go:117] "RemoveContainer" containerID="6ddb68775c5c349683be7c8546a2cab505f745fa8178dcdde2229605483c8633" Nov 28 14:22:50 crc kubenswrapper[4631]: E1128 14:22:50.856414 4631 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6ddb68775c5c349683be7c8546a2cab505f745fa8178dcdde2229605483c8633\": container with ID starting with 6ddb68775c5c349683be7c8546a2cab505f745fa8178dcdde2229605483c8633 not found: ID does not exist" containerID="6ddb68775c5c349683be7c8546a2cab505f745fa8178dcdde2229605483c8633" Nov 28 14:22:50 crc kubenswrapper[4631]: I1128 14:22:50.856459 4631 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6ddb68775c5c349683be7c8546a2cab505f745fa8178dcdde2229605483c8633"} err="failed to get container status \"6ddb68775c5c349683be7c8546a2cab505f745fa8178dcdde2229605483c8633\": rpc error: code = NotFound desc = could not find container \"6ddb68775c5c349683be7c8546a2cab505f745fa8178dcdde2229605483c8633\": container with ID starting with 6ddb68775c5c349683be7c8546a2cab505f745fa8178dcdde2229605483c8633 not found: ID does not exist" Nov 28 14:22:51 crc kubenswrapper[4631]: I1128 14:22:51.522890 4631 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9cd4efd2-7af0-4384-b565-a15063290cdf" path="/var/lib/kubelet/pods/9cd4efd2-7af0-4384-b565-a15063290cdf/volumes" Nov 28 14:23:05 crc kubenswrapper[4631]: I1128 14:23:05.383534 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7b64f4fb85-hdr72_00a81f80-35b9-4eb4-a508-3f104ca426be/kube-rbac-proxy/0.log" Nov 28 14:23:05 crc kubenswrapper[4631]: I1128 14:23:05.472598 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7b64f4fb85-hdr72_00a81f80-35b9-4eb4-a508-3f104ca426be/manager/0.log" Nov 28 14:23:05 crc kubenswrapper[4631]: I1128 14:23:05.597801 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-6b7f75547b-rmlgq_8028d28d-2414-4ffe-9c5d-4854f77e980f/kube-rbac-proxy/0.log" Nov 28 14:23:05 crc kubenswrapper[4631]: I1128 14:23:05.704116 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-6b7f75547b-rmlgq_8028d28d-2414-4ffe-9c5d-4854f77e980f/manager/0.log" Nov 28 14:23:05 crc kubenswrapper[4631]: I1128 14:23:05.834835 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_dee8d58050df5e5c4d52f5078d12ac93857da536112fc471b2066b915e9wbcx_57ea9be9-7467-4c95-b00b-14f7df44eae4/util/0.log" Nov 28 14:23:06 crc kubenswrapper[4631]: I1128 14:23:06.001317 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_dee8d58050df5e5c4d52f5078d12ac93857da536112fc471b2066b915e9wbcx_57ea9be9-7467-4c95-b00b-14f7df44eae4/util/0.log" Nov 28 14:23:06 crc kubenswrapper[4631]: I1128 14:23:06.035224 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_dee8d58050df5e5c4d52f5078d12ac93857da536112fc471b2066b915e9wbcx_57ea9be9-7467-4c95-b00b-14f7df44eae4/pull/0.log" Nov 28 14:23:06 crc kubenswrapper[4631]: I1128 14:23:06.073195 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_dee8d58050df5e5c4d52f5078d12ac93857da536112fc471b2066b915e9wbcx_57ea9be9-7467-4c95-b00b-14f7df44eae4/pull/0.log" Nov 28 14:23:06 crc kubenswrapper[4631]: I1128 14:23:06.246305 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_dee8d58050df5e5c4d52f5078d12ac93857da536112fc471b2066b915e9wbcx_57ea9be9-7467-4c95-b00b-14f7df44eae4/pull/0.log" Nov 28 14:23:06 crc kubenswrapper[4631]: I1128 14:23:06.249307 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_dee8d58050df5e5c4d52f5078d12ac93857da536112fc471b2066b915e9wbcx_57ea9be9-7467-4c95-b00b-14f7df44eae4/extract/0.log" Nov 28 14:23:06 crc kubenswrapper[4631]: I1128 14:23:06.282722 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_dee8d58050df5e5c4d52f5078d12ac93857da536112fc471b2066b915e9wbcx_57ea9be9-7467-4c95-b00b-14f7df44eae4/util/0.log" Nov 28 14:23:06 crc kubenswrapper[4631]: I1128 14:23:06.472267 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-589cbd6b5b-npdtc_59ee25ad-f169-47b9-9d5c-f7a7a253a591/kube-rbac-proxy/0.log" Nov 28 14:23:06 crc kubenswrapper[4631]: I1128 14:23:06.495961 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-955677c94-6bt2f_06994d92-3dcb-461c-83ac-c7e65d4bd640/manager/0.log" Nov 28 14:23:06 crc kubenswrapper[4631]: I1128 14:23:06.518318 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-955677c94-6bt2f_06994d92-3dcb-461c-83ac-c7e65d4bd640/kube-rbac-proxy/0.log" Nov 28 14:23:06 crc kubenswrapper[4631]: I1128 14:23:06.764311 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-589cbd6b5b-npdtc_59ee25ad-f169-47b9-9d5c-f7a7a253a591/manager/0.log" Nov 28 14:23:06 crc kubenswrapper[4631]: I1128 14:23:06.811720 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5b77f656f-8pvq9_9cc92c5c-285e-4604-bc60-7637938a78c8/kube-rbac-proxy/0.log" Nov 28 14:23:06 crc kubenswrapper[4631]: I1128 14:23:06.813632 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5b77f656f-8pvq9_9cc92c5c-285e-4604-bc60-7637938a78c8/manager/0.log" Nov 28 14:23:06 crc kubenswrapper[4631]: I1128 14:23:06.955695 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-5d494799bf-2mq7v_aafff68d-b989-4edc-ab1c-b9d33edb8ee2/kube-rbac-proxy/0.log" Nov 28 14:23:07 crc kubenswrapper[4631]: I1128 14:23:07.040908 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-5d494799bf-2mq7v_aafff68d-b989-4edc-ab1c-b9d33edb8ee2/manager/0.log" Nov 28 14:23:07 crc kubenswrapper[4631]: I1128 14:23:07.231462 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-57548d458d-h7npd_0079ab2d-c5a8-45eb-b6e8-9574158de32f/kube-rbac-proxy/0.log" Nov 28 14:23:07 crc kubenswrapper[4631]: I1128 14:23:07.289612 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-57548d458d-h7npd_0079ab2d-c5a8-45eb-b6e8-9574158de32f/manager/0.log" Nov 28 14:23:07 crc kubenswrapper[4631]: I1128 14:23:07.380174 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-67cb4dc6d4-db8kg_ed17d71f-b72c-4f12-b5f7-d5cb3bb130dc/kube-rbac-proxy/0.log" Nov 28 14:23:07 crc kubenswrapper[4631]: I1128 14:23:07.503108 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-67cb4dc6d4-db8kg_ed17d71f-b72c-4f12-b5f7-d5cb3bb130dc/manager/0.log" Nov 28 14:23:07 crc kubenswrapper[4631]: I1128 14:23:07.557244 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-6b49bfffcd-5ks5f_f3e12436-8a74-4443-94a4-3b9717105b05/kube-rbac-proxy/0.log" Nov 28 14:23:07 crc kubenswrapper[4631]: I1128 14:23:07.653667 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-6b49bfffcd-5ks5f_f3e12436-8a74-4443-94a4-3b9717105b05/manager/0.log" Nov 28 14:23:07 crc kubenswrapper[4631]: I1128 14:23:07.800166 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-5d499bf58b-dck2d_dbbf77e7-3c97-480f-83b1-0eb8bb8f386d/manager/0.log" Nov 28 14:23:07 crc kubenswrapper[4631]: I1128 14:23:07.802935 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-5d499bf58b-dck2d_dbbf77e7-3c97-480f-83b1-0eb8bb8f386d/kube-rbac-proxy/0.log" Nov 28 14:23:07 crc kubenswrapper[4631]: I1128 14:23:07.945606 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-66f4dd4bc7-m9t96_cefc6050-9552-4dda-9864-39ac05626785/kube-rbac-proxy/0.log" Nov 28 14:23:08 crc kubenswrapper[4631]: I1128 14:23:08.036348 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-66f4dd4bc7-m9t96_cefc6050-9552-4dda-9864-39ac05626785/manager/0.log" Nov 28 14:23:08 crc kubenswrapper[4631]: I1128 14:23:08.124622 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-6fdcddb789-msc6n_2dab52af-5e82-4119-b4d5-56fb31abcb4d/kube-rbac-proxy/0.log" Nov 28 14:23:08 crc kubenswrapper[4631]: I1128 14:23:08.220373 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-6fdcddb789-msc6n_2dab52af-5e82-4119-b4d5-56fb31abcb4d/manager/0.log" Nov 28 14:23:08 crc kubenswrapper[4631]: I1128 14:23:08.343565 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-79556f57fc-tpbhk_bb908ace-bb62-4f55-8131-90e48da7585b/kube-rbac-proxy/0.log" Nov 28 14:23:08 crc kubenswrapper[4631]: I1128 14:23:08.408136 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-79556f57fc-tpbhk_bb908ace-bb62-4f55-8131-90e48da7585b/manager/0.log" Nov 28 14:23:08 crc kubenswrapper[4631]: I1128 14:23:08.540137 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-64cdc6ff96-4qzz7_ecfd67ad-5136-4e6e-9e41-3d933d10ea2c/kube-rbac-proxy/0.log" Nov 28 14:23:08 crc kubenswrapper[4631]: I1128 14:23:08.646436 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-64cdc6ff96-4qzz7_ecfd67ad-5136-4e6e-9e41-3d933d10ea2c/manager/0.log" Nov 28 14:23:08 crc kubenswrapper[4631]: I1128 14:23:08.777091 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-5fcdb54b6b2f9vf_ca18d816-8a63-4dcd-ac27-b1e7ad9ce4f6/kube-rbac-proxy/0.log" Nov 28 14:23:08 crc kubenswrapper[4631]: I1128 14:23:08.807240 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-5fcdb54b6b2f9vf_ca18d816-8a63-4dcd-ac27-b1e7ad9ce4f6/manager/0.log" Nov 28 14:23:09 crc kubenswrapper[4631]: I1128 14:23:09.318517 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-5bcf64b564-zsrq9_1104dd12-c822-4ffa-9634-d82898f8f575/operator/0.log" Nov 28 14:23:09 crc kubenswrapper[4631]: I1128 14:23:09.392883 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-qclsp_8c8c3b22-0767-4d9f-9ec4-bf42ddcd8433/registry-server/0.log" Nov 28 14:23:09 crc kubenswrapper[4631]: I1128 14:23:09.685152 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-56897c768d-wkpql_d09fac54-dd24-47df-83a1-23cae3102e7a/kube-rbac-proxy/0.log" Nov 28 14:23:09 crc kubenswrapper[4631]: I1128 14:23:09.748188 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-56897c768d-wkpql_d09fac54-dd24-47df-83a1-23cae3102e7a/manager/0.log" Nov 28 14:23:09 crc kubenswrapper[4631]: I1128 14:23:09.899660 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-57988cc5b5-wqpdp_05a6dcba-27b4-4b20-9275-f434b8943be9/kube-rbac-proxy/0.log" Nov 28 14:23:09 crc kubenswrapper[4631]: I1128 14:23:09.997365 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-57988cc5b5-wqpdp_05a6dcba-27b4-4b20-9275-f434b8943be9/manager/0.log" Nov 28 14:23:10 crc kubenswrapper[4631]: I1128 14:23:10.146159 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-854b9f67c6-hfb26_4bc38647-eefa-4ad6-bef5-6a7ed535b794/manager/0.log" Nov 28 14:23:10 crc kubenswrapper[4631]: I1128 14:23:10.185936 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-nrpfx_10a95153-0898-4016-92a6-d56633555348/operator/0.log" Nov 28 14:23:10 crc kubenswrapper[4631]: I1128 14:23:10.307269 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-d77b94747-2g446_2d59108a-2b92-421b-beb8-157c2eca09ac/manager/0.log" Nov 28 14:23:10 crc kubenswrapper[4631]: I1128 14:23:10.309475 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-d77b94747-2g446_2d59108a-2b92-421b-beb8-157c2eca09ac/kube-rbac-proxy/0.log" Nov 28 14:23:10 crc kubenswrapper[4631]: I1128 14:23:10.351483 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-76cc84c6bb-r94wn_f2aea9a5-85cf-464f-8eba-382b3db8d71d/kube-rbac-proxy/0.log" Nov 28 14:23:10 crc kubenswrapper[4631]: I1128 14:23:10.468776 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-76cc84c6bb-r94wn_f2aea9a5-85cf-464f-8eba-382b3db8d71d/manager/0.log" Nov 28 14:23:10 crc kubenswrapper[4631]: I1128 14:23:10.565083 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5cd6c7f4c8-jwd25_fa36877c-8fc2-4074-a75d-135c2006807a/manager/0.log" Nov 28 14:23:10 crc kubenswrapper[4631]: I1128 14:23:10.604513 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5cd6c7f4c8-jwd25_fa36877c-8fc2-4074-a75d-135c2006807a/kube-rbac-proxy/0.log" Nov 28 14:23:10 crc kubenswrapper[4631]: I1128 14:23:10.686199 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-656dcb59d4-wh9v9_826f27ac-a938-4292-a357-ff54c4117986/kube-rbac-proxy/0.log" Nov 28 14:23:10 crc kubenswrapper[4631]: I1128 14:23:10.780669 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-656dcb59d4-wh9v9_826f27ac-a938-4292-a357-ff54c4117986/manager/0.log" Nov 28 14:23:19 crc kubenswrapper[4631]: I1128 14:23:19.635274 4631 patch_prober.go:28] interesting pod/machine-config-daemon-47sxw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 28 14:23:19 crc kubenswrapper[4631]: I1128 14:23:19.635849 4631 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" podUID="8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 28 14:23:19 crc kubenswrapper[4631]: I1128 14:23:19.635904 4631 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" Nov 28 14:23:19 crc kubenswrapper[4631]: I1128 14:23:19.636796 4631 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"b2f5e32c4eeb164aad43a3d1940c5d648cc8679d4f599da50345a14cc66f0979"} pod="openshift-machine-config-operator/machine-config-daemon-47sxw" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 28 14:23:19 crc kubenswrapper[4631]: I1128 14:23:19.636854 4631 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" podUID="8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1" containerName="machine-config-daemon" containerID="cri-o://b2f5e32c4eeb164aad43a3d1940c5d648cc8679d4f599da50345a14cc66f0979" gracePeriod=600 Nov 28 14:23:19 crc kubenswrapper[4631]: E1128 14:23:19.775041 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-47sxw_openshift-machine-config-operator(8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1)\"" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" podUID="8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1" Nov 28 14:23:19 crc kubenswrapper[4631]: I1128 14:23:19.997716 4631 generic.go:334] "Generic (PLEG): container finished" podID="8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1" containerID="b2f5e32c4eeb164aad43a3d1940c5d648cc8679d4f599da50345a14cc66f0979" exitCode=0 Nov 28 14:23:19 crc kubenswrapper[4631]: I1128 14:23:19.997757 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" event={"ID":"8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1","Type":"ContainerDied","Data":"b2f5e32c4eeb164aad43a3d1940c5d648cc8679d4f599da50345a14cc66f0979"} Nov 28 14:23:19 crc kubenswrapper[4631]: I1128 14:23:19.997789 4631 scope.go:117] "RemoveContainer" containerID="a4c0198e8942036cb2b3c39a8f93d0e5314f141d5b92505f69b7cefcad77bdfc" Nov 28 14:23:19 crc kubenswrapper[4631]: I1128 14:23:19.998414 4631 scope.go:117] "RemoveContainer" containerID="b2f5e32c4eeb164aad43a3d1940c5d648cc8679d4f599da50345a14cc66f0979" Nov 28 14:23:19 crc kubenswrapper[4631]: E1128 14:23:19.998701 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-47sxw_openshift-machine-config-operator(8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1)\"" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" podUID="8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1" Nov 28 14:23:31 crc kubenswrapper[4631]: I1128 14:23:31.876147 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-2v5d6_2a2f7727-f05e-4deb-8eba-cb3ec3561621/control-plane-machine-set-operator/0.log" Nov 28 14:23:32 crc kubenswrapper[4631]: I1128 14:23:32.063150 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-2ch2b_889c117d-ea46-4f02-a2e0-42a47c6e4683/machine-api-operator/0.log" Nov 28 14:23:32 crc kubenswrapper[4631]: I1128 14:23:32.070471 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-2ch2b_889c117d-ea46-4f02-a2e0-42a47c6e4683/kube-rbac-proxy/0.log" Nov 28 14:23:34 crc kubenswrapper[4631]: I1128 14:23:34.512923 4631 scope.go:117] "RemoveContainer" containerID="b2f5e32c4eeb164aad43a3d1940c5d648cc8679d4f599da50345a14cc66f0979" Nov 28 14:23:34 crc kubenswrapper[4631]: E1128 14:23:34.513632 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-47sxw_openshift-machine-config-operator(8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1)\"" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" podUID="8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1" Nov 28 14:23:45 crc kubenswrapper[4631]: I1128 14:23:45.516709 4631 scope.go:117] "RemoveContainer" containerID="b2f5e32c4eeb164aad43a3d1940c5d648cc8679d4f599da50345a14cc66f0979" Nov 28 14:23:45 crc kubenswrapper[4631]: E1128 14:23:45.517445 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-47sxw_openshift-machine-config-operator(8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1)\"" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" podUID="8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1" Nov 28 14:23:47 crc kubenswrapper[4631]: I1128 14:23:47.834129 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-kf7cv_4e46e751-ddac-4bc4-b463-d7cf9b772871/cert-manager-controller/0.log" Nov 28 14:23:48 crc kubenswrapper[4631]: I1128 14:23:48.036699 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-rslr5_6591b4ca-08f0-4346-aed6-8e9d1caf2b7f/cert-manager-cainjector/0.log" Nov 28 14:23:48 crc kubenswrapper[4631]: I1128 14:23:48.091012 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-l6tzn_11540749-0bb3-4c94-9d22-12a889ac10ce/cert-manager-webhook/0.log" Nov 28 14:23:57 crc kubenswrapper[4631]: I1128 14:23:57.512869 4631 scope.go:117] "RemoveContainer" containerID="b2f5e32c4eeb164aad43a3d1940c5d648cc8679d4f599da50345a14cc66f0979" Nov 28 14:23:57 crc kubenswrapper[4631]: E1128 14:23:57.514656 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-47sxw_openshift-machine-config-operator(8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1)\"" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" podUID="8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1" Nov 28 14:24:01 crc kubenswrapper[4631]: I1128 14:24:01.740708 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-7fbb5f6569-k9dxt_61c15dca-2c0d-4c3f-b63c-6a6b7be6ad99/nmstate-console-plugin/0.log" Nov 28 14:24:01 crc kubenswrapper[4631]: I1128 14:24:01.894370 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-9t5sn_6dd14cda-0401-4840-8e08-aebcb536ffb9/nmstate-handler/0.log" Nov 28 14:24:01 crc kubenswrapper[4631]: I1128 14:24:01.989233 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-z7fvn_40f6a5ae-0fc9-4af6-a001-667ce456f18a/kube-rbac-proxy/0.log" Nov 28 14:24:02 crc kubenswrapper[4631]: I1128 14:24:02.007940 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-z7fvn_40f6a5ae-0fc9-4af6-a001-667ce456f18a/nmstate-metrics/0.log" Nov 28 14:24:02 crc kubenswrapper[4631]: I1128 14:24:02.097075 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-5b5b58f5c8-8swhv_9124a25d-f057-4244-bd4d-a612b7a4a01b/nmstate-operator/0.log" Nov 28 14:24:02 crc kubenswrapper[4631]: I1128 14:24:02.273907 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-5f6d4c5ccb-b2ssz_f4c0815f-e4ee-4be8-855a-9bf949bf91a0/nmstate-webhook/0.log" Nov 28 14:24:11 crc kubenswrapper[4631]: I1128 14:24:11.513137 4631 scope.go:117] "RemoveContainer" containerID="b2f5e32c4eeb164aad43a3d1940c5d648cc8679d4f599da50345a14cc66f0979" Nov 28 14:24:11 crc kubenswrapper[4631]: E1128 14:24:11.513867 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-47sxw_openshift-machine-config-operator(8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1)\"" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" podUID="8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1" Nov 28 14:24:19 crc kubenswrapper[4631]: I1128 14:24:19.133876 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-fpkw8_f9e35a8e-a1f5-40f9-b673-533ee2f1425f/controller/0.log" Nov 28 14:24:19 crc kubenswrapper[4631]: I1128 14:24:19.170150 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-fpkw8_f9e35a8e-a1f5-40f9-b673-533ee2f1425f/kube-rbac-proxy/0.log" Nov 28 14:24:19 crc kubenswrapper[4631]: I1128 14:24:19.399813 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-5cpds_15c5a33d-b3c0-45aa-963f-b6d9aa639edd/cp-frr-files/0.log" Nov 28 14:24:19 crc kubenswrapper[4631]: I1128 14:24:19.632714 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-5cpds_15c5a33d-b3c0-45aa-963f-b6d9aa639edd/cp-frr-files/0.log" Nov 28 14:24:19 crc kubenswrapper[4631]: I1128 14:24:19.646384 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-5cpds_15c5a33d-b3c0-45aa-963f-b6d9aa639edd/cp-metrics/0.log" Nov 28 14:24:19 crc kubenswrapper[4631]: I1128 14:24:19.661252 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-5cpds_15c5a33d-b3c0-45aa-963f-b6d9aa639edd/cp-reloader/0.log" Nov 28 14:24:19 crc kubenswrapper[4631]: I1128 14:24:19.713864 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-5cpds_15c5a33d-b3c0-45aa-963f-b6d9aa639edd/cp-reloader/0.log" Nov 28 14:24:19 crc kubenswrapper[4631]: I1128 14:24:19.927839 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-5cpds_15c5a33d-b3c0-45aa-963f-b6d9aa639edd/cp-frr-files/0.log" Nov 28 14:24:19 crc kubenswrapper[4631]: I1128 14:24:19.979693 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-5cpds_15c5a33d-b3c0-45aa-963f-b6d9aa639edd/cp-metrics/0.log" Nov 28 14:24:19 crc kubenswrapper[4631]: I1128 14:24:19.986142 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-5cpds_15c5a33d-b3c0-45aa-963f-b6d9aa639edd/cp-metrics/0.log" Nov 28 14:24:20 crc kubenswrapper[4631]: I1128 14:24:20.026090 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-5cpds_15c5a33d-b3c0-45aa-963f-b6d9aa639edd/cp-reloader/0.log" Nov 28 14:24:20 crc kubenswrapper[4631]: I1128 14:24:20.129239 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-5cpds_15c5a33d-b3c0-45aa-963f-b6d9aa639edd/cp-frr-files/0.log" Nov 28 14:24:20 crc kubenswrapper[4631]: I1128 14:24:20.196017 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-5cpds_15c5a33d-b3c0-45aa-963f-b6d9aa639edd/cp-metrics/0.log" Nov 28 14:24:20 crc kubenswrapper[4631]: I1128 14:24:20.205652 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-5cpds_15c5a33d-b3c0-45aa-963f-b6d9aa639edd/cp-reloader/0.log" Nov 28 14:24:20 crc kubenswrapper[4631]: I1128 14:24:20.233548 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-5cpds_15c5a33d-b3c0-45aa-963f-b6d9aa639edd/controller/0.log" Nov 28 14:24:20 crc kubenswrapper[4631]: I1128 14:24:20.458992 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-5cpds_15c5a33d-b3c0-45aa-963f-b6d9aa639edd/frr-metrics/0.log" Nov 28 14:24:20 crc kubenswrapper[4631]: I1128 14:24:20.473513 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-5cpds_15c5a33d-b3c0-45aa-963f-b6d9aa639edd/kube-rbac-proxy/0.log" Nov 28 14:24:21 crc kubenswrapper[4631]: I1128 14:24:21.159700 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-5cpds_15c5a33d-b3c0-45aa-963f-b6d9aa639edd/kube-rbac-proxy-frr/0.log" Nov 28 14:24:21 crc kubenswrapper[4631]: I1128 14:24:21.388149 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-5cpds_15c5a33d-b3c0-45aa-963f-b6d9aa639edd/reloader/0.log" Nov 28 14:24:21 crc kubenswrapper[4631]: I1128 14:24:21.462040 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-7fcb986d4-9vxk7_020cecb2-380e-4c2a-8741-eee44e230abf/frr-k8s-webhook-server/0.log" Nov 28 14:24:21 crc kubenswrapper[4631]: I1128 14:24:21.750071 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-7c98947894-pw5mq_299f5184-243e-4fe0-b60e-2d40eee5a3f6/manager/0.log" Nov 28 14:24:22 crc kubenswrapper[4631]: I1128 14:24:22.020991 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-d24rl_255213e3-20a3-4b0b-a165-c598ecce4766/kube-rbac-proxy/0.log" Nov 28 14:24:22 crc kubenswrapper[4631]: I1128 14:24:22.089318 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-fc587766c-d92lf_dc71ed9c-9a56-449a-afab-1a28a746fb2e/webhook-server/0.log" Nov 28 14:24:22 crc kubenswrapper[4631]: I1128 14:24:22.290212 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-5cpds_15c5a33d-b3c0-45aa-963f-b6d9aa639edd/frr/0.log" Nov 28 14:24:22 crc kubenswrapper[4631]: I1128 14:24:22.513735 4631 scope.go:117] "RemoveContainer" containerID="b2f5e32c4eeb164aad43a3d1940c5d648cc8679d4f599da50345a14cc66f0979" Nov 28 14:24:22 crc kubenswrapper[4631]: E1128 14:24:22.513968 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-47sxw_openshift-machine-config-operator(8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1)\"" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" podUID="8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1" Nov 28 14:24:22 crc kubenswrapper[4631]: I1128 14:24:22.629974 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-d24rl_255213e3-20a3-4b0b-a165-c598ecce4766/speaker/0.log" Nov 28 14:24:33 crc kubenswrapper[4631]: I1128 14:24:33.519266 4631 scope.go:117] "RemoveContainer" containerID="b2f5e32c4eeb164aad43a3d1940c5d648cc8679d4f599da50345a14cc66f0979" Nov 28 14:24:33 crc kubenswrapper[4631]: E1128 14:24:33.519885 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-47sxw_openshift-machine-config-operator(8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1)\"" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" podUID="8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1" Nov 28 14:24:36 crc kubenswrapper[4631]: I1128 14:24:36.052651 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f6xhhl_86f665f5-dfec-46bd-91aa-b07aa5803661/util/0.log" Nov 28 14:24:36 crc kubenswrapper[4631]: I1128 14:24:36.265664 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f6xhhl_86f665f5-dfec-46bd-91aa-b07aa5803661/pull/0.log" Nov 28 14:24:36 crc kubenswrapper[4631]: I1128 14:24:36.266868 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f6xhhl_86f665f5-dfec-46bd-91aa-b07aa5803661/util/0.log" Nov 28 14:24:36 crc kubenswrapper[4631]: I1128 14:24:36.326862 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f6xhhl_86f665f5-dfec-46bd-91aa-b07aa5803661/pull/0.log" Nov 28 14:24:36 crc kubenswrapper[4631]: I1128 14:24:36.575700 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f6xhhl_86f665f5-dfec-46bd-91aa-b07aa5803661/extract/0.log" Nov 28 14:24:36 crc kubenswrapper[4631]: I1128 14:24:36.682048 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f6xhhl_86f665f5-dfec-46bd-91aa-b07aa5803661/pull/0.log" Nov 28 14:24:36 crc kubenswrapper[4631]: I1128 14:24:36.683714 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f6xhhl_86f665f5-dfec-46bd-91aa-b07aa5803661/util/0.log" Nov 28 14:24:36 crc kubenswrapper[4631]: I1128 14:24:36.803306 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83rfhlr_ddb177f8-0125-45bc-bbc4-130d689e10c9/util/0.log" Nov 28 14:24:37 crc kubenswrapper[4631]: I1128 14:24:37.068956 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83rfhlr_ddb177f8-0125-45bc-bbc4-130d689e10c9/pull/0.log" Nov 28 14:24:37 crc kubenswrapper[4631]: I1128 14:24:37.069224 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83rfhlr_ddb177f8-0125-45bc-bbc4-130d689e10c9/pull/0.log" Nov 28 14:24:37 crc kubenswrapper[4631]: I1128 14:24:37.074167 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83rfhlr_ddb177f8-0125-45bc-bbc4-130d689e10c9/util/0.log" Nov 28 14:24:37 crc kubenswrapper[4631]: I1128 14:24:37.330319 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83rfhlr_ddb177f8-0125-45bc-bbc4-130d689e10c9/pull/0.log" Nov 28 14:24:37 crc kubenswrapper[4631]: I1128 14:24:37.364042 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83rfhlr_ddb177f8-0125-45bc-bbc4-130d689e10c9/util/0.log" Nov 28 14:24:37 crc kubenswrapper[4631]: I1128 14:24:37.370125 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83rfhlr_ddb177f8-0125-45bc-bbc4-130d689e10c9/extract/0.log" Nov 28 14:24:37 crc kubenswrapper[4631]: I1128 14:24:37.527622 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-j6d86_e55d89ce-6920-4cb5-951f-5b36cb73b6df/extract-utilities/0.log" Nov 28 14:24:37 crc kubenswrapper[4631]: I1128 14:24:37.726203 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-j6d86_e55d89ce-6920-4cb5-951f-5b36cb73b6df/extract-utilities/0.log" Nov 28 14:24:37 crc kubenswrapper[4631]: I1128 14:24:37.749434 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-j6d86_e55d89ce-6920-4cb5-951f-5b36cb73b6df/extract-content/0.log" Nov 28 14:24:37 crc kubenswrapper[4631]: I1128 14:24:37.822029 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-j6d86_e55d89ce-6920-4cb5-951f-5b36cb73b6df/extract-content/0.log" Nov 28 14:24:37 crc kubenswrapper[4631]: I1128 14:24:37.963474 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-j6d86_e55d89ce-6920-4cb5-951f-5b36cb73b6df/extract-utilities/0.log" Nov 28 14:24:38 crc kubenswrapper[4631]: I1128 14:24:38.024745 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-j6d86_e55d89ce-6920-4cb5-951f-5b36cb73b6df/extract-content/0.log" Nov 28 14:24:38 crc kubenswrapper[4631]: I1128 14:24:38.285654 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-7hvgh_bbd0356c-2316-430e-81d3-c3c505c86e9e/extract-utilities/0.log" Nov 28 14:24:38 crc kubenswrapper[4631]: I1128 14:24:38.428488 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-j6d86_e55d89ce-6920-4cb5-951f-5b36cb73b6df/registry-server/0.log" Nov 28 14:24:38 crc kubenswrapper[4631]: I1128 14:24:38.598769 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-7hvgh_bbd0356c-2316-430e-81d3-c3c505c86e9e/extract-content/0.log" Nov 28 14:24:38 crc kubenswrapper[4631]: I1128 14:24:38.604706 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-7hvgh_bbd0356c-2316-430e-81d3-c3c505c86e9e/extract-content/0.log" Nov 28 14:24:38 crc kubenswrapper[4631]: I1128 14:24:38.644692 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-7hvgh_bbd0356c-2316-430e-81d3-c3c505c86e9e/extract-utilities/0.log" Nov 28 14:24:38 crc kubenswrapper[4631]: I1128 14:24:38.861884 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-7hvgh_bbd0356c-2316-430e-81d3-c3c505c86e9e/extract-utilities/0.log" Nov 28 14:24:38 crc kubenswrapper[4631]: I1128 14:24:38.864221 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-7hvgh_bbd0356c-2316-430e-81d3-c3c505c86e9e/extract-content/0.log" Nov 28 14:24:39 crc kubenswrapper[4631]: I1128 14:24:39.205111 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-gsrzv_4eb86ab5-b771-49dc-8bfb-49ce46c0c712/marketplace-operator/0.log" Nov 28 14:24:39 crc kubenswrapper[4631]: I1128 14:24:39.422490 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-fsz47_abdc286b-6b18-47b2-b974-410a0f9614b9/extract-utilities/0.log" Nov 28 14:24:39 crc kubenswrapper[4631]: I1128 14:24:39.716554 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-fsz47_abdc286b-6b18-47b2-b974-410a0f9614b9/extract-utilities/0.log" Nov 28 14:24:39 crc kubenswrapper[4631]: I1128 14:24:39.719970 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-7hvgh_bbd0356c-2316-430e-81d3-c3c505c86e9e/registry-server/0.log" Nov 28 14:24:39 crc kubenswrapper[4631]: I1128 14:24:39.727912 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-fsz47_abdc286b-6b18-47b2-b974-410a0f9614b9/extract-content/0.log" Nov 28 14:24:39 crc kubenswrapper[4631]: I1128 14:24:39.784633 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-fsz47_abdc286b-6b18-47b2-b974-410a0f9614b9/extract-content/0.log" Nov 28 14:24:39 crc kubenswrapper[4631]: I1128 14:24:39.962530 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-fsz47_abdc286b-6b18-47b2-b974-410a0f9614b9/extract-content/0.log" Nov 28 14:24:39 crc kubenswrapper[4631]: I1128 14:24:39.979784 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-fsz47_abdc286b-6b18-47b2-b974-410a0f9614b9/extract-utilities/0.log" Nov 28 14:24:40 crc kubenswrapper[4631]: I1128 14:24:40.163411 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-fsz47_abdc286b-6b18-47b2-b974-410a0f9614b9/registry-server/0.log" Nov 28 14:24:40 crc kubenswrapper[4631]: I1128 14:24:40.292540 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-mzsr9_d2115a7c-dc8e-4f02-bded-25e080c7de51/extract-utilities/0.log" Nov 28 14:24:40 crc kubenswrapper[4631]: I1128 14:24:40.376845 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-mzsr9_d2115a7c-dc8e-4f02-bded-25e080c7de51/extract-utilities/0.log" Nov 28 14:24:40 crc kubenswrapper[4631]: I1128 14:24:40.458216 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-mzsr9_d2115a7c-dc8e-4f02-bded-25e080c7de51/extract-content/0.log" Nov 28 14:24:40 crc kubenswrapper[4631]: I1128 14:24:40.467798 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-mzsr9_d2115a7c-dc8e-4f02-bded-25e080c7de51/extract-content/0.log" Nov 28 14:24:40 crc kubenswrapper[4631]: I1128 14:24:40.623242 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-mzsr9_d2115a7c-dc8e-4f02-bded-25e080c7de51/extract-utilities/0.log" Nov 28 14:24:40 crc kubenswrapper[4631]: I1128 14:24:40.739505 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-mzsr9_d2115a7c-dc8e-4f02-bded-25e080c7de51/extract-content/0.log" Nov 28 14:24:41 crc kubenswrapper[4631]: I1128 14:24:41.074538 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-mzsr9_d2115a7c-dc8e-4f02-bded-25e080c7de51/registry-server/0.log" Nov 28 14:24:48 crc kubenswrapper[4631]: I1128 14:24:48.513204 4631 scope.go:117] "RemoveContainer" containerID="b2f5e32c4eeb164aad43a3d1940c5d648cc8679d4f599da50345a14cc66f0979" Nov 28 14:24:48 crc kubenswrapper[4631]: E1128 14:24:48.513882 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-47sxw_openshift-machine-config-operator(8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1)\"" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" podUID="8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1" Nov 28 14:24:59 crc kubenswrapper[4631]: I1128 14:24:59.524580 4631 scope.go:117] "RemoveContainer" containerID="b2f5e32c4eeb164aad43a3d1940c5d648cc8679d4f599da50345a14cc66f0979" Nov 28 14:24:59 crc kubenswrapper[4631]: E1128 14:24:59.525400 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-47sxw_openshift-machine-config-operator(8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1)\"" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" podUID="8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1" Nov 28 14:25:12 crc kubenswrapper[4631]: I1128 14:25:12.514488 4631 scope.go:117] "RemoveContainer" containerID="b2f5e32c4eeb164aad43a3d1940c5d648cc8679d4f599da50345a14cc66f0979" Nov 28 14:25:12 crc kubenswrapper[4631]: E1128 14:25:12.515169 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-47sxw_openshift-machine-config-operator(8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1)\"" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" podUID="8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1" Nov 28 14:25:24 crc kubenswrapper[4631]: I1128 14:25:24.513826 4631 scope.go:117] "RemoveContainer" containerID="b2f5e32c4eeb164aad43a3d1940c5d648cc8679d4f599da50345a14cc66f0979" Nov 28 14:25:24 crc kubenswrapper[4631]: E1128 14:25:24.514742 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-47sxw_openshift-machine-config-operator(8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1)\"" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" podUID="8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1" Nov 28 14:25:35 crc kubenswrapper[4631]: I1128 14:25:35.513539 4631 scope.go:117] "RemoveContainer" containerID="b2f5e32c4eeb164aad43a3d1940c5d648cc8679d4f599da50345a14cc66f0979" Nov 28 14:25:35 crc kubenswrapper[4631]: E1128 14:25:35.514348 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-47sxw_openshift-machine-config-operator(8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1)\"" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" podUID="8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1" Nov 28 14:25:49 crc kubenswrapper[4631]: I1128 14:25:49.514636 4631 scope.go:117] "RemoveContainer" containerID="b2f5e32c4eeb164aad43a3d1940c5d648cc8679d4f599da50345a14cc66f0979" Nov 28 14:25:49 crc kubenswrapper[4631]: E1128 14:25:49.515341 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-47sxw_openshift-machine-config-operator(8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1)\"" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" podUID="8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1" Nov 28 14:26:00 crc kubenswrapper[4631]: I1128 14:26:00.512949 4631 scope.go:117] "RemoveContainer" containerID="b2f5e32c4eeb164aad43a3d1940c5d648cc8679d4f599da50345a14cc66f0979" Nov 28 14:26:00 crc kubenswrapper[4631]: E1128 14:26:00.513769 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-47sxw_openshift-machine-config-operator(8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1)\"" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" podUID="8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1" Nov 28 14:26:14 crc kubenswrapper[4631]: I1128 14:26:14.515442 4631 scope.go:117] "RemoveContainer" containerID="b2f5e32c4eeb164aad43a3d1940c5d648cc8679d4f599da50345a14cc66f0979" Nov 28 14:26:14 crc kubenswrapper[4631]: E1128 14:26:14.518164 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-47sxw_openshift-machine-config-operator(8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1)\"" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" podUID="8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1" Nov 28 14:26:27 crc kubenswrapper[4631]: I1128 14:26:27.513122 4631 scope.go:117] "RemoveContainer" containerID="b2f5e32c4eeb164aad43a3d1940c5d648cc8679d4f599da50345a14cc66f0979" Nov 28 14:26:27 crc kubenswrapper[4631]: E1128 14:26:27.514100 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-47sxw_openshift-machine-config-operator(8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1)\"" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" podUID="8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1" Nov 28 14:26:39 crc kubenswrapper[4631]: I1128 14:26:39.444568 4631 generic.go:334] "Generic (PLEG): container finished" podID="a5418d8d-9e54-4825-b19e-22eed3133a61" containerID="1afa63395df40354a7ce29c0fba3b0bc410af13cdc4292c6f37a5479f29e656e" exitCode=0 Nov 28 14:26:39 crc kubenswrapper[4631]: I1128 14:26:39.444628 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-95xfg/must-gather-8gjjm" event={"ID":"a5418d8d-9e54-4825-b19e-22eed3133a61","Type":"ContainerDied","Data":"1afa63395df40354a7ce29c0fba3b0bc410af13cdc4292c6f37a5479f29e656e"} Nov 28 14:26:39 crc kubenswrapper[4631]: I1128 14:26:39.445588 4631 scope.go:117] "RemoveContainer" containerID="1afa63395df40354a7ce29c0fba3b0bc410af13cdc4292c6f37a5479f29e656e" Nov 28 14:26:40 crc kubenswrapper[4631]: I1128 14:26:40.461487 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-95xfg_must-gather-8gjjm_a5418d8d-9e54-4825-b19e-22eed3133a61/gather/0.log" Nov 28 14:26:40 crc kubenswrapper[4631]: I1128 14:26:40.513265 4631 scope.go:117] "RemoveContainer" containerID="b2f5e32c4eeb164aad43a3d1940c5d648cc8679d4f599da50345a14cc66f0979" Nov 28 14:26:40 crc kubenswrapper[4631]: E1128 14:26:40.513541 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-47sxw_openshift-machine-config-operator(8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1)\"" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" podUID="8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1" Nov 28 14:26:52 crc kubenswrapper[4631]: I1128 14:26:52.397269 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-95xfg/must-gather-8gjjm"] Nov 28 14:26:52 crc kubenswrapper[4631]: I1128 14:26:52.398727 4631 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-95xfg/must-gather-8gjjm" podUID="a5418d8d-9e54-4825-b19e-22eed3133a61" containerName="copy" containerID="cri-o://7f767ef445a32dd22acba63789c626676072dc383b0e4417c654f46e21f680ef" gracePeriod=2 Nov 28 14:26:52 crc kubenswrapper[4631]: I1128 14:26:52.404859 4631 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-95xfg/must-gather-8gjjm"] Nov 28 14:26:52 crc kubenswrapper[4631]: I1128 14:26:52.632475 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-95xfg_must-gather-8gjjm_a5418d8d-9e54-4825-b19e-22eed3133a61/copy/0.log" Nov 28 14:26:52 crc kubenswrapper[4631]: I1128 14:26:52.632859 4631 generic.go:334] "Generic (PLEG): container finished" podID="a5418d8d-9e54-4825-b19e-22eed3133a61" containerID="7f767ef445a32dd22acba63789c626676072dc383b0e4417c654f46e21f680ef" exitCode=143 Nov 28 14:26:53 crc kubenswrapper[4631]: I1128 14:26:53.062352 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-95xfg_must-gather-8gjjm_a5418d8d-9e54-4825-b19e-22eed3133a61/copy/0.log" Nov 28 14:26:53 crc kubenswrapper[4631]: I1128 14:26:53.063262 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-95xfg/must-gather-8gjjm" Nov 28 14:26:53 crc kubenswrapper[4631]: I1128 14:26:53.110622 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tq9qx\" (UniqueName: \"kubernetes.io/projected/a5418d8d-9e54-4825-b19e-22eed3133a61-kube-api-access-tq9qx\") pod \"a5418d8d-9e54-4825-b19e-22eed3133a61\" (UID: \"a5418d8d-9e54-4825-b19e-22eed3133a61\") " Nov 28 14:26:53 crc kubenswrapper[4631]: I1128 14:26:53.110836 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/a5418d8d-9e54-4825-b19e-22eed3133a61-must-gather-output\") pod \"a5418d8d-9e54-4825-b19e-22eed3133a61\" (UID: \"a5418d8d-9e54-4825-b19e-22eed3133a61\") " Nov 28 14:26:53 crc kubenswrapper[4631]: I1128 14:26:53.129698 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a5418d8d-9e54-4825-b19e-22eed3133a61-kube-api-access-tq9qx" (OuterVolumeSpecName: "kube-api-access-tq9qx") pod "a5418d8d-9e54-4825-b19e-22eed3133a61" (UID: "a5418d8d-9e54-4825-b19e-22eed3133a61"). InnerVolumeSpecName "kube-api-access-tq9qx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:26:53 crc kubenswrapper[4631]: I1128 14:26:53.213656 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tq9qx\" (UniqueName: \"kubernetes.io/projected/a5418d8d-9e54-4825-b19e-22eed3133a61-kube-api-access-tq9qx\") on node \"crc\" DevicePath \"\"" Nov 28 14:26:53 crc kubenswrapper[4631]: I1128 14:26:53.274770 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a5418d8d-9e54-4825-b19e-22eed3133a61-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "a5418d8d-9e54-4825-b19e-22eed3133a61" (UID: "a5418d8d-9e54-4825-b19e-22eed3133a61"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 14:26:53 crc kubenswrapper[4631]: I1128 14:26:53.316471 4631 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/a5418d8d-9e54-4825-b19e-22eed3133a61-must-gather-output\") on node \"crc\" DevicePath \"\"" Nov 28 14:26:53 crc kubenswrapper[4631]: I1128 14:26:53.530910 4631 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a5418d8d-9e54-4825-b19e-22eed3133a61" path="/var/lib/kubelet/pods/a5418d8d-9e54-4825-b19e-22eed3133a61/volumes" Nov 28 14:26:53 crc kubenswrapper[4631]: I1128 14:26:53.648131 4631 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-95xfg_must-gather-8gjjm_a5418d8d-9e54-4825-b19e-22eed3133a61/copy/0.log" Nov 28 14:26:53 crc kubenswrapper[4631]: I1128 14:26:53.649843 4631 scope.go:117] "RemoveContainer" containerID="7f767ef445a32dd22acba63789c626676072dc383b0e4417c654f46e21f680ef" Nov 28 14:26:53 crc kubenswrapper[4631]: I1128 14:26:53.650141 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-95xfg/must-gather-8gjjm" Nov 28 14:26:53 crc kubenswrapper[4631]: I1128 14:26:53.674762 4631 scope.go:117] "RemoveContainer" containerID="1afa63395df40354a7ce29c0fba3b0bc410af13cdc4292c6f37a5479f29e656e" Nov 28 14:26:54 crc kubenswrapper[4631]: I1128 14:26:54.513690 4631 scope.go:117] "RemoveContainer" containerID="b2f5e32c4eeb164aad43a3d1940c5d648cc8679d4f599da50345a14cc66f0979" Nov 28 14:26:54 crc kubenswrapper[4631]: E1128 14:26:54.514790 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-47sxw_openshift-machine-config-operator(8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1)\"" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" podUID="8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1" Nov 28 14:27:07 crc kubenswrapper[4631]: I1128 14:27:07.069118 4631 scope.go:117] "RemoveContainer" containerID="24f6c67a028a1e168937286b3a24cfa1cde8b0d45198cc361e109ac977ce98f1" Nov 28 14:27:09 crc kubenswrapper[4631]: I1128 14:27:09.513226 4631 scope.go:117] "RemoveContainer" containerID="b2f5e32c4eeb164aad43a3d1940c5d648cc8679d4f599da50345a14cc66f0979" Nov 28 14:27:09 crc kubenswrapper[4631]: E1128 14:27:09.515074 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-47sxw_openshift-machine-config-operator(8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1)\"" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" podUID="8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1" Nov 28 14:27:20 crc kubenswrapper[4631]: I1128 14:27:20.968574 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-s5xnh"] Nov 28 14:27:20 crc kubenswrapper[4631]: E1128 14:27:20.969467 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9cd4efd2-7af0-4384-b565-a15063290cdf" containerName="extract-utilities" Nov 28 14:27:20 crc kubenswrapper[4631]: I1128 14:27:20.969480 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="9cd4efd2-7af0-4384-b565-a15063290cdf" containerName="extract-utilities" Nov 28 14:27:20 crc kubenswrapper[4631]: E1128 14:27:20.969501 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a5418d8d-9e54-4825-b19e-22eed3133a61" containerName="gather" Nov 28 14:27:20 crc kubenswrapper[4631]: I1128 14:27:20.969508 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="a5418d8d-9e54-4825-b19e-22eed3133a61" containerName="gather" Nov 28 14:27:20 crc kubenswrapper[4631]: E1128 14:27:20.969536 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9cd4efd2-7af0-4384-b565-a15063290cdf" containerName="registry-server" Nov 28 14:27:20 crc kubenswrapper[4631]: I1128 14:27:20.969544 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="9cd4efd2-7af0-4384-b565-a15063290cdf" containerName="registry-server" Nov 28 14:27:20 crc kubenswrapper[4631]: E1128 14:27:20.969562 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9cd4efd2-7af0-4384-b565-a15063290cdf" containerName="extract-content" Nov 28 14:27:20 crc kubenswrapper[4631]: I1128 14:27:20.969568 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="9cd4efd2-7af0-4384-b565-a15063290cdf" containerName="extract-content" Nov 28 14:27:20 crc kubenswrapper[4631]: E1128 14:27:20.969582 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a5418d8d-9e54-4825-b19e-22eed3133a61" containerName="copy" Nov 28 14:27:20 crc kubenswrapper[4631]: I1128 14:27:20.969588 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="a5418d8d-9e54-4825-b19e-22eed3133a61" containerName="copy" Nov 28 14:27:20 crc kubenswrapper[4631]: I1128 14:27:20.969751 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="a5418d8d-9e54-4825-b19e-22eed3133a61" containerName="copy" Nov 28 14:27:20 crc kubenswrapper[4631]: I1128 14:27:20.969770 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="9cd4efd2-7af0-4384-b565-a15063290cdf" containerName="registry-server" Nov 28 14:27:20 crc kubenswrapper[4631]: I1128 14:27:20.969785 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="a5418d8d-9e54-4825-b19e-22eed3133a61" containerName="gather" Nov 28 14:27:20 crc kubenswrapper[4631]: I1128 14:27:20.971050 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-s5xnh" Nov 28 14:27:20 crc kubenswrapper[4631]: I1128 14:27:20.989457 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-s5xnh"] Nov 28 14:27:21 crc kubenswrapper[4631]: I1128 14:27:21.051795 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0fea5e92-96a0-4eb9-9c98-061cd06184be-catalog-content\") pod \"redhat-operators-s5xnh\" (UID: \"0fea5e92-96a0-4eb9-9c98-061cd06184be\") " pod="openshift-marketplace/redhat-operators-s5xnh" Nov 28 14:27:21 crc kubenswrapper[4631]: I1128 14:27:21.051927 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fjpns\" (UniqueName: \"kubernetes.io/projected/0fea5e92-96a0-4eb9-9c98-061cd06184be-kube-api-access-fjpns\") pod \"redhat-operators-s5xnh\" (UID: \"0fea5e92-96a0-4eb9-9c98-061cd06184be\") " pod="openshift-marketplace/redhat-operators-s5xnh" Nov 28 14:27:21 crc kubenswrapper[4631]: I1128 14:27:21.052249 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0fea5e92-96a0-4eb9-9c98-061cd06184be-utilities\") pod \"redhat-operators-s5xnh\" (UID: \"0fea5e92-96a0-4eb9-9c98-061cd06184be\") " pod="openshift-marketplace/redhat-operators-s5xnh" Nov 28 14:27:21 crc kubenswrapper[4631]: I1128 14:27:21.153630 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0fea5e92-96a0-4eb9-9c98-061cd06184be-catalog-content\") pod \"redhat-operators-s5xnh\" (UID: \"0fea5e92-96a0-4eb9-9c98-061cd06184be\") " pod="openshift-marketplace/redhat-operators-s5xnh" Nov 28 14:27:21 crc kubenswrapper[4631]: I1128 14:27:21.153691 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fjpns\" (UniqueName: \"kubernetes.io/projected/0fea5e92-96a0-4eb9-9c98-061cd06184be-kube-api-access-fjpns\") pod \"redhat-operators-s5xnh\" (UID: \"0fea5e92-96a0-4eb9-9c98-061cd06184be\") " pod="openshift-marketplace/redhat-operators-s5xnh" Nov 28 14:27:21 crc kubenswrapper[4631]: I1128 14:27:21.153745 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0fea5e92-96a0-4eb9-9c98-061cd06184be-utilities\") pod \"redhat-operators-s5xnh\" (UID: \"0fea5e92-96a0-4eb9-9c98-061cd06184be\") " pod="openshift-marketplace/redhat-operators-s5xnh" Nov 28 14:27:21 crc kubenswrapper[4631]: I1128 14:27:21.154248 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0fea5e92-96a0-4eb9-9c98-061cd06184be-utilities\") pod \"redhat-operators-s5xnh\" (UID: \"0fea5e92-96a0-4eb9-9c98-061cd06184be\") " pod="openshift-marketplace/redhat-operators-s5xnh" Nov 28 14:27:21 crc kubenswrapper[4631]: I1128 14:27:21.154243 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0fea5e92-96a0-4eb9-9c98-061cd06184be-catalog-content\") pod \"redhat-operators-s5xnh\" (UID: \"0fea5e92-96a0-4eb9-9c98-061cd06184be\") " pod="openshift-marketplace/redhat-operators-s5xnh" Nov 28 14:27:21 crc kubenswrapper[4631]: I1128 14:27:21.184147 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fjpns\" (UniqueName: \"kubernetes.io/projected/0fea5e92-96a0-4eb9-9c98-061cd06184be-kube-api-access-fjpns\") pod \"redhat-operators-s5xnh\" (UID: \"0fea5e92-96a0-4eb9-9c98-061cd06184be\") " pod="openshift-marketplace/redhat-operators-s5xnh" Nov 28 14:27:21 crc kubenswrapper[4631]: I1128 14:27:21.292816 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-s5xnh" Nov 28 14:27:22 crc kubenswrapper[4631]: I1128 14:27:22.058170 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-s5xnh"] Nov 28 14:27:22 crc kubenswrapper[4631]: I1128 14:27:22.513078 4631 scope.go:117] "RemoveContainer" containerID="b2f5e32c4eeb164aad43a3d1940c5d648cc8679d4f599da50345a14cc66f0979" Nov 28 14:27:22 crc kubenswrapper[4631]: E1128 14:27:22.513691 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-47sxw_openshift-machine-config-operator(8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1)\"" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" podUID="8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1" Nov 28 14:27:23 crc kubenswrapper[4631]: I1128 14:27:23.019094 4631 generic.go:334] "Generic (PLEG): container finished" podID="0fea5e92-96a0-4eb9-9c98-061cd06184be" containerID="82a4330537993d9574aebf0331f3999e1390170b93b85686b64350ebf1f7b96e" exitCode=0 Nov 28 14:27:23 crc kubenswrapper[4631]: I1128 14:27:23.019135 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-s5xnh" event={"ID":"0fea5e92-96a0-4eb9-9c98-061cd06184be","Type":"ContainerDied","Data":"82a4330537993d9574aebf0331f3999e1390170b93b85686b64350ebf1f7b96e"} Nov 28 14:27:23 crc kubenswrapper[4631]: I1128 14:27:23.019167 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-s5xnh" event={"ID":"0fea5e92-96a0-4eb9-9c98-061cd06184be","Type":"ContainerStarted","Data":"eb184d0e5d96edc0e8dccda36b4acd746b60db1243195fa618821771b94b1880"} Nov 28 14:27:24 crc kubenswrapper[4631]: I1128 14:27:24.033674 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-s5xnh" event={"ID":"0fea5e92-96a0-4eb9-9c98-061cd06184be","Type":"ContainerStarted","Data":"fe1c46591d76025d1b303795640416c863d25bde02cb7b167aec5e5c2860d563"} Nov 28 14:27:27 crc kubenswrapper[4631]: I1128 14:27:27.061708 4631 generic.go:334] "Generic (PLEG): container finished" podID="0fea5e92-96a0-4eb9-9c98-061cd06184be" containerID="fe1c46591d76025d1b303795640416c863d25bde02cb7b167aec5e5c2860d563" exitCode=0 Nov 28 14:27:27 crc kubenswrapper[4631]: I1128 14:27:27.061792 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-s5xnh" event={"ID":"0fea5e92-96a0-4eb9-9c98-061cd06184be","Type":"ContainerDied","Data":"fe1c46591d76025d1b303795640416c863d25bde02cb7b167aec5e5c2860d563"} Nov 28 14:27:28 crc kubenswrapper[4631]: I1128 14:27:28.096437 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-s5xnh" event={"ID":"0fea5e92-96a0-4eb9-9c98-061cd06184be","Type":"ContainerStarted","Data":"37508331fe3708c3a6252d54d1d024c4cf9c09cded73705f1f7b0f73dcfe12a1"} Nov 28 14:27:28 crc kubenswrapper[4631]: I1128 14:27:28.126235 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-s5xnh" podStartSLOduration=3.662809851 podStartE2EDuration="8.126211788s" podCreationTimestamp="2025-11-28 14:27:20 +0000 UTC" firstStartedPulling="2025-11-28 14:27:23.020462603 +0000 UTC m=+3999.827765947" lastFinishedPulling="2025-11-28 14:27:27.48386454 +0000 UTC m=+4004.291167884" observedRunningTime="2025-11-28 14:27:28.123836209 +0000 UTC m=+4004.931139553" watchObservedRunningTime="2025-11-28 14:27:28.126211788 +0000 UTC m=+4004.933515132" Nov 28 14:27:31 crc kubenswrapper[4631]: I1128 14:27:31.293937 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-s5xnh" Nov 28 14:27:31 crc kubenswrapper[4631]: I1128 14:27:31.296210 4631 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-s5xnh" Nov 28 14:27:32 crc kubenswrapper[4631]: I1128 14:27:32.380952 4631 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-s5xnh" podUID="0fea5e92-96a0-4eb9-9c98-061cd06184be" containerName="registry-server" probeResult="failure" output=< Nov 28 14:27:32 crc kubenswrapper[4631]: timeout: failed to connect service ":50051" within 1s Nov 28 14:27:32 crc kubenswrapper[4631]: > Nov 28 14:27:36 crc kubenswrapper[4631]: I1128 14:27:36.513429 4631 scope.go:117] "RemoveContainer" containerID="b2f5e32c4eeb164aad43a3d1940c5d648cc8679d4f599da50345a14cc66f0979" Nov 28 14:27:36 crc kubenswrapper[4631]: E1128 14:27:36.514582 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-47sxw_openshift-machine-config-operator(8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1)\"" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" podUID="8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1" Nov 28 14:27:41 crc kubenswrapper[4631]: I1128 14:27:41.383606 4631 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-s5xnh" Nov 28 14:27:41 crc kubenswrapper[4631]: I1128 14:27:41.448406 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-s5xnh" Nov 28 14:27:41 crc kubenswrapper[4631]: I1128 14:27:41.618416 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-s5xnh"] Nov 28 14:27:43 crc kubenswrapper[4631]: I1128 14:27:43.317564 4631 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-s5xnh" podUID="0fea5e92-96a0-4eb9-9c98-061cd06184be" containerName="registry-server" containerID="cri-o://37508331fe3708c3a6252d54d1d024c4cf9c09cded73705f1f7b0f73dcfe12a1" gracePeriod=2 Nov 28 14:27:43 crc kubenswrapper[4631]: I1128 14:27:43.785657 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-s5xnh" Nov 28 14:27:43 crc kubenswrapper[4631]: I1128 14:27:43.876376 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fjpns\" (UniqueName: \"kubernetes.io/projected/0fea5e92-96a0-4eb9-9c98-061cd06184be-kube-api-access-fjpns\") pod \"0fea5e92-96a0-4eb9-9c98-061cd06184be\" (UID: \"0fea5e92-96a0-4eb9-9c98-061cd06184be\") " Nov 28 14:27:43 crc kubenswrapper[4631]: I1128 14:27:43.876433 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0fea5e92-96a0-4eb9-9c98-061cd06184be-catalog-content\") pod \"0fea5e92-96a0-4eb9-9c98-061cd06184be\" (UID: \"0fea5e92-96a0-4eb9-9c98-061cd06184be\") " Nov 28 14:27:43 crc kubenswrapper[4631]: I1128 14:27:43.876523 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0fea5e92-96a0-4eb9-9c98-061cd06184be-utilities\") pod \"0fea5e92-96a0-4eb9-9c98-061cd06184be\" (UID: \"0fea5e92-96a0-4eb9-9c98-061cd06184be\") " Nov 28 14:27:43 crc kubenswrapper[4631]: I1128 14:27:43.877761 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0fea5e92-96a0-4eb9-9c98-061cd06184be-utilities" (OuterVolumeSpecName: "utilities") pod "0fea5e92-96a0-4eb9-9c98-061cd06184be" (UID: "0fea5e92-96a0-4eb9-9c98-061cd06184be"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 14:27:43 crc kubenswrapper[4631]: I1128 14:27:43.887937 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0fea5e92-96a0-4eb9-9c98-061cd06184be-kube-api-access-fjpns" (OuterVolumeSpecName: "kube-api-access-fjpns") pod "0fea5e92-96a0-4eb9-9c98-061cd06184be" (UID: "0fea5e92-96a0-4eb9-9c98-061cd06184be"). InnerVolumeSpecName "kube-api-access-fjpns". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:27:43 crc kubenswrapper[4631]: I1128 14:27:43.978260 4631 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0fea5e92-96a0-4eb9-9c98-061cd06184be-utilities\") on node \"crc\" DevicePath \"\"" Nov 28 14:27:43 crc kubenswrapper[4631]: I1128 14:27:43.978648 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fjpns\" (UniqueName: \"kubernetes.io/projected/0fea5e92-96a0-4eb9-9c98-061cd06184be-kube-api-access-fjpns\") on node \"crc\" DevicePath \"\"" Nov 28 14:27:44 crc kubenswrapper[4631]: I1128 14:27:44.010210 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0fea5e92-96a0-4eb9-9c98-061cd06184be-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0fea5e92-96a0-4eb9-9c98-061cd06184be" (UID: "0fea5e92-96a0-4eb9-9c98-061cd06184be"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 14:27:44 crc kubenswrapper[4631]: I1128 14:27:44.079924 4631 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0fea5e92-96a0-4eb9-9c98-061cd06184be-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 28 14:27:44 crc kubenswrapper[4631]: I1128 14:27:44.335786 4631 generic.go:334] "Generic (PLEG): container finished" podID="0fea5e92-96a0-4eb9-9c98-061cd06184be" containerID="37508331fe3708c3a6252d54d1d024c4cf9c09cded73705f1f7b0f73dcfe12a1" exitCode=0 Nov 28 14:27:44 crc kubenswrapper[4631]: I1128 14:27:44.335856 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-s5xnh" event={"ID":"0fea5e92-96a0-4eb9-9c98-061cd06184be","Type":"ContainerDied","Data":"37508331fe3708c3a6252d54d1d024c4cf9c09cded73705f1f7b0f73dcfe12a1"} Nov 28 14:27:44 crc kubenswrapper[4631]: I1128 14:27:44.335935 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-s5xnh" event={"ID":"0fea5e92-96a0-4eb9-9c98-061cd06184be","Type":"ContainerDied","Data":"eb184d0e5d96edc0e8dccda36b4acd746b60db1243195fa618821771b94b1880"} Nov 28 14:27:44 crc kubenswrapper[4631]: I1128 14:27:44.335965 4631 scope.go:117] "RemoveContainer" containerID="37508331fe3708c3a6252d54d1d024c4cf9c09cded73705f1f7b0f73dcfe12a1" Nov 28 14:27:44 crc kubenswrapper[4631]: I1128 14:27:44.337481 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-s5xnh" Nov 28 14:27:44 crc kubenswrapper[4631]: I1128 14:27:44.365501 4631 scope.go:117] "RemoveContainer" containerID="fe1c46591d76025d1b303795640416c863d25bde02cb7b167aec5e5c2860d563" Nov 28 14:27:44 crc kubenswrapper[4631]: I1128 14:27:44.406523 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-s5xnh"] Nov 28 14:27:44 crc kubenswrapper[4631]: I1128 14:27:44.421169 4631 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-s5xnh"] Nov 28 14:27:44 crc kubenswrapper[4631]: I1128 14:27:44.424800 4631 scope.go:117] "RemoveContainer" containerID="82a4330537993d9574aebf0331f3999e1390170b93b85686b64350ebf1f7b96e" Nov 28 14:27:44 crc kubenswrapper[4631]: I1128 14:27:44.463370 4631 scope.go:117] "RemoveContainer" containerID="37508331fe3708c3a6252d54d1d024c4cf9c09cded73705f1f7b0f73dcfe12a1" Nov 28 14:27:44 crc kubenswrapper[4631]: E1128 14:27:44.464038 4631 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"37508331fe3708c3a6252d54d1d024c4cf9c09cded73705f1f7b0f73dcfe12a1\": container with ID starting with 37508331fe3708c3a6252d54d1d024c4cf9c09cded73705f1f7b0f73dcfe12a1 not found: ID does not exist" containerID="37508331fe3708c3a6252d54d1d024c4cf9c09cded73705f1f7b0f73dcfe12a1" Nov 28 14:27:44 crc kubenswrapper[4631]: I1128 14:27:44.464080 4631 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"37508331fe3708c3a6252d54d1d024c4cf9c09cded73705f1f7b0f73dcfe12a1"} err="failed to get container status \"37508331fe3708c3a6252d54d1d024c4cf9c09cded73705f1f7b0f73dcfe12a1\": rpc error: code = NotFound desc = could not find container \"37508331fe3708c3a6252d54d1d024c4cf9c09cded73705f1f7b0f73dcfe12a1\": container with ID starting with 37508331fe3708c3a6252d54d1d024c4cf9c09cded73705f1f7b0f73dcfe12a1 not found: ID does not exist" Nov 28 14:27:44 crc kubenswrapper[4631]: I1128 14:27:44.464110 4631 scope.go:117] "RemoveContainer" containerID="fe1c46591d76025d1b303795640416c863d25bde02cb7b167aec5e5c2860d563" Nov 28 14:27:44 crc kubenswrapper[4631]: E1128 14:27:44.464416 4631 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fe1c46591d76025d1b303795640416c863d25bde02cb7b167aec5e5c2860d563\": container with ID starting with fe1c46591d76025d1b303795640416c863d25bde02cb7b167aec5e5c2860d563 not found: ID does not exist" containerID="fe1c46591d76025d1b303795640416c863d25bde02cb7b167aec5e5c2860d563" Nov 28 14:27:44 crc kubenswrapper[4631]: I1128 14:27:44.464441 4631 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fe1c46591d76025d1b303795640416c863d25bde02cb7b167aec5e5c2860d563"} err="failed to get container status \"fe1c46591d76025d1b303795640416c863d25bde02cb7b167aec5e5c2860d563\": rpc error: code = NotFound desc = could not find container \"fe1c46591d76025d1b303795640416c863d25bde02cb7b167aec5e5c2860d563\": container with ID starting with fe1c46591d76025d1b303795640416c863d25bde02cb7b167aec5e5c2860d563 not found: ID does not exist" Nov 28 14:27:44 crc kubenswrapper[4631]: I1128 14:27:44.464456 4631 scope.go:117] "RemoveContainer" containerID="82a4330537993d9574aebf0331f3999e1390170b93b85686b64350ebf1f7b96e" Nov 28 14:27:44 crc kubenswrapper[4631]: E1128 14:27:44.464880 4631 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"82a4330537993d9574aebf0331f3999e1390170b93b85686b64350ebf1f7b96e\": container with ID starting with 82a4330537993d9574aebf0331f3999e1390170b93b85686b64350ebf1f7b96e not found: ID does not exist" containerID="82a4330537993d9574aebf0331f3999e1390170b93b85686b64350ebf1f7b96e" Nov 28 14:27:44 crc kubenswrapper[4631]: I1128 14:27:44.464991 4631 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"82a4330537993d9574aebf0331f3999e1390170b93b85686b64350ebf1f7b96e"} err="failed to get container status \"82a4330537993d9574aebf0331f3999e1390170b93b85686b64350ebf1f7b96e\": rpc error: code = NotFound desc = could not find container \"82a4330537993d9574aebf0331f3999e1390170b93b85686b64350ebf1f7b96e\": container with ID starting with 82a4330537993d9574aebf0331f3999e1390170b93b85686b64350ebf1f7b96e not found: ID does not exist" Nov 28 14:27:45 crc kubenswrapper[4631]: I1128 14:27:45.531633 4631 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0fea5e92-96a0-4eb9-9c98-061cd06184be" path="/var/lib/kubelet/pods/0fea5e92-96a0-4eb9-9c98-061cd06184be/volumes" Nov 28 14:27:49 crc kubenswrapper[4631]: I1128 14:27:49.514841 4631 scope.go:117] "RemoveContainer" containerID="b2f5e32c4eeb164aad43a3d1940c5d648cc8679d4f599da50345a14cc66f0979" Nov 28 14:27:49 crc kubenswrapper[4631]: E1128 14:27:49.516522 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-47sxw_openshift-machine-config-operator(8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1)\"" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" podUID="8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1" Nov 28 14:28:00 crc kubenswrapper[4631]: I1128 14:28:00.513356 4631 scope.go:117] "RemoveContainer" containerID="b2f5e32c4eeb164aad43a3d1940c5d648cc8679d4f599da50345a14cc66f0979" Nov 28 14:28:00 crc kubenswrapper[4631]: E1128 14:28:00.513973 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-47sxw_openshift-machine-config-operator(8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1)\"" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" podUID="8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1" Nov 28 14:28:12 crc kubenswrapper[4631]: I1128 14:28:12.109864 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-mb575"] Nov 28 14:28:12 crc kubenswrapper[4631]: E1128 14:28:12.117148 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0fea5e92-96a0-4eb9-9c98-061cd06184be" containerName="registry-server" Nov 28 14:28:12 crc kubenswrapper[4631]: I1128 14:28:12.117191 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="0fea5e92-96a0-4eb9-9c98-061cd06184be" containerName="registry-server" Nov 28 14:28:12 crc kubenswrapper[4631]: E1128 14:28:12.117211 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0fea5e92-96a0-4eb9-9c98-061cd06184be" containerName="extract-utilities" Nov 28 14:28:12 crc kubenswrapper[4631]: I1128 14:28:12.117220 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="0fea5e92-96a0-4eb9-9c98-061cd06184be" containerName="extract-utilities" Nov 28 14:28:12 crc kubenswrapper[4631]: E1128 14:28:12.117271 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0fea5e92-96a0-4eb9-9c98-061cd06184be" containerName="extract-content" Nov 28 14:28:12 crc kubenswrapper[4631]: I1128 14:28:12.117279 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="0fea5e92-96a0-4eb9-9c98-061cd06184be" containerName="extract-content" Nov 28 14:28:12 crc kubenswrapper[4631]: I1128 14:28:12.117561 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="0fea5e92-96a0-4eb9-9c98-061cd06184be" containerName="registry-server" Nov 28 14:28:12 crc kubenswrapper[4631]: I1128 14:28:12.119384 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mb575" Nov 28 14:28:12 crc kubenswrapper[4631]: I1128 14:28:12.122131 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-mb575"] Nov 28 14:28:12 crc kubenswrapper[4631]: I1128 14:28:12.171811 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4c335782-89ec-40c3-8625-6f1865e44d4d-catalog-content\") pod \"certified-operators-mb575\" (UID: \"4c335782-89ec-40c3-8625-6f1865e44d4d\") " pod="openshift-marketplace/certified-operators-mb575" Nov 28 14:28:12 crc kubenswrapper[4631]: I1128 14:28:12.172111 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vkxqn\" (UniqueName: \"kubernetes.io/projected/4c335782-89ec-40c3-8625-6f1865e44d4d-kube-api-access-vkxqn\") pod \"certified-operators-mb575\" (UID: \"4c335782-89ec-40c3-8625-6f1865e44d4d\") " pod="openshift-marketplace/certified-operators-mb575" Nov 28 14:28:12 crc kubenswrapper[4631]: I1128 14:28:12.172391 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4c335782-89ec-40c3-8625-6f1865e44d4d-utilities\") pod \"certified-operators-mb575\" (UID: \"4c335782-89ec-40c3-8625-6f1865e44d4d\") " pod="openshift-marketplace/certified-operators-mb575" Nov 28 14:28:12 crc kubenswrapper[4631]: I1128 14:28:12.273764 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4c335782-89ec-40c3-8625-6f1865e44d4d-catalog-content\") pod \"certified-operators-mb575\" (UID: \"4c335782-89ec-40c3-8625-6f1865e44d4d\") " pod="openshift-marketplace/certified-operators-mb575" Nov 28 14:28:12 crc kubenswrapper[4631]: I1128 14:28:12.273917 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vkxqn\" (UniqueName: \"kubernetes.io/projected/4c335782-89ec-40c3-8625-6f1865e44d4d-kube-api-access-vkxqn\") pod \"certified-operators-mb575\" (UID: \"4c335782-89ec-40c3-8625-6f1865e44d4d\") " pod="openshift-marketplace/certified-operators-mb575" Nov 28 14:28:12 crc kubenswrapper[4631]: I1128 14:28:12.274001 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4c335782-89ec-40c3-8625-6f1865e44d4d-utilities\") pod \"certified-operators-mb575\" (UID: \"4c335782-89ec-40c3-8625-6f1865e44d4d\") " pod="openshift-marketplace/certified-operators-mb575" Nov 28 14:28:12 crc kubenswrapper[4631]: I1128 14:28:12.274603 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4c335782-89ec-40c3-8625-6f1865e44d4d-catalog-content\") pod \"certified-operators-mb575\" (UID: \"4c335782-89ec-40c3-8625-6f1865e44d4d\") " pod="openshift-marketplace/certified-operators-mb575" Nov 28 14:28:12 crc kubenswrapper[4631]: I1128 14:28:12.274778 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4c335782-89ec-40c3-8625-6f1865e44d4d-utilities\") pod \"certified-operators-mb575\" (UID: \"4c335782-89ec-40c3-8625-6f1865e44d4d\") " pod="openshift-marketplace/certified-operators-mb575" Nov 28 14:28:12 crc kubenswrapper[4631]: I1128 14:28:12.302338 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vkxqn\" (UniqueName: \"kubernetes.io/projected/4c335782-89ec-40c3-8625-6f1865e44d4d-kube-api-access-vkxqn\") pod \"certified-operators-mb575\" (UID: \"4c335782-89ec-40c3-8625-6f1865e44d4d\") " pod="openshift-marketplace/certified-operators-mb575" Nov 28 14:28:12 crc kubenswrapper[4631]: I1128 14:28:12.314671 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-vj7fh"] Nov 28 14:28:12 crc kubenswrapper[4631]: I1128 14:28:12.317075 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vj7fh" Nov 28 14:28:12 crc kubenswrapper[4631]: I1128 14:28:12.327704 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-vj7fh"] Nov 28 14:28:12 crc kubenswrapper[4631]: I1128 14:28:12.376154 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eb9c922e-7595-46c3-8063-20e003f802a4-utilities\") pod \"community-operators-vj7fh\" (UID: \"eb9c922e-7595-46c3-8063-20e003f802a4\") " pod="openshift-marketplace/community-operators-vj7fh" Nov 28 14:28:12 crc kubenswrapper[4631]: I1128 14:28:12.376741 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fv8v6\" (UniqueName: \"kubernetes.io/projected/eb9c922e-7595-46c3-8063-20e003f802a4-kube-api-access-fv8v6\") pod \"community-operators-vj7fh\" (UID: \"eb9c922e-7595-46c3-8063-20e003f802a4\") " pod="openshift-marketplace/community-operators-vj7fh" Nov 28 14:28:12 crc kubenswrapper[4631]: I1128 14:28:12.376826 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eb9c922e-7595-46c3-8063-20e003f802a4-catalog-content\") pod \"community-operators-vj7fh\" (UID: \"eb9c922e-7595-46c3-8063-20e003f802a4\") " pod="openshift-marketplace/community-operators-vj7fh" Nov 28 14:28:12 crc kubenswrapper[4631]: I1128 14:28:12.449516 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mb575" Nov 28 14:28:12 crc kubenswrapper[4631]: I1128 14:28:12.478436 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eb9c922e-7595-46c3-8063-20e003f802a4-utilities\") pod \"community-operators-vj7fh\" (UID: \"eb9c922e-7595-46c3-8063-20e003f802a4\") " pod="openshift-marketplace/community-operators-vj7fh" Nov 28 14:28:12 crc kubenswrapper[4631]: I1128 14:28:12.478505 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fv8v6\" (UniqueName: \"kubernetes.io/projected/eb9c922e-7595-46c3-8063-20e003f802a4-kube-api-access-fv8v6\") pod \"community-operators-vj7fh\" (UID: \"eb9c922e-7595-46c3-8063-20e003f802a4\") " pod="openshift-marketplace/community-operators-vj7fh" Nov 28 14:28:12 crc kubenswrapper[4631]: I1128 14:28:12.478557 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eb9c922e-7595-46c3-8063-20e003f802a4-catalog-content\") pod \"community-operators-vj7fh\" (UID: \"eb9c922e-7595-46c3-8063-20e003f802a4\") " pod="openshift-marketplace/community-operators-vj7fh" Nov 28 14:28:12 crc kubenswrapper[4631]: I1128 14:28:12.479036 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eb9c922e-7595-46c3-8063-20e003f802a4-utilities\") pod \"community-operators-vj7fh\" (UID: \"eb9c922e-7595-46c3-8063-20e003f802a4\") " pod="openshift-marketplace/community-operators-vj7fh" Nov 28 14:28:12 crc kubenswrapper[4631]: I1128 14:28:12.479045 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eb9c922e-7595-46c3-8063-20e003f802a4-catalog-content\") pod \"community-operators-vj7fh\" (UID: \"eb9c922e-7595-46c3-8063-20e003f802a4\") " pod="openshift-marketplace/community-operators-vj7fh" Nov 28 14:28:12 crc kubenswrapper[4631]: I1128 14:28:12.512851 4631 scope.go:117] "RemoveContainer" containerID="b2f5e32c4eeb164aad43a3d1940c5d648cc8679d4f599da50345a14cc66f0979" Nov 28 14:28:12 crc kubenswrapper[4631]: I1128 14:28:12.512934 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fv8v6\" (UniqueName: \"kubernetes.io/projected/eb9c922e-7595-46c3-8063-20e003f802a4-kube-api-access-fv8v6\") pod \"community-operators-vj7fh\" (UID: \"eb9c922e-7595-46c3-8063-20e003f802a4\") " pod="openshift-marketplace/community-operators-vj7fh" Nov 28 14:28:12 crc kubenswrapper[4631]: E1128 14:28:12.513108 4631 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-47sxw_openshift-machine-config-operator(8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1)\"" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" podUID="8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1" Nov 28 14:28:12 crc kubenswrapper[4631]: I1128 14:28:12.714978 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vj7fh" Nov 28 14:28:12 crc kubenswrapper[4631]: I1128 14:28:12.979401 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-mb575"] Nov 28 14:28:13 crc kubenswrapper[4631]: I1128 14:28:13.375929 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-vj7fh"] Nov 28 14:28:13 crc kubenswrapper[4631]: W1128 14:28:13.510375 4631 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podeb9c922e_7595_46c3_8063_20e003f802a4.slice/crio-0c3e1d89c8157d2883e49b91f143effd6a02db7f67116aef4640521cf95e5caf WatchSource:0}: Error finding container 0c3e1d89c8157d2883e49b91f143effd6a02db7f67116aef4640521cf95e5caf: Status 404 returned error can't find the container with id 0c3e1d89c8157d2883e49b91f143effd6a02db7f67116aef4640521cf95e5caf Nov 28 14:28:13 crc kubenswrapper[4631]: I1128 14:28:13.612866 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mb575" event={"ID":"4c335782-89ec-40c3-8625-6f1865e44d4d","Type":"ContainerStarted","Data":"79020394c2ef6713cb0313d6abebe7fe6b6c8fe87d8314000681fd5507cfceaf"} Nov 28 14:28:13 crc kubenswrapper[4631]: I1128 14:28:13.614446 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vj7fh" event={"ID":"eb9c922e-7595-46c3-8063-20e003f802a4","Type":"ContainerStarted","Data":"0c3e1d89c8157d2883e49b91f143effd6a02db7f67116aef4640521cf95e5caf"} Nov 28 14:28:14 crc kubenswrapper[4631]: I1128 14:28:14.625411 4631 generic.go:334] "Generic (PLEG): container finished" podID="4c335782-89ec-40c3-8625-6f1865e44d4d" containerID="00f627797097229503a2ffc8f1ea9124210c27c7c486b9556d34734fb982035f" exitCode=0 Nov 28 14:28:14 crc kubenswrapper[4631]: I1128 14:28:14.625478 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mb575" event={"ID":"4c335782-89ec-40c3-8625-6f1865e44d4d","Type":"ContainerDied","Data":"00f627797097229503a2ffc8f1ea9124210c27c7c486b9556d34734fb982035f"} Nov 28 14:28:14 crc kubenswrapper[4631]: I1128 14:28:14.627504 4631 generic.go:334] "Generic (PLEG): container finished" podID="eb9c922e-7595-46c3-8063-20e003f802a4" containerID="7363396023e5fdf5a70e19d811321e479394fb32354e86abf89b2341c733411e" exitCode=0 Nov 28 14:28:14 crc kubenswrapper[4631]: I1128 14:28:14.627532 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vj7fh" event={"ID":"eb9c922e-7595-46c3-8063-20e003f802a4","Type":"ContainerDied","Data":"7363396023e5fdf5a70e19d811321e479394fb32354e86abf89b2341c733411e"} Nov 28 14:28:14 crc kubenswrapper[4631]: I1128 14:28:14.629998 4631 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 28 14:28:16 crc kubenswrapper[4631]: I1128 14:28:16.646697 4631 generic.go:334] "Generic (PLEG): container finished" podID="eb9c922e-7595-46c3-8063-20e003f802a4" containerID="8c8a46ecde928782f71c7e2c32354f5dc80f980124ce1fb2db1dd0149ce09a81" exitCode=0 Nov 28 14:28:16 crc kubenswrapper[4631]: I1128 14:28:16.646856 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vj7fh" event={"ID":"eb9c922e-7595-46c3-8063-20e003f802a4","Type":"ContainerDied","Data":"8c8a46ecde928782f71c7e2c32354f5dc80f980124ce1fb2db1dd0149ce09a81"} Nov 28 14:28:19 crc kubenswrapper[4631]: I1128 14:28:19.679188 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mb575" event={"ID":"4c335782-89ec-40c3-8625-6f1865e44d4d","Type":"ContainerStarted","Data":"330ab5bca05e37918bf26995e47e617d4d587c10583e056ccc017908183f2977"} Nov 28 14:28:19 crc kubenswrapper[4631]: I1128 14:28:19.685905 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vj7fh" event={"ID":"eb9c922e-7595-46c3-8063-20e003f802a4","Type":"ContainerStarted","Data":"e891746dfdc7435115332c93e468912774262aa37527b83167ba37f51f4b2b8b"} Nov 28 14:28:19 crc kubenswrapper[4631]: I1128 14:28:19.724417 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-vj7fh" podStartSLOduration=3.241160963 podStartE2EDuration="7.724400966s" podCreationTimestamp="2025-11-28 14:28:12 +0000 UTC" firstStartedPulling="2025-11-28 14:28:14.63292246 +0000 UTC m=+4051.440225804" lastFinishedPulling="2025-11-28 14:28:19.116162463 +0000 UTC m=+4055.923465807" observedRunningTime="2025-11-28 14:28:19.719755321 +0000 UTC m=+4056.527058675" watchObservedRunningTime="2025-11-28 14:28:19.724400966 +0000 UTC m=+4056.531704310" Nov 28 14:28:20 crc kubenswrapper[4631]: I1128 14:28:20.696262 4631 generic.go:334] "Generic (PLEG): container finished" podID="4c335782-89ec-40c3-8625-6f1865e44d4d" containerID="330ab5bca05e37918bf26995e47e617d4d587c10583e056ccc017908183f2977" exitCode=0 Nov 28 14:28:20 crc kubenswrapper[4631]: I1128 14:28:20.696342 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mb575" event={"ID":"4c335782-89ec-40c3-8625-6f1865e44d4d","Type":"ContainerDied","Data":"330ab5bca05e37918bf26995e47e617d4d587c10583e056ccc017908183f2977"} Nov 28 14:28:22 crc kubenswrapper[4631]: I1128 14:28:22.716740 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-vj7fh" Nov 28 14:28:22 crc kubenswrapper[4631]: I1128 14:28:22.717087 4631 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-vj7fh" Nov 28 14:28:22 crc kubenswrapper[4631]: I1128 14:28:22.740846 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mb575" event={"ID":"4c335782-89ec-40c3-8625-6f1865e44d4d","Type":"ContainerStarted","Data":"deae587445475427e2f2dbb4972a47acce6f185bdf05186ebfc0cc5fcdb0ab23"} Nov 28 14:28:22 crc kubenswrapper[4631]: I1128 14:28:22.763423 4631 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-mb575" podStartSLOduration=3.584288135 podStartE2EDuration="10.763405121s" podCreationTimestamp="2025-11-28 14:28:12 +0000 UTC" firstStartedPulling="2025-11-28 14:28:14.629756503 +0000 UTC m=+4051.437059847" lastFinishedPulling="2025-11-28 14:28:21.808873489 +0000 UTC m=+4058.616176833" observedRunningTime="2025-11-28 14:28:22.759887105 +0000 UTC m=+4059.567190449" watchObservedRunningTime="2025-11-28 14:28:22.763405121 +0000 UTC m=+4059.570708465" Nov 28 14:28:22 crc kubenswrapper[4631]: I1128 14:28:22.783298 4631 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-vj7fh" Nov 28 14:28:25 crc kubenswrapper[4631]: I1128 14:28:25.513902 4631 scope.go:117] "RemoveContainer" containerID="b2f5e32c4eeb164aad43a3d1940c5d648cc8679d4f599da50345a14cc66f0979" Nov 28 14:28:26 crc kubenswrapper[4631]: I1128 14:28:26.781486 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-47sxw" event={"ID":"8f2b70e1-4fcc-43ae-891a-ea0d7959a7e1","Type":"ContainerStarted","Data":"2a8464cddf97545c0654f6b652627398d60e98cf6f57dbc01c23252a6a5d7bd8"} Nov 28 14:28:32 crc kubenswrapper[4631]: I1128 14:28:32.449968 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-mb575" Nov 28 14:28:32 crc kubenswrapper[4631]: I1128 14:28:32.450538 4631 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-mb575" Nov 28 14:28:32 crc kubenswrapper[4631]: I1128 14:28:32.504880 4631 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-mb575" Nov 28 14:28:32 crc kubenswrapper[4631]: I1128 14:28:32.761101 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-vj7fh" Nov 28 14:28:32 crc kubenswrapper[4631]: I1128 14:28:32.808987 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-vj7fh"] Nov 28 14:28:32 crc kubenswrapper[4631]: I1128 14:28:32.832824 4631 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-vj7fh" podUID="eb9c922e-7595-46c3-8063-20e003f802a4" containerName="registry-server" containerID="cri-o://e891746dfdc7435115332c93e468912774262aa37527b83167ba37f51f4b2b8b" gracePeriod=2 Nov 28 14:28:33 crc kubenswrapper[4631]: I1128 14:28:33.155204 4631 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-mb575" Nov 28 14:28:33 crc kubenswrapper[4631]: I1128 14:28:33.477784 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vj7fh" Nov 28 14:28:33 crc kubenswrapper[4631]: I1128 14:28:33.628874 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eb9c922e-7595-46c3-8063-20e003f802a4-utilities\") pod \"eb9c922e-7595-46c3-8063-20e003f802a4\" (UID: \"eb9c922e-7595-46c3-8063-20e003f802a4\") " Nov 28 14:28:33 crc kubenswrapper[4631]: I1128 14:28:33.629472 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fv8v6\" (UniqueName: \"kubernetes.io/projected/eb9c922e-7595-46c3-8063-20e003f802a4-kube-api-access-fv8v6\") pod \"eb9c922e-7595-46c3-8063-20e003f802a4\" (UID: \"eb9c922e-7595-46c3-8063-20e003f802a4\") " Nov 28 14:28:33 crc kubenswrapper[4631]: I1128 14:28:33.629587 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eb9c922e-7595-46c3-8063-20e003f802a4-catalog-content\") pod \"eb9c922e-7595-46c3-8063-20e003f802a4\" (UID: \"eb9c922e-7595-46c3-8063-20e003f802a4\") " Nov 28 14:28:33 crc kubenswrapper[4631]: I1128 14:28:33.630420 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/eb9c922e-7595-46c3-8063-20e003f802a4-utilities" (OuterVolumeSpecName: "utilities") pod "eb9c922e-7595-46c3-8063-20e003f802a4" (UID: "eb9c922e-7595-46c3-8063-20e003f802a4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 14:28:33 crc kubenswrapper[4631]: I1128 14:28:33.638997 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eb9c922e-7595-46c3-8063-20e003f802a4-kube-api-access-fv8v6" (OuterVolumeSpecName: "kube-api-access-fv8v6") pod "eb9c922e-7595-46c3-8063-20e003f802a4" (UID: "eb9c922e-7595-46c3-8063-20e003f802a4"). InnerVolumeSpecName "kube-api-access-fv8v6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:28:33 crc kubenswrapper[4631]: I1128 14:28:33.691485 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/eb9c922e-7595-46c3-8063-20e003f802a4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "eb9c922e-7595-46c3-8063-20e003f802a4" (UID: "eb9c922e-7595-46c3-8063-20e003f802a4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 14:28:33 crc kubenswrapper[4631]: I1128 14:28:33.731756 4631 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eb9c922e-7595-46c3-8063-20e003f802a4-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 28 14:28:33 crc kubenswrapper[4631]: I1128 14:28:33.731799 4631 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eb9c922e-7595-46c3-8063-20e003f802a4-utilities\") on node \"crc\" DevicePath \"\"" Nov 28 14:28:33 crc kubenswrapper[4631]: I1128 14:28:33.731809 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fv8v6\" (UniqueName: \"kubernetes.io/projected/eb9c922e-7595-46c3-8063-20e003f802a4-kube-api-access-fv8v6\") on node \"crc\" DevicePath \"\"" Nov 28 14:28:33 crc kubenswrapper[4631]: I1128 14:28:33.846870 4631 generic.go:334] "Generic (PLEG): container finished" podID="eb9c922e-7595-46c3-8063-20e003f802a4" containerID="e891746dfdc7435115332c93e468912774262aa37527b83167ba37f51f4b2b8b" exitCode=0 Nov 28 14:28:33 crc kubenswrapper[4631]: I1128 14:28:33.846925 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vj7fh" event={"ID":"eb9c922e-7595-46c3-8063-20e003f802a4","Type":"ContainerDied","Data":"e891746dfdc7435115332c93e468912774262aa37527b83167ba37f51f4b2b8b"} Nov 28 14:28:33 crc kubenswrapper[4631]: I1128 14:28:33.847102 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vj7fh" Nov 28 14:28:33 crc kubenswrapper[4631]: I1128 14:28:33.847429 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vj7fh" event={"ID":"eb9c922e-7595-46c3-8063-20e003f802a4","Type":"ContainerDied","Data":"0c3e1d89c8157d2883e49b91f143effd6a02db7f67116aef4640521cf95e5caf"} Nov 28 14:28:33 crc kubenswrapper[4631]: I1128 14:28:33.847484 4631 scope.go:117] "RemoveContainer" containerID="e891746dfdc7435115332c93e468912774262aa37527b83167ba37f51f4b2b8b" Nov 28 14:28:33 crc kubenswrapper[4631]: I1128 14:28:33.897636 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-vj7fh"] Nov 28 14:28:33 crc kubenswrapper[4631]: I1128 14:28:33.897776 4631 scope.go:117] "RemoveContainer" containerID="8c8a46ecde928782f71c7e2c32354f5dc80f980124ce1fb2db1dd0149ce09a81" Nov 28 14:28:33 crc kubenswrapper[4631]: I1128 14:28:33.907264 4631 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-vj7fh"] Nov 28 14:28:33 crc kubenswrapper[4631]: I1128 14:28:33.927497 4631 scope.go:117] "RemoveContainer" containerID="7363396023e5fdf5a70e19d811321e479394fb32354e86abf89b2341c733411e" Nov 28 14:28:33 crc kubenswrapper[4631]: I1128 14:28:33.966913 4631 scope.go:117] "RemoveContainer" containerID="e891746dfdc7435115332c93e468912774262aa37527b83167ba37f51f4b2b8b" Nov 28 14:28:33 crc kubenswrapper[4631]: E1128 14:28:33.967566 4631 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e891746dfdc7435115332c93e468912774262aa37527b83167ba37f51f4b2b8b\": container with ID starting with e891746dfdc7435115332c93e468912774262aa37527b83167ba37f51f4b2b8b not found: ID does not exist" containerID="e891746dfdc7435115332c93e468912774262aa37527b83167ba37f51f4b2b8b" Nov 28 14:28:33 crc kubenswrapper[4631]: I1128 14:28:33.967680 4631 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e891746dfdc7435115332c93e468912774262aa37527b83167ba37f51f4b2b8b"} err="failed to get container status \"e891746dfdc7435115332c93e468912774262aa37527b83167ba37f51f4b2b8b\": rpc error: code = NotFound desc = could not find container \"e891746dfdc7435115332c93e468912774262aa37527b83167ba37f51f4b2b8b\": container with ID starting with e891746dfdc7435115332c93e468912774262aa37527b83167ba37f51f4b2b8b not found: ID does not exist" Nov 28 14:28:33 crc kubenswrapper[4631]: I1128 14:28:33.967802 4631 scope.go:117] "RemoveContainer" containerID="8c8a46ecde928782f71c7e2c32354f5dc80f980124ce1fb2db1dd0149ce09a81" Nov 28 14:28:33 crc kubenswrapper[4631]: E1128 14:28:33.968200 4631 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8c8a46ecde928782f71c7e2c32354f5dc80f980124ce1fb2db1dd0149ce09a81\": container with ID starting with 8c8a46ecde928782f71c7e2c32354f5dc80f980124ce1fb2db1dd0149ce09a81 not found: ID does not exist" containerID="8c8a46ecde928782f71c7e2c32354f5dc80f980124ce1fb2db1dd0149ce09a81" Nov 28 14:28:33 crc kubenswrapper[4631]: I1128 14:28:33.968253 4631 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8c8a46ecde928782f71c7e2c32354f5dc80f980124ce1fb2db1dd0149ce09a81"} err="failed to get container status \"8c8a46ecde928782f71c7e2c32354f5dc80f980124ce1fb2db1dd0149ce09a81\": rpc error: code = NotFound desc = could not find container \"8c8a46ecde928782f71c7e2c32354f5dc80f980124ce1fb2db1dd0149ce09a81\": container with ID starting with 8c8a46ecde928782f71c7e2c32354f5dc80f980124ce1fb2db1dd0149ce09a81 not found: ID does not exist" Nov 28 14:28:33 crc kubenswrapper[4631]: I1128 14:28:33.968397 4631 scope.go:117] "RemoveContainer" containerID="7363396023e5fdf5a70e19d811321e479394fb32354e86abf89b2341c733411e" Nov 28 14:28:33 crc kubenswrapper[4631]: E1128 14:28:33.968761 4631 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7363396023e5fdf5a70e19d811321e479394fb32354e86abf89b2341c733411e\": container with ID starting with 7363396023e5fdf5a70e19d811321e479394fb32354e86abf89b2341c733411e not found: ID does not exist" containerID="7363396023e5fdf5a70e19d811321e479394fb32354e86abf89b2341c733411e" Nov 28 14:28:33 crc kubenswrapper[4631]: I1128 14:28:33.968885 4631 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7363396023e5fdf5a70e19d811321e479394fb32354e86abf89b2341c733411e"} err="failed to get container status \"7363396023e5fdf5a70e19d811321e479394fb32354e86abf89b2341c733411e\": rpc error: code = NotFound desc = could not find container \"7363396023e5fdf5a70e19d811321e479394fb32354e86abf89b2341c733411e\": container with ID starting with 7363396023e5fdf5a70e19d811321e479394fb32354e86abf89b2341c733411e not found: ID does not exist" Nov 28 14:28:35 crc kubenswrapper[4631]: I1128 14:28:35.167962 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-mb575"] Nov 28 14:28:35 crc kubenswrapper[4631]: I1128 14:28:35.544840 4631 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eb9c922e-7595-46c3-8063-20e003f802a4" path="/var/lib/kubelet/pods/eb9c922e-7595-46c3-8063-20e003f802a4/volumes" Nov 28 14:28:35 crc kubenswrapper[4631]: I1128 14:28:35.546666 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-j6d86"] Nov 28 14:28:35 crc kubenswrapper[4631]: I1128 14:28:35.546868 4631 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-j6d86" podUID="e55d89ce-6920-4cb5-951f-5b36cb73b6df" containerName="registry-server" containerID="cri-o://bf3bde6fa4d0b3a9fa83477c1e20b76eb3779b1e7c943d18b7e4edb87cd5d0b5" gracePeriod=2 Nov 28 14:28:35 crc kubenswrapper[4631]: E1128 14:28:35.684320 4631 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of bf3bde6fa4d0b3a9fa83477c1e20b76eb3779b1e7c943d18b7e4edb87cd5d0b5 is running failed: container process not found" containerID="bf3bde6fa4d0b3a9fa83477c1e20b76eb3779b1e7c943d18b7e4edb87cd5d0b5" cmd=["grpc_health_probe","-addr=:50051"] Nov 28 14:28:35 crc kubenswrapper[4631]: E1128 14:28:35.684838 4631 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of bf3bde6fa4d0b3a9fa83477c1e20b76eb3779b1e7c943d18b7e4edb87cd5d0b5 is running failed: container process not found" containerID="bf3bde6fa4d0b3a9fa83477c1e20b76eb3779b1e7c943d18b7e4edb87cd5d0b5" cmd=["grpc_health_probe","-addr=:50051"] Nov 28 14:28:35 crc kubenswrapper[4631]: E1128 14:28:35.685217 4631 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of bf3bde6fa4d0b3a9fa83477c1e20b76eb3779b1e7c943d18b7e4edb87cd5d0b5 is running failed: container process not found" containerID="bf3bde6fa4d0b3a9fa83477c1e20b76eb3779b1e7c943d18b7e4edb87cd5d0b5" cmd=["grpc_health_probe","-addr=:50051"] Nov 28 14:28:35 crc kubenswrapper[4631]: E1128 14:28:35.685250 4631 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of bf3bde6fa4d0b3a9fa83477c1e20b76eb3779b1e7c943d18b7e4edb87cd5d0b5 is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/certified-operators-j6d86" podUID="e55d89ce-6920-4cb5-951f-5b36cb73b6df" containerName="registry-server" Nov 28 14:28:35 crc kubenswrapper[4631]: I1128 14:28:35.875999 4631 generic.go:334] "Generic (PLEG): container finished" podID="e55d89ce-6920-4cb5-951f-5b36cb73b6df" containerID="bf3bde6fa4d0b3a9fa83477c1e20b76eb3779b1e7c943d18b7e4edb87cd5d0b5" exitCode=0 Nov 28 14:28:35 crc kubenswrapper[4631]: I1128 14:28:35.876040 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-j6d86" event={"ID":"e55d89ce-6920-4cb5-951f-5b36cb73b6df","Type":"ContainerDied","Data":"bf3bde6fa4d0b3a9fa83477c1e20b76eb3779b1e7c943d18b7e4edb87cd5d0b5"} Nov 28 14:28:36 crc kubenswrapper[4631]: I1128 14:28:36.042916 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-j6d86" Nov 28 14:28:36 crc kubenswrapper[4631]: I1128 14:28:36.186427 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w8wm8\" (UniqueName: \"kubernetes.io/projected/e55d89ce-6920-4cb5-951f-5b36cb73b6df-kube-api-access-w8wm8\") pod \"e55d89ce-6920-4cb5-951f-5b36cb73b6df\" (UID: \"e55d89ce-6920-4cb5-951f-5b36cb73b6df\") " Nov 28 14:28:36 crc kubenswrapper[4631]: I1128 14:28:36.186500 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e55d89ce-6920-4cb5-951f-5b36cb73b6df-utilities\") pod \"e55d89ce-6920-4cb5-951f-5b36cb73b6df\" (UID: \"e55d89ce-6920-4cb5-951f-5b36cb73b6df\") " Nov 28 14:28:36 crc kubenswrapper[4631]: I1128 14:28:36.186899 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e55d89ce-6920-4cb5-951f-5b36cb73b6df-utilities" (OuterVolumeSpecName: "utilities") pod "e55d89ce-6920-4cb5-951f-5b36cb73b6df" (UID: "e55d89ce-6920-4cb5-951f-5b36cb73b6df"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 14:28:36 crc kubenswrapper[4631]: I1128 14:28:36.187853 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e55d89ce-6920-4cb5-951f-5b36cb73b6df-catalog-content\") pod \"e55d89ce-6920-4cb5-951f-5b36cb73b6df\" (UID: \"e55d89ce-6920-4cb5-951f-5b36cb73b6df\") " Nov 28 14:28:36 crc kubenswrapper[4631]: I1128 14:28:36.189548 4631 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e55d89ce-6920-4cb5-951f-5b36cb73b6df-utilities\") on node \"crc\" DevicePath \"\"" Nov 28 14:28:36 crc kubenswrapper[4631]: I1128 14:28:36.198409 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e55d89ce-6920-4cb5-951f-5b36cb73b6df-kube-api-access-w8wm8" (OuterVolumeSpecName: "kube-api-access-w8wm8") pod "e55d89ce-6920-4cb5-951f-5b36cb73b6df" (UID: "e55d89ce-6920-4cb5-951f-5b36cb73b6df"). InnerVolumeSpecName "kube-api-access-w8wm8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:28:36 crc kubenswrapper[4631]: I1128 14:28:36.251383 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e55d89ce-6920-4cb5-951f-5b36cb73b6df-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e55d89ce-6920-4cb5-951f-5b36cb73b6df" (UID: "e55d89ce-6920-4cb5-951f-5b36cb73b6df"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 14:28:36 crc kubenswrapper[4631]: I1128 14:28:36.291957 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w8wm8\" (UniqueName: \"kubernetes.io/projected/e55d89ce-6920-4cb5-951f-5b36cb73b6df-kube-api-access-w8wm8\") on node \"crc\" DevicePath \"\"" Nov 28 14:28:36 crc kubenswrapper[4631]: I1128 14:28:36.292004 4631 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e55d89ce-6920-4cb5-951f-5b36cb73b6df-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 28 14:28:36 crc kubenswrapper[4631]: I1128 14:28:36.886954 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-j6d86" event={"ID":"e55d89ce-6920-4cb5-951f-5b36cb73b6df","Type":"ContainerDied","Data":"89b23b1cf431c44dba24f085d59c592553c3429350295eed27045c7376f90c3a"} Nov 28 14:28:36 crc kubenswrapper[4631]: I1128 14:28:36.887015 4631 scope.go:117] "RemoveContainer" containerID="bf3bde6fa4d0b3a9fa83477c1e20b76eb3779b1e7c943d18b7e4edb87cd5d0b5" Nov 28 14:28:36 crc kubenswrapper[4631]: I1128 14:28:36.887355 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-j6d86" Nov 28 14:28:36 crc kubenswrapper[4631]: I1128 14:28:36.908426 4631 scope.go:117] "RemoveContainer" containerID="7e682c5ae6d59cdc8567819f7f55ff90c05902f477a5998ed7b2fe6480af3f31" Nov 28 14:28:36 crc kubenswrapper[4631]: I1128 14:28:36.940758 4631 scope.go:117] "RemoveContainer" containerID="28e1b5bfd08f3699b9ccea95f82293a45eb357d4f5f079759f7d9317b0087988" Nov 28 14:28:36 crc kubenswrapper[4631]: I1128 14:28:36.949209 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-j6d86"] Nov 28 14:28:36 crc kubenswrapper[4631]: I1128 14:28:36.973831 4631 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-j6d86"] Nov 28 14:28:37 crc kubenswrapper[4631]: I1128 14:28:37.526357 4631 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e55d89ce-6920-4cb5-951f-5b36cb73b6df" path="/var/lib/kubelet/pods/e55d89ce-6920-4cb5-951f-5b36cb73b6df/volumes" Nov 28 14:30:00 crc kubenswrapper[4631]: I1128 14:30:00.180141 4631 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29405670-h5gqh"] Nov 28 14:30:00 crc kubenswrapper[4631]: E1128 14:30:00.181071 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e55d89ce-6920-4cb5-951f-5b36cb73b6df" containerName="extract-utilities" Nov 28 14:30:00 crc kubenswrapper[4631]: I1128 14:30:00.181086 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="e55d89ce-6920-4cb5-951f-5b36cb73b6df" containerName="extract-utilities" Nov 28 14:30:00 crc kubenswrapper[4631]: E1128 14:30:00.181102 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eb9c922e-7595-46c3-8063-20e003f802a4" containerName="extract-content" Nov 28 14:30:00 crc kubenswrapper[4631]: I1128 14:30:00.181110 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="eb9c922e-7595-46c3-8063-20e003f802a4" containerName="extract-content" Nov 28 14:30:00 crc kubenswrapper[4631]: E1128 14:30:00.181130 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eb9c922e-7595-46c3-8063-20e003f802a4" containerName="extract-utilities" Nov 28 14:30:00 crc kubenswrapper[4631]: I1128 14:30:00.181141 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="eb9c922e-7595-46c3-8063-20e003f802a4" containerName="extract-utilities" Nov 28 14:30:00 crc kubenswrapper[4631]: E1128 14:30:00.181152 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e55d89ce-6920-4cb5-951f-5b36cb73b6df" containerName="registry-server" Nov 28 14:30:00 crc kubenswrapper[4631]: I1128 14:30:00.181159 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="e55d89ce-6920-4cb5-951f-5b36cb73b6df" containerName="registry-server" Nov 28 14:30:00 crc kubenswrapper[4631]: E1128 14:30:00.181182 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e55d89ce-6920-4cb5-951f-5b36cb73b6df" containerName="extract-content" Nov 28 14:30:00 crc kubenswrapper[4631]: I1128 14:30:00.181189 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="e55d89ce-6920-4cb5-951f-5b36cb73b6df" containerName="extract-content" Nov 28 14:30:00 crc kubenswrapper[4631]: E1128 14:30:00.181201 4631 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eb9c922e-7595-46c3-8063-20e003f802a4" containerName="registry-server" Nov 28 14:30:00 crc kubenswrapper[4631]: I1128 14:30:00.181209 4631 state_mem.go:107] "Deleted CPUSet assignment" podUID="eb9c922e-7595-46c3-8063-20e003f802a4" containerName="registry-server" Nov 28 14:30:00 crc kubenswrapper[4631]: I1128 14:30:00.181493 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="e55d89ce-6920-4cb5-951f-5b36cb73b6df" containerName="registry-server" Nov 28 14:30:00 crc kubenswrapper[4631]: I1128 14:30:00.181521 4631 memory_manager.go:354] "RemoveStaleState removing state" podUID="eb9c922e-7595-46c3-8063-20e003f802a4" containerName="registry-server" Nov 28 14:30:00 crc kubenswrapper[4631]: I1128 14:30:00.182603 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29405670-h5gqh" Nov 28 14:30:00 crc kubenswrapper[4631]: I1128 14:30:00.186452 4631 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 28 14:30:00 crc kubenswrapper[4631]: I1128 14:30:00.195883 4631 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 28 14:30:00 crc kubenswrapper[4631]: I1128 14:30:00.222925 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29405670-h5gqh"] Nov 28 14:30:00 crc kubenswrapper[4631]: I1128 14:30:00.273976 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3d968638-2820-4912-8b47-e44c300d119e-config-volume\") pod \"collect-profiles-29405670-h5gqh\" (UID: \"3d968638-2820-4912-8b47-e44c300d119e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405670-h5gqh" Nov 28 14:30:00 crc kubenswrapper[4631]: I1128 14:30:00.274109 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8fvc5\" (UniqueName: \"kubernetes.io/projected/3d968638-2820-4912-8b47-e44c300d119e-kube-api-access-8fvc5\") pod \"collect-profiles-29405670-h5gqh\" (UID: \"3d968638-2820-4912-8b47-e44c300d119e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405670-h5gqh" Nov 28 14:30:00 crc kubenswrapper[4631]: I1128 14:30:00.274189 4631 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3d968638-2820-4912-8b47-e44c300d119e-secret-volume\") pod \"collect-profiles-29405670-h5gqh\" (UID: \"3d968638-2820-4912-8b47-e44c300d119e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405670-h5gqh" Nov 28 14:30:00 crc kubenswrapper[4631]: I1128 14:30:00.376115 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3d968638-2820-4912-8b47-e44c300d119e-config-volume\") pod \"collect-profiles-29405670-h5gqh\" (UID: \"3d968638-2820-4912-8b47-e44c300d119e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405670-h5gqh" Nov 28 14:30:00 crc kubenswrapper[4631]: I1128 14:30:00.376184 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8fvc5\" (UniqueName: \"kubernetes.io/projected/3d968638-2820-4912-8b47-e44c300d119e-kube-api-access-8fvc5\") pod \"collect-profiles-29405670-h5gqh\" (UID: \"3d968638-2820-4912-8b47-e44c300d119e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405670-h5gqh" Nov 28 14:30:00 crc kubenswrapper[4631]: I1128 14:30:00.376251 4631 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3d968638-2820-4912-8b47-e44c300d119e-secret-volume\") pod \"collect-profiles-29405670-h5gqh\" (UID: \"3d968638-2820-4912-8b47-e44c300d119e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405670-h5gqh" Nov 28 14:30:00 crc kubenswrapper[4631]: I1128 14:30:00.377211 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3d968638-2820-4912-8b47-e44c300d119e-config-volume\") pod \"collect-profiles-29405670-h5gqh\" (UID: \"3d968638-2820-4912-8b47-e44c300d119e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405670-h5gqh" Nov 28 14:30:00 crc kubenswrapper[4631]: I1128 14:30:00.382691 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3d968638-2820-4912-8b47-e44c300d119e-secret-volume\") pod \"collect-profiles-29405670-h5gqh\" (UID: \"3d968638-2820-4912-8b47-e44c300d119e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405670-h5gqh" Nov 28 14:30:00 crc kubenswrapper[4631]: I1128 14:30:00.401934 4631 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8fvc5\" (UniqueName: \"kubernetes.io/projected/3d968638-2820-4912-8b47-e44c300d119e-kube-api-access-8fvc5\") pod \"collect-profiles-29405670-h5gqh\" (UID: \"3d968638-2820-4912-8b47-e44c300d119e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405670-h5gqh" Nov 28 14:30:00 crc kubenswrapper[4631]: I1128 14:30:00.519299 4631 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29405670-h5gqh" Nov 28 14:30:00 crc kubenswrapper[4631]: I1128 14:30:00.984828 4631 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29405670-h5gqh"] Nov 28 14:30:01 crc kubenswrapper[4631]: I1128 14:30:01.651088 4631 generic.go:334] "Generic (PLEG): container finished" podID="3d968638-2820-4912-8b47-e44c300d119e" containerID="e4d1d328f96db4187b8e580eb2adc7965dfe91aa723bc623d5e7b9bbc71467a4" exitCode=0 Nov 28 14:30:01 crc kubenswrapper[4631]: I1128 14:30:01.651165 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29405670-h5gqh" event={"ID":"3d968638-2820-4912-8b47-e44c300d119e","Type":"ContainerDied","Data":"e4d1d328f96db4187b8e580eb2adc7965dfe91aa723bc623d5e7b9bbc71467a4"} Nov 28 14:30:01 crc kubenswrapper[4631]: I1128 14:30:01.651452 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29405670-h5gqh" event={"ID":"3d968638-2820-4912-8b47-e44c300d119e","Type":"ContainerStarted","Data":"c0b60bb717cb51503ac6fc5afa175d7eff9d1963258ff2650a5788e20636e429"} Nov 28 14:30:03 crc kubenswrapper[4631]: I1128 14:30:03.081763 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29405670-h5gqh" Nov 28 14:30:03 crc kubenswrapper[4631]: I1128 14:30:03.139078 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8fvc5\" (UniqueName: \"kubernetes.io/projected/3d968638-2820-4912-8b47-e44c300d119e-kube-api-access-8fvc5\") pod \"3d968638-2820-4912-8b47-e44c300d119e\" (UID: \"3d968638-2820-4912-8b47-e44c300d119e\") " Nov 28 14:30:03 crc kubenswrapper[4631]: I1128 14:30:03.140305 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3d968638-2820-4912-8b47-e44c300d119e-secret-volume\") pod \"3d968638-2820-4912-8b47-e44c300d119e\" (UID: \"3d968638-2820-4912-8b47-e44c300d119e\") " Nov 28 14:30:03 crc kubenswrapper[4631]: I1128 14:30:03.140341 4631 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3d968638-2820-4912-8b47-e44c300d119e-config-volume\") pod \"3d968638-2820-4912-8b47-e44c300d119e\" (UID: \"3d968638-2820-4912-8b47-e44c300d119e\") " Nov 28 14:30:03 crc kubenswrapper[4631]: I1128 14:30:03.141381 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3d968638-2820-4912-8b47-e44c300d119e-config-volume" (OuterVolumeSpecName: "config-volume") pod "3d968638-2820-4912-8b47-e44c300d119e" (UID: "3d968638-2820-4912-8b47-e44c300d119e"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 14:30:03 crc kubenswrapper[4631]: I1128 14:30:03.150547 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3d968638-2820-4912-8b47-e44c300d119e-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "3d968638-2820-4912-8b47-e44c300d119e" (UID: "3d968638-2820-4912-8b47-e44c300d119e"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 14:30:03 crc kubenswrapper[4631]: I1128 14:30:03.150688 4631 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3d968638-2820-4912-8b47-e44c300d119e-kube-api-access-8fvc5" (OuterVolumeSpecName: "kube-api-access-8fvc5") pod "3d968638-2820-4912-8b47-e44c300d119e" (UID: "3d968638-2820-4912-8b47-e44c300d119e"). InnerVolumeSpecName "kube-api-access-8fvc5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 14:30:03 crc kubenswrapper[4631]: I1128 14:30:03.242848 4631 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3d968638-2820-4912-8b47-e44c300d119e-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 28 14:30:03 crc kubenswrapper[4631]: I1128 14:30:03.242895 4631 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3d968638-2820-4912-8b47-e44c300d119e-config-volume\") on node \"crc\" DevicePath \"\"" Nov 28 14:30:03 crc kubenswrapper[4631]: I1128 14:30:03.242907 4631 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8fvc5\" (UniqueName: \"kubernetes.io/projected/3d968638-2820-4912-8b47-e44c300d119e-kube-api-access-8fvc5\") on node \"crc\" DevicePath \"\"" Nov 28 14:30:03 crc kubenswrapper[4631]: I1128 14:30:03.671117 4631 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29405670-h5gqh" event={"ID":"3d968638-2820-4912-8b47-e44c300d119e","Type":"ContainerDied","Data":"c0b60bb717cb51503ac6fc5afa175d7eff9d1963258ff2650a5788e20636e429"} Nov 28 14:30:03 crc kubenswrapper[4631]: I1128 14:30:03.671476 4631 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c0b60bb717cb51503ac6fc5afa175d7eff9d1963258ff2650a5788e20636e429" Nov 28 14:30:03 crc kubenswrapper[4631]: I1128 14:30:03.671195 4631 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29405670-h5gqh" Nov 28 14:30:04 crc kubenswrapper[4631]: I1128 14:30:04.172871 4631 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29405625-ncrbh"] Nov 28 14:30:04 crc kubenswrapper[4631]: I1128 14:30:04.183651 4631 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29405625-ncrbh"] Nov 28 14:30:05 crc kubenswrapper[4631]: I1128 14:30:05.528345 4631 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9583083a-b161-492e-93c6-088fb16c06fe" path="/var/lib/kubelet/pods/9583083a-b161-492e-93c6-088fb16c06fe/volumes" Nov 28 14:30:07 crc kubenswrapper[4631]: I1128 14:30:07.245004 4631 scope.go:117] "RemoveContainer" containerID="19f40940349ba6d21797abd21b448bca573afc98fe0bfafd6332b170448653c7" var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515112330765024450 0ustar coreroot‹íÁ  ÷Om7 €7šÞ'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015112330765017365 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015112320254016500 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015112320255015451 5ustar corecore